mkdir datadownload and unzip the following datasets into data
- reason identification and classification for evaluation
- claim sentences search for sentence modeling (use dirname
ibm_claim) - internet argument corpus v1.1 for post modeling (use dirname
iac_v1.1)
mkdir trial trial/data trial/ckpt
cd srcrun ./data_ibm.py or ./data_iac.py
./train.pycheck ./train.py --help for arguments
and see config.json for more configurations.
this part is messy due to the experimental nature of this project, but basically:
- run
eval_prep_data_*.pyto prepare the evaluation data - run
eval_embed*.pyto embed the prepared data using the trained model - run
eval_classification*.pyfor classification andeval_clustering*.pyfor clustering - there are also a few scripts named
explore*.pyfor explorations
datafor storing raw datatrial/datafor storing processed datatrial/ckptfor model checkpointstrial/logfor tensorboard summaries