- Download the dataset from huggingface
- Place the downloaded dataset in the
demo/inputdirectory. - Download the models from:
- GCN model: GCN
- GCN without edges: GCN_no_edges
- Place the downloaded models in the
demo/models/directory. - Rename the folders as
undirected_gnnandno_edge_gnnrespectively.
- Set PYTHONPATH to the parent directory:
export PYTHONPATH=demo/ - Navigate to the
demodirectory in your terminal. - Install the required dependencies using:
pip install -r requirements.txt - Run the Streamlit app using:
streamlit run src/demo.py
Alternatively you can visit the hosted demo at: Demo
- Download the dataset from huggingface
- Place the downloaded dataset in the
modeling/directory. - Install the required dependencies using:
pip install -r demo/requirements.txt - Navigate to the
modelingdirectory in your terminal. - Run gnn.ipynb or heuristic_methods.ipynb using Jupyter Notebook or Jupyter Lab.
- Navigate to the
preprocessingdirectory in your terminal.- preprocess.ipynb: Contains all the preprocessing steps to create the dataset from raw wikipedia articles collected using Wikipedia API.
- generate_embeddings.ipynb: Contains steps to generate document embeddings for the articles using a pre-trained embedding model.
- Metin Usta
- 504251504