You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Co-authored-by: Kevin M Jablonka <32935233+kjappelbaum@users.noreply.github.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Copy file name to clipboardExpand all lines: README.md
+23Lines changed: 23 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -62,3 +62,26 @@ and log-in to `wandb` with your API key per [here](https://docs.wandb.ai/quickst
62
62
### Adding a new dataset (to the model training pipline)
63
63
64
64
We specify datasets by creating a new function [here](src/chemnlp/data/hf_datasets.py) which is named per the dataset on Hugging Face. At present the function must accept a tokenizer and return back the tokenized train and validation datasets.
65
+
66
+
### Installing submodules
67
+
68
+
In order to ensure you also clone and install the required submodules (i.e. gpt-neox) you will have to do one of the following;
> This will automatically initialize and update each submodule in the repository, including nested submodules if any of the submodules in the repository have submodules themselve
81
+
82
+
- Initialise and install the submodule after cloning
83
+
84
+
```
85
+
git submodule init # registers submodule
86
+
git submodule update # clones and updates submodule
0 commit comments