I resigned from using jupyter on docker and I choosed to run notebooks locally. Libraries were installed manually. File requirements.txt will be provided later.
Data are stored in data directory.
You can download .zip from: link (permission needed).
The purpose of these notebooks is to convert files to dataframes stored in pickle files. The preffered way to execute pipeline is to run notebooks in order from 1 to 6.\
- grace_process.ipynb - notebook with GRACE data processing code
- era5_process.ipynb - notebook with ERA5 data processing code
- prepare_gpm_imerg_data.ipynb - notebook with GPM IMERG data processing code
- measurements_excel_to_geopandas.ipynb - notebook for converting excel file with measurements to geopandas dataframe
- prepare_measurements_input.ipynb - notebook for preparing measurements data from excel to pickle file
- merge_datasets.ipynb - notebook for merging all datasets to one pickle file
train_v1.ipynb - notebook with training code
For training you need to have prepared pickle file with dataframes from data processing step. Training was conducted on Google Colab due to problem with installing ray tune on MacOS with M1 chip.
test_model.ipynb - notebook with testing code
- Install docker on your machine.
- Build image from Dockerfile:
docker build https://raw.githubusercontent.com/radekszostak/grace-pl/master/Dockerfile -t grace-pl --no-cache - Create container from image:
docker create -p 8888:8888 --name grace-pl-container grace-pl:latest
docker start --interactive grace-pl-container- Open Jupyter Lab at http://localhost:8888/ or attach running container to IDE.