We using the pre-extracted features coming from this awesome paper R$^2$-Tuning, can be downloaded from HuggingFace Hub directly. And We express our sincere gratitude for their contribution to the community.
Please follow our baseline to prepare the dataset and place the corresponding files in the correct directory. And change the config file to the correct path.
Here are the origin video datasets download links:
# Single GPU
python tools/launch.py <path-to-config>
# Multiple GPUs on a single node (elastic)
torchrun --nproc_per_node=<num-gpus> tools/launch.py <path-to-config>Arguments of tools/launch.py
configThe config file to use--checkpointThe checkpoint file to load from--resumeThe checkpoint file to resume from--work_dirWorking directory--evalEvaluation only--dumpDump inference outputs--seedThe random seed to use--ampWhether to use automatic mixed precision training--debugDebug mode (detectnanduring training)--launcherThe job launcher to use
python tools/launch.py <path-to-config> --checkpoint <path-to-checkpoint> --evalIf problems occur when reproducing the results, please feel free to contact us at github or email.
Maybe you need to change the config file to the correct path.
Some issues may be fixed by these issues in Baseline Repository
We would like to express our sincere gratitude to the following authors for their contributions to the community: