Major requirements
pytorch=1.13.1hydracv2pandas
We exported all packages installed in environment on which this code is tested to run with conda list -e > requirements.txt and generated requirements are available at ./requirements.txt. Please note that these requirements may contain unnecessary libraries unrelated to STEP.
Download the APT-36K dataset and put it under ./datasets/APT-36k folder. Full consolidated json apt36k_annotations.json used for training is available at Link. Organise the data in following architecture
├── data_root
│ └── Sequences
| : └── clips
| : └── im1.png, im1.json
| : :
| : └── imn.png, imn.json
| :
| └── Sequences
| └── clips
| └── im1.png, im1.json
| :
| └── imn.png, im2.json
├── apt36k_annotations.json
Download the APT-10K dataset and put it under ./datasets/APT10k folder. Full consolidated json ap10k.json used for training is available at Link. Follow similar setup as APT-36K, and place .json in root.
Download the CrowdPose dataset and put it under ./datasets/CrowdPose folder. Full consolidated json ap10k.json used for training is available at Link. Follow similar setup as APT-36K, and place .json in root.
We provide various settings in run.sh. Detailed configs for all datasets and ablation studies are availabe in ./configs/ folder.
python -W ignore main.py config=step_config.yaml "+run_title=$run_title" \ "pipeline.train=True" "data=aptmmpose_nokpts_kptsemb.yaml" "config: which config to use from./configs/folderpipeline.train: Sets mode for Inference/Trainingdata: change config accordingly atconfig/datawith appropriate paths pointing to training datasets, Validation datasets.
Download the trained models and consolidated Jsons from Here
Place downloaded models at location pointed by the key snaps.model_save_dir in config flag; and the .json file as mentioned in Dataset section.
Coming Soon
Code in this repository uses utilities and implementation style from : MMPose, PyTracking. We thank the authors for their amazing works and sharing the code.
@article{verma2025step,
title={STEP: Simultaneous Tracking and Estimation of Pose for Animals and Humans},
author={Verma, Shashikant and Katti, Harish and Debnath, Soumyaratna and Swamy, Yamuna and Raman, Shanmuganathan},
journal={arXiv preprint arXiv:2503.13344},
year={2025}
}