This is the fusion stage training and inference of the VIPL-SLP submission for CV-ISLR challenge.
Download the estimated keypoints, extracted RGB and depth features from Google Drive, and put them under the root of project:
Download the processed information dicts from Google Drive, and put them in the ./data
Feel free to contact us if the link is invalid.
conda env create -f environment.ymlDownload the pretrained weights from Google Drive and put them in the ./weights
- For RGB data:
python main.py --config ./configs/test_single_rgb.yaml --load-weights weights/single_rgb.ptExpected performance: Average Topk-1 : 34.55%
- For skeleton data:
python main.py --config ./configs/test_single_skeleton.yaml --load-weights weights/sk_phase2.ptExpected performance: Average Topk-1 : 46.00%
- For RGB+Skeleton data:
python main.py --config ./configs/test_fusion_rgbd.yaml --load-weights ./weights/fusion_rgbd.ptExpected performance: Average Topk-1 : 56.87%
- For Depth data:
python main.py --config ./configs/test_single_depth.yaml --load-weights ./weights/single_depth.ptExpected performance: Average Topk-1 : 28.84%
- For RGB+Skeleton+Depth data:
python main.py --config ./configs/test_fusion_rgbd.yaml --load-weights ./weights/fusion_rgbd.ptExpected performance: Average Topk-1 : 57.98%
python main.py --config ./configs/train_fusion_rgb.yamlpython main.py --config ./configs/train_fusion_rgbd.yaml