You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The code is based on SiamFC++ and tested on Ubuntu 20.04 with PyTorch 1.8.0.
Test on FE240hz Dataset
Download our preprocessed test dataset of FE240hz. (The whole FE240hz dataset can be downloaded here).
Download the pretrained model and put it into ./snapshots/stnet.
Change dataset path at line 32 in videoanalyst/engine/tester/tester_impl/eventdata.py. data_root="/your_data_path/img_120_split"
run python main/test.py --config experiments/test/fe240/fe240.yaml the predicted bounding boxes are saved in logs/EVENT-Benchmark/.
The predicted bounding box format: An N×4 matrix with each line representing object location [xmin, ymin, width, height] in one event frame.
Test on VisEvent Dataset
Download our preprocessing test dataset of VisEvent. (The whole VisEvent dataset can be downloaded here).
Download the pretrained model and put it into ./snapshots/stnet.
Change dataset path at line 32 in videoanalyst/engine/tester/tester_impl/eventdata.py, data_root="/your_data_path/img_120_split"
Change model path at line 25 in experiments/test/fe240/fe240.yaml, pretrain_model_path: "snapshots/stnet/fe240.pkl"
run python main/test.py --config experiments/test/fe240/fe240.yaml the predicted bounding boxes are be saved in logs/EVENT-Benchmark/.
The predicted bounding box format: An N×4 matrix with each line representing object location [xmin, ymin, width, height] in one event frame.
Citation
Please cite our paper if you find the work useful:
@inproceedings{zhang2022stnet,
title={Spiking Transformers for Event-based Single Object Tracking},
author={Zhang, Jiqing and Dong, Bo and Zhang, Haiwei and Ding, Jianchuan and Heide, Felix and Yin, Baocai and Yang, Xin},
booktitle={Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition},
year={2022}
}