Jiqing Zhang, Bo Dong, Haiwei Zhang, Jianchuan Ding, Felix Heide, Baocai Yin, Xin Yang
[project] [paper]
The code is based on SiamFC++ and tested on Ubuntu 20.04 with PyTorch 1.8.0.
Download our preprocessed test dataset of FE240hz. (The whole FE240hz dataset can be downloaded here).
Download the pretrained model and put it into ./snapshots/stnet.
Change dataset path at line 32 in videoanalyst/engine/tester/tester_impl/eventdata.py. data_root="/your_data_path/img_120_split"
run python main/test.py --config experiments/test/fe240/fe240.yaml
the predicted bounding boxes are saved in logs/EVENT-Benchmark/.
Download our preprocessing test dataset of VisEvent. (The whole VisEvent dataset can be downloaded here).
Download the pretrained model and put it into ./snapshots/stnet.
Change dataset path at line 32 in videoanalyst/engine/tester/tester_impl/eventdata.py, data_root="/your_data_path/img_120_split"
Change model path at line 25 in experiments/test/fe240/fe240.yaml, pretrain_model_path: "snapshots/stnet/fe240.pkl"
run python main/test.py --config experiments/test/fe240/fe240.yaml
the predicted bounding boxes are be saved in logs/EVENT-Benchmark/.
Please cite our paper if you find the work useful:
@inproceedings{zhang2022stnet,
title={Spiking Transformers for Event-based Single Object Tracking},
author={Zhang, Jiqing and Dong, Bo and Zhang, Haiwei and Ding, Jianchuan and Heide, Felix and Yin, Baocai and Yang, Xin},
booktitle={Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition},
year={2022}
}