You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
*Equally contributing first authors,
✉Correspondence,
†Project Leader
Work done when interning at Bytedance.
We proposed Flash-VStream, an efficient VLM with a novel Flash Memory mechanism that enables real-time understanding and Q&A of extremely long video streams. Our model achieves outstanding accuracy and efficiency on EgoSchema, MLVU, LVBench, MVBench and Video-MME Benchmarks.
News
[2025/6/26] 🔥 [ICCV 2025] Flash-VStream-Qwen is coming! We release the
homepage,
paper,
Code,
and model.
[2024/06/12] Flash-VStream-LLaVA is coming! We release the
homepage,
paper,
code
and model
for Flash-VStream.
We release the dataset for VStream-QA benchmark.
If you find this project useful in your research, please consider citing:
@article{zhang2025flashvstream,
title={Flash-VStream: Efficient Real-Time Understanding for Long Video Streams},
author={Haoji Zhang and Yiqin Wang and Yansong Tang and Yong Liu and Jiashi Feng and Xiaojie Jin},
journal={arXiv preprint arXiv:2506.23825},
year={2025},
}
@article{zhang2024flashvstream,
title={Flash-vstream: Memory-based real-time understanding for long video streams},
author={Zhang, Haoji and Wang, Yiqin and Tang, Yansong and Liu, Yong and Feng, Jiashi and Dai, Jifeng and Jin, Xiaojie},
journal={arXiv preprint arXiv:2406.08085},
year={2024}
}
Acknowledgement
We would like to thank the following repos for their great work: