You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We propose SAM-PT, an extension of the Segment Anything Model (SAM) for zero-shot video segmentation. Our work offers a simple yet effective point-based perspective in video object segmentation research. For more details, refer to our paper.
Video Object Segmentation Demo
Annotators only provide a few points to denote the target object at the first video frame to get video segmentation results. Please visit our project page for more visualizations, including qualitative results on DAVIS 2017 videos and more Avatar clips.
Interactive Point-Based Video Segmentation
Annotators can interactively add or remove points to refine the segmentation results.
Documentation
Explore our step-by-step guides to get up and running:
Getting Started: Learn how to set up your environment and run the demo.
Prepare Datasets: Instructions on acquiring and prepping necessary datasets.
If you find SAM-PT useful in your research or if you refer to the results mentioned in our work, please star ⭐ this repository and consider citing 📝:
@article{sam-pt,
title = {Segment Anything Meets Point Tracking},
author = {Rajič, Frano and Ke, Lei and Tai, Yu-Wing and Tang, Chi-Keung and Danelljan, Martin and Yu, Fisher},
journal = {arXiv:2307.01197},
year = {2023}
}
About
SAM-PT: Extending SAM to zero-shot video segmentation with point-based tracking.