You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
by Kate Rakelly*, Evan Shelhamer*, Trevor Darrell, Alexei A. Efros, and Sergey Levine
UC Berkeley
Learning-based methods for visual segmentation have made progress on particular
types of segmentation tasks, but are limited by the necessary supervision, the
narrow definitions of fixed tasks, and the lack of control during inference for
correcting errors. To remedy the rigidity and annotation burden of standard
approaches, we address the problem of few-shot segmentation: given few image
and few pixel supervision, segment any images accordingly. We propose guided
networks, which extract a latent task representation from any amount of
supervision, and optimize our architecture end-to-end for fast, accurate
few-shot segmentation. Our method can switch tasks without further optimization
and quickly update when given more guidance. We report the first results for
segmentation from one pixel per concept and show real-time interactive video
segmentation. Our unified approach propagates pixel annotations across space
for interactive segmentation, across time for video segmentation, and across
scenes for semantic segmentation. Our guided segmentor is state-of-the-art in
accuracy for the amount of annotation and time.
This is a work-in-progress, not yet a reference implementation of the paper, and could change at any time.
for few-shot interactive image segmentation and few-shot semantic segmentation, see this branch.
for few-shot video object segmentation, see branch video-seg (note: this is older code, for pytorch 0.3.1)
TODO
port to pytorch 1.0
push branch for interactive video segmentation
reconcile branches into unified implementation of few-shot segmentation
Please check back soon for improvements, pre-trained models, and usage notebooks!