| CARVIEW |
Select Language
HTTP/2 200
last-modified: Thu, 10 Feb 2022 01:06:17 GMT
via: 1.1 google, 1.1 varnish, 1.1 varnish, 1.1 varnish
cache-control: max-age=3600
x-cloud-trace-context: a61c2ec8228e35133dd0c7df58728f7a
x-frame-options: SAMEORIGIN
server: Google Frontend
content-security-policy: frame-ancestors 'none'
content-type: text/html; charset=utf-8
accept-ranges: bytes
age: 1934036
date: Thu, 01 Jan 2026 00:58:14 GMT
x-served-by: cache-lga21956-LGA, cache-lga21939-LGA, cache-bom-vanm7210066-BOM
x-cache: MISS, HIT, HIT
x-timer: S1767229094.184250,VS0,VE1
content-length: 44910
[2202.04200] MaskGIT: Masked Generative Image Transformer
Skip to main content
We gratefully acknowledge support from the Simons Foundation, member institutions, and all contributors.
Donate
Computer Science > Computer Vision and Pattern Recognition
arXiv:2202.04200 (cs)
[Submitted on 8 Feb 2022]
Title:MaskGIT: Masked Generative Image Transformer
View a PDF of the paper titled MaskGIT: Masked Generative Image Transformer, by Huiwen Chang and 4 other authors
View PDF
Abstract:Generative transformers have experienced rapid popularity growth in the computer vision community in synthesizing high-fidelity and high-resolution images. The best generative transformer models so far, however, still treat an image naively as a sequence of tokens, and decode an image sequentially following the raster scan ordering (i.e. line-by-line). We find this strategy neither optimal nor efficient. This paper proposes a novel image synthesis paradigm using a bidirectional transformer decoder, which we term MaskGIT. During training, MaskGIT learns to predict randomly masked tokens by attending to tokens in all directions. At inference time, the model begins with generating all tokens of an image simultaneously, and then refines the image iteratively conditioned on the previous generation. Our experiments demonstrate that MaskGIT significantly outperforms the state-of-the-art transformer model on the ImageNet dataset, and accelerates autoregressive decoding by up to 64x. Besides, we illustrate that MaskGIT can be easily extended to various image editing tasks, such as inpainting, extrapolation, and image manipulation.
| Subjects: | Computer Vision and Pattern Recognition (cs.CV) |
| Cite as: | arXiv:2202.04200 [cs.CV] |
| (or arXiv:2202.04200v1 [cs.CV] for this version) | |
| https://doi.org/10.48550/arXiv.2202.04200
arXiv-issued DOI via DataCite
|
Full-text links:
Access Paper:
- View PDF
- TeX Source
View a PDF of the paper titled MaskGIT: Masked Generative Image Transformer, by Huiwen Chang and 4 other authors
References & Citations
export BibTeX citation
Loading...
Bibliographic and Citation Tools
Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)
Code, Data and Media Associated with this Article
alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)
Demos
Recommenders and Search Tools
Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
arXivLabs: experimental projects with community collaborators
arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.
Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.
Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.