| CARVIEW |
CVPR 2024, Seattle
Overview and Call for Papers
The First Workshop on Efficient and On-Device Generation (EDGE) at CVPR 2024 will focus on the latest advancements of generative AI in the computer vision domain, with an emphasis on efficiencies across multiple aspects. We encourage techniques that enable generative models to be trained more efficiently and/or run on resource-constrained devices, such as mobile phones and edge devices. Through these efforts, we envision a future where these permeating generative AI capabilities become significantly more accessible with virtuous scalability and plateauing carbon footprint.
The topics involved in the workshop include but are not limited to:
- Training efficiency aspect:
- Methods for reducing the memory footprint of generative models.
- Using knowledge distillation to support model compression.
- Design of efficient computational modules and algorithms for generative models
- Hardware-aware training acceleration and energy efficient computing.
- Online training under edge constraints.
- Parameter efficient tuning.
- Inference efficiency aspect:
- Techniques for reducing the computational complexity of generative models, such as model compression, quantization, sparsification, and distillation.
- Advanced sampling techniques for fast inference.
- Approaches for efficient deployment and distribution generative models across multiple devices or edge devices.
- Hardware-aware inference acceleration and energy efficient computing.
- Data efficiency aspect:
- Few-shot diffusion model.
- Generative model finetuning.
- Data efficiency for generative models and/or foundation models.
- Evaluation efficiency aspect:
- Efficient evaluation of generative models.
- Methods for benchmarking the efficiency of generative models.
- Challenges in evaluating the efficiency of generative models in diverse applications.
- Application-specific efficiency aspect:
- Applications of efficient generative AI in computer vision, such as image generation, style transfer, and super-resolution.
- Efficient personalization and post-generation editing.
- Real-time and efficient neural rendering.
- Efficient novel view synthesis.
- Interdisciplinary approaches that combine efficient generative AI with other fields, such as natural language processing, robotics, and human-computer interaction.
- Efficient generative models for video synthesis.
- Efficient interplay between image generation, video generation, and language generation such as efficient text2img, img2text, text2video, video2text, img2video, video2img, etc.
- Other related topics:
- Challenges and opportunities.
- Exploratory direction.
- New evaluation metrics.
- Privacy considerations.
Format: Submissions must use the CVPR 2024 Author Kit for Latex/Word Zip file and follow the CVPR 2024 author instructions and submission policies. Submissions need to be anonymized. The workshop considers two submission tracks:
- Long Paper: submissions are limited to 8 pages excluding references;
- Extended Abstract: submissions are limited to 4 pages excluding references.
Only long papers will be included in the CVPR 2024 proceedings.
Submission Site: https://cmt3.research.microsoft.com/CVPREDGE2024
Submission Deadline:
March 22, 2024 (AOE)
March 27, 2024 (AOE)
Workshop Date: June 17, 2024 (Full Day)
Workshop Location: Summit 432
Virtual Audience: https://cvpr.thecvf.com/virtual/2024/workshop/23675
Speakers

Georgios Kopanas
Inria

Thomas Leimkühler
Max-Planck-Institut für Informatik

Shanghang Zhang
Peking University

Song Han
Massachusetts Institute of Technology

Björn Ommer
Ludwig Maximilian University of Munich

Jiaming Song
Luma AI

Jian Ren
Snap Inc.

Tim Salimans

Cheng Lu
Tsinghua University

Qiang Liu
University of Texas at Austin

Ziwei Liu
Nanyang Technological University
Schedule
Each talk is assigned a 40 min slot (2 min intro + 30 min talk + 5 min QA + 3 min buffer)| Time | Activity | Title |
|---|---|---|
| 08:30 - 08:40 | Opening remarks | |
| 08:40 - 09:20 | Georgios Kapanas, Inria / Thomas Leimkühler, MPI | "3D Gaussian Splatting: An Introduction and Recent Advancements" |
| 09:20 - 10:00 | Shanghang Zhang, Peking University | "Towards Machine Learning Generalization in the Open World" |
| 10:00 - 10:10 | Morning Break | |
| 10:10 - 10:50 | Song Han, MIT | "Efficient Multi-modal LLM on the Edge" |
| 10:50 - 11:30 | Björn Ommer, LMU | "Beyond Diffusion: Efficient Models for Visual Synthesis" |
| 11:30 - 12:10 | Jiaming Song, Luma AI | "Dream Machine" |
| 12:10 - 13:40 | Poster Session (Level 4, 4E), Lunch Break | |
| 13:50 - 14:30 | Jian Ren, Snap Inc. | "Content Generation on Mobile Devices" |
| 14:30 - 15:10 | Tim Salimans, Google | "Two New Distillation Methods for Few-step Generation with Diffusion Models" |
| 15:10 - 15:50 | Cheng Lu, Tsinghua University | "DPM-Solver: Training-Free Fast Samplers for Diffusion Models" |
| 15:50 - 16:00 | Afternoon Break | |
| 16:00 - 16:40 | Qing Liu, UT Austin | "Rectified Flow and Straight Coupling: What is the Mathematical Essense of Fast Generation?" |
| 16:40 - 17:20 | Ziwei Liu, NTU | "Vchitect: Efficient and Scalable Video Generation" |
| 17:20 - 17:30 | Closing remarks | |
Accepted Papers
We are pleased to announce the accepted papers for the First Workshop on Efficient and On-Device Generation (EDGE) at CVPR 2024. Congratulations to all authors!
Long Papers
-
Title: LD-Pruner: Efficient Pruning of Latent Diffusion Models using Task-Agnostic Insights
Authors: Thibault Castells (Nota Inc.), Hyoung-Kyu Song (Nota Inc.), Bo-Kyeong Kim (Nota Inc.), Shinkook Choi (Nota Inc.)
[Poster # 36] -
Title: EdgeRelight360: Text-Conditioned 360-Degree HDR Image Generation for Real-Time On-Device Video Portrait Relighting
Authors: Min-Hui Lin (Qualcomm Technologies Inc.), Mahesh Kumar Krishna Reddy (Qualcomm Technologies Inc), Guillaume Berger (Qualcomm Technologies Inc.), Michel Sarkis (Qualcomm Technologies Inc.), Fatih Porikli (Qualcomm AI Research), Ning Bi (Qualcomm)
[Poster # 37]
Short Papers
-
Title: FoLD: Efficient Fourier-series-based Score Estimation for Langevin Diffusion
Authors: Siddarth Asokan (Microsoft Research Lab India), Aadithya Srikanth (Indian Institute of Science), Nishanth Shetty (Indian Institute of Science), Chandra Sekhar Seelamantula (IISc Bangalore)
[Poster # 38] -
Title: EdgeFusion: On-Device Text-to-Image Generation
Authors: Thibault Castells (Nota Inc.) Hyoung-Kyu Song (Nota Inc.), Tairen Piao (Nota Inc), Shinkook Choi (Nota Inc.), Bo-Kyeong Kim (Nota Inc.), Yim Hanyoung (Samsung Electronics.), Changgwun Lee (Samsung Electronics), Jae Gon Kim (Samsung Electronics), Tae-Ho Kim (Nota, Inc.)
[Poster # 39] -
Title: Observation-Guided Diffusion Probabilistic Models
Authors: Junoh Kang (Seoul National University), Jinyoung Choi (Seoul National University), Sungik Choi (LG AI Research), Bohyung Han (Seoul National University)
[Poster # 40] -
Title: Efficiently Quantize Latent Diffusion Models
Authors: Yuewei Yang (Meta Platforms Inc.), Jialiang Wang (Meta Platforms Inc), Xiaoliang Dai (Meta Platforms Inc.), Peizhao Zhang (Meta), Hongbo Zhang (Meta)
[Poster # 41] -
Title: Knowledge Distillation Cross Domain Diffusion Models for Weakly Supervised Defect Detection
Authors: Yuan-Fu Yang (National Yang Ming Chiao Tung University), Min Sun (NTHU)
[Poster # 42] -
Title: Bigger is not Always Better: Scaling Properties of Latent Diffusion Models
Authors: Kangfu Mei (Johns Hopkins University), Zhengzhong Tu (University of Texas at Austin), Mauricio Delbracio (Google Research), Hossein Talebi (Google Inc.), Vishal Patel (Johns Hopkins University), Peyman Milanfar (Google)
[Poster # 43] -
Title: ECLIPSE: A Resource-Efficient Text-to-Image Prior for Image Generations
Authors: Maitreya Patel (ASU), Changhoon Kim (Arizona State University), Sheng Cheng (Arizona State University), Chitta Baral (Arizona State University), Yezhou Yang (Arizona State University)
[Poster # 44]
Organizing Committee

Felix Juefei-Xu*
GenAI, Meta

Tingbo Hou*
GenAI, Meta

Licheng Yu
GenAI, Meta

Ruiqi Gao
Google DeepMind

Xiaoliang Dai
GenAI, Meta

Huiwen Chang
OpenAI

Bichen Wu
GenAI, Meta

Chenlin Meng
Stanford University

Ning Zhang
GenAI, Meta

Yanwu Xu
Boston University

Xi Yin
GenAI, Meta

Sirui Xie
UCLA

Camille Couprie
FAIR, Meta

Yunzhi Zhang
Stanford University

Andrew Brown
GenAI, Meta

Yang Zhao

Ali Thabet
GenAI, Meta

Zhisheng Xiao

Peizhao Zhang
GenAI, Meta

Peter Vajda
GenAI, Meta
Important Dates
-
Submission Deadline:
March 22, 2024 (AOE)March 27, 2024 (AOE) - Author Notification: April 5, 2024
- Camera-Ready Deadline: 11:59 PM PST, April 12, 2024 (Firm)
- Workshop Date: June 17, 2024 (Full Day)