| CARVIEW |
Select Language
Comparisons with Baselines for Given Body Shapes ↑ back to top
We showcase comparative results here among all of the baselines, including T2M-GPT [1], MotionDiffuse [2], MotionGPT [3], MLD [4], and MDM [5]. We show the ground truth motion and body shape in the first column for reference. This page contains 30 samples categorized by different actions. Selecting a particular Action, e.g., 'Run', shows a list of sample videos corresponding to the selected action type. Each individual video in this list can then be selected and viewed.
Actions
Videos
Ground Truth
Ours
MotionDiffuse
.
T2M-GPT
Ground Truth Body Shape
MotionGPT
MLD
MDM
.
References
[1] Jianrong Zhang, Yangsong Zhang, Xiaodong Cun, Shaoli Huang, Yong Zhang, Hongwei Zhao, Hongtao Lu, and Xi Shen. T2m-gpt: Generating human motion from textual descriptions with discrete representations. In CVPR, 2023.
[2] Mingyuan Zhang, Zhongang Cai, Liang Pan, Fangzhou Hong, Xinying Guo, Lei Yang, and Ziwei Liu. Motiondiffuse: Text-driven human motion generation with diffusion model. arXiv preprint, 2022.
[3] Biao Jiang, Xin Chen, Wen Liu, Jingyi Yu, Gang Yu, and Tao Chen. Motiongpt: Human motion as a foreign language. NIPS, 2024.
[4] Xin Chen, Biao Jiang, Wen Liu, Zilong Huang, Bin Fu, Tao Chen, and Gang Yu. Executing your commands via motion diffusion in latent space. In CVPR, 2023.
[5] Guy Tevet, Sigal Raab, Brian Gordon, Yoni Shafir, Daniel Cohen-or, and Amit Haim Bermano. Human motion diffusion model. In ICLR, 2023.
BibTeX
@inproceedings{shapemove,
author = {Liao, Ting-Hsuan and Zhou, Yi and Shen, Yu and Huang, Chun-Hao Paul and Mitra, Saayan and Huang, Jia-Bin and Bhattacharya, Uttaran},
title = {Shape My Moves: Text-Driven Shape-Aware Synthesis of Human Motions},
journal = {Proceedings of the Computer Vision and Pattern Recognition Conference},
year = {2025},
}