You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Generating CAD Code with Vision-Language Models for 3D Designs: |Paper|
Contribution
1. CADCodeVerify approach: a novel approach to iteratively verify and improve the design output of 3D objects generated from CAD code. 2. CADPrompt dataset: the first benchmark for CAD code generation, consisting of 200 natural language prompts paired with expert-annotated scripting code for 3D objects to benchmark progress.
1. CADCodeVerify approach
We introduce CADCodeVerify, a novel approach to iteratively verify and improve 3D objects generated from CAD code. Our approach works by producing ameliorative feedback by prompting a Vision-Language Model (VLM) to generate and answer a set of validation questions to verify the generated object and prompt the VLM to correct deviations.
2. CADPrompt dataset
We introduce a new benchmark, CADPrompt, featuring 200 3D objects represented in the Standard Triangle Language (STL) format. Each sample includes:
The 3D object in STL format sourced from DeepCAD [1].
A language-based description of the 3D object.
The 3D object in OBJ format, detailing vertices, faces, and edges in standard 3D geometry.
A .json file containing the CAD commands and their parameters.
The Python code that generates the 3D object, written by a CAD design expert.
Statistics of the CADPrompt dataset
The table below provides statistics for the CADPrompt dataset, including vertex and face counts of ground truth 3D objects, the lengths of language descriptions, and the corresponding Python code. The dataset is divided into simple and complex categories.
Examples from CADPrompt
Reference
[1] Wu, Rundi, Chang Xiao, and Changxi Zheng. "Deepcad: A deep generative network for computer-aided design models." Proceedings of the IEEE/CVF International Conference on Computer Vision. 2021.
Citation
If this dataset contributes to your work, please consider citing the following publications.
@article{alrashedy2024generating,
title={Generating CAD Code with Vision-Language Models for 3D Designs},
author={Alrashedy, Kamel and Tambwekar, Pradyumna and Zaidi, Zulfiqar and Langwasser, Megan and Xu, Wei and Gombolay, Matthew},
booktitle={International conference on learning representations (ICLR)},
year={2025}
}
@inproceedings{wu2021deepcad,
title={Deepcad: A deep generative network for computer-aided design models},
author={Wu, Rundi and Xiao, Chang and Zheng, Changxi},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={6772--6782},
year={2021}
}
About
[ICLR 2025] Generating CAD Code with Vision-Language Models for 3D Designs