CadVLM: Bridging Language and Vision in the Generation of Parametric CAD Sketches
Parametric Computer-Aided Design (CAD) is central to contemporary mechanical design. However, it encounters challenges in achieving precise parametric sketch modeling and lacks practical evaluation metrics suitable for mechanical design. We harness the capabilities of pre-trained foundation models,...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Parametric Computer-Aided Design (CAD) is central to contemporary mechanical
design. However, it encounters challenges in achieving precise parametric
sketch modeling and lacks practical evaluation metrics suitable for mechanical
design. We harness the capabilities of pre-trained foundation models, renowned
for their successes in natural language processing and computer vision, to
develop generative models specifically for CAD. These models are adept at
understanding complex geometries and design reasoning, a crucial advancement in
CAD technology. In this paper, we propose CadVLM, an end-to-end vision language
model for CAD generation. Our approach involves adapting pre-trained foundation
models to manipulate engineering sketches effectively, integrating both sketch
primitive sequences and sketch images. Extensive experiments demonstrate
superior performance on multiple CAD sketch generation tasks such as CAD
autocompletion, CAD autoconstraint, and image conditional generation. To our
knowledge, this is the first instance of a multimodal Large Language Model
(LLM) being successfully applied to parametric CAD generation, representing a
pioneering step in the field of computer-aided mechanical design. |
---|---|
DOI: | 10.48550/arxiv.2409.17457 |