I’m Zach one of the co-founders of Adam (https://www.adamcad.com). We're building AI-powered tools for CAD and 3D modeling [1].
We’ve recently been exploring a new way to bring GPT-style image editing directly into 3D model generation and are excited to showcase this in our web-app today. We’re calling it creative mode and are intrigued by the fun use cases this could create by making 3D generation more conversational!
For example you can put a prompt in such as “an elephant” then follow it up by “have it ride a skateboard” and it preserves the context, identity and maintains consistency with the previous model. We believe this lends itself better to an iterative design process when prototyping creative 3D assets or models for printing.
We’re offering everyone 10 free generations to start (ramping up soon!). Here’s a short video explaining how it works: https://www.loom.com/share/cf9ab91375374a4f93d6cc89619a043b
We’d also love you to try our parametric mode (free!) which uses LLMs to create a conversational interface for solid modeling as touched on in a recent HN thread [2]. We are leveraging the code generation capabilities of these models to generate OpenSCAD code (an open-source script based CAD) and are surfacing the variables as sliders the user can toggle to adjust their design. We hope this can give a glimpse into what it could be like to “vibe-CAD”. We will soon be releasing our results on Will Patrick's Text to CAD eval [3] and adding B-rep compatible export!
We’d love to hear what you think and where we should take this next :)
[1]https://x.com/zachdive/status/1882858765613228287
[2]https://news.ycombinator.com/item?id=43774990
[3]https://willpatrick.xyz/technology/2025/04/23/teaching-llms-...
dekuma•1d ago
zachdive•1d ago