Excited to share our #ICCV2025 work Reusing Computation in Text-to-Image Diffusion for Efficient Generation of Image Sets!
Our method generates large sets of images using significantly less compute than standard diffusion.
πhttps://ddecatur.github.io/hierarchical-diffusion/
1/
22.10.2025 20:22
π 8
π 2
π¬ 1
π 2
This work was led by Sining Lu and Guan Chen, in collaboration with Nam Anh Dinh, Itai Lang, Ari Holtzman, and me.
Check out our paper: arxiv.org/abs/2508.08228
Weβre still actively developing LL3M, and weβd love to hear your thoughts! 7/
15.08.2025 04:15
π 1
π 1
π¬ 0
π 0
Another cool thing about LL3M: the Blender code it writes is actually readable. Clear structure, detailed comments, intuitive variable names. Easy to tweak a single parameter (e.g. key width) or even change the algorithmic logic (e.g. the keyboard pattern). 6/
15.08.2025 04:15
π 0
π 0
π¬ 1
π 0
LL3M generates 3D assets in 3 phases with specialized agents:
1οΈβ£ Initial Creation β break prompt into subtasks, retrieve relevant code snippets (BlenderRAG)
2οΈβ£ Auto-refine β critic spots issues, verification checks fixes
3οΈβ£ User-guided β iterative edits via user feedback
5/
15.08.2025 04:15
π 0
π 0
π¬ 1
π 0
LL3M can create a wide range of shapes, without requiring specialized 3D datasets or fine-tuning. Every asset created is represented under the hood as editable Blender code. 4/
15.08.2025 04:15
π 0
π 0
π¬ 1
π 0
Even non-experts can jump right in and easily edit 3D shapes. Blender code created by LL3M generates a node graph that is packed with tunable parameters, enabling users to tweak colors, textures, patterns, lengths, heights, and more. 3/
15.08.2025 04:15
π 0
π 0
π¬ 1
π 0
What we β€οΈ about LL3M: You're in the loop! If you want to make a tweak, LL3M can be your collaborative 3D design partner. And there's no need to regenerate the entire model each time - target a specific part, provide follow-up prompts, and the rest stays intact. 2/
15.08.2025 04:15
π 0
π 0
π¬ 1
π 0
Weβve been building something weβre ππππππ¦ excited about β LL3M: LLM-powered agents that turn text into editable 3D assets. LL3M models shapes as interpretable Blender code, making geometry, appearance, and style easy to modify. π threedle.github.io/ll3m 1/
15.08.2025 04:15
π 2
π 0
π¬ 1
π 0
Our work βGeometry in Styleβ will be presented at #CVPR2025 on Sunday at 4pm in ExHall D, poster 219. Drop by and say hi!
Our technique is capable of performing expressive text-driven deformations that preserve the input shape identity.
1/
15.06.2025 18:33
π 22
π 4
π¬ 1
π 0
Big update from The Workshop on Computer Vision For Mixed Reality @ CVPR 2025:
π Papers and schedule are now up!
Our Speakers:
Richard Newcombe (Meta)
Anjul Patney (NVIDIA)
Rana Hanocka (UChicago)
Laura Leal-Taixe βͺ(NVIDIA)
Margarita Grinvald (Meta)
π
June 11, 8AM πRoom 109.
cv4mr.github.io
03.06.2025 15:50
π 15
π 3
π¬ 2
π 4
Congrats Brian and team on the best paper honorable mention at #3dv2025 π₯³π
Brian (hywkim-brian.github.io/site/) is going to start his PhD next year at Columbia with @silviasellan.bsky.social Make sure to watch out for their awesome works πππ
26.03.2025 03:53
π 15
π 3
π¬ 1
π 0
Tweet advertising that Rana Hanocka at UChicago CS is recruiting PhD students.
My colleague Rana Hanocka does exciting work on the frontier for 3D graphics and AI/ML, and sheβs recruiting PhD students this cycle!
11.12.2024 06:18
π 6
π 1
π¬ 1
π 0