InteractVLM (#CVPR2025) is a great collaboration MPI-IS, UvA and Inria.
Authors: @saidwivedi.in, @anticdimi.bsky.social, S. Tripathi, O. Taheri, C. Schmid, @michael-j-black.bsky.social and @dimtzionas.bsky.social.
Code & models available at: interactvlm.is.tue.mpg.de (10/10)
15.06.2025 12:23
π 3
π 0
π¬ 0
π 0
InteractVLM is the first method that infers 3D contacts on both humans and objects from in-the-wild images, and exploits these for 3D reconstruction via an optimization pipeline. In contrast, existing methods like PHOSA rely on handcrafted or heuristic-based contacts. (9/10)
15.06.2025 12:23
π 2
π 0
π¬ 1
π 0
With just 5% of DAMONβs 3D body contact annotations, InteractVLM surpasses the fully-supervised DECO baseline trained on 100% of 3D annotations. This is promising for minimizing reliance on costly 3D data by using foundational models. (8/10)
15.06.2025 12:23
π 1
π 0
π¬ 1
π 0
InteractVLM also shows strong outperformance on object affordance prediction on the PIAD dataset. Here affordance is defined as contact probabilities on the object. (7/10)
15.06.2025 12:23
π 2
π 0
π¬ 1
π 0
InteractVLM significantly outperforms prior work, both qualitatively and quantitatively, on in-the-wild 3D human (binary & semantic) contact prediction on the DAMON dataset. (6/10)
15.06.2025 12:23
π 1
π 0
π¬ 1
π 0
To bridge this 2D-to-3D gap, we propose "Render-Localize-Lift":
- Render: 3D human/object meshes into multiview 2D images.
- Localize: A Multiview Localization (MV-Loc) model, guided by VLM tokens, predicts 2D contact masks.
- Lift: 2D contact masks to 3D.
(5/10)
15.06.2025 12:23
π 1
π 1
π¬ 1
π 0
How can we infer 3D contact with limited 3D data? InteractVLM exploits foundational modelsβa VLM & localization model fine tuned to reason about contact. Given an image & prompt, the VLM outputs tokens for localization. But these models work in 2D, while contact is 3D. (4/10)
15.06.2025 12:23
π 1
π 1
π¬ 1
π 0
Furthermore, simple binary contact (touching βanyβ object) misses the rich semantics of real multi-object interactions. Thus, we introduce a novel task - "Semantic Human Contact" estimation: predicting contact points on a human related to a specified object. (3/10)
15.06.2025 12:23
π 2
π 0
π¬ 1
π 0
Precisely inferring where humans contact objects from an image is hard due to occlusion & depth ambiguity. Current datasets of images with 3D contact are small as theyβre costly & tedious to create (mocap/manual labeling), limiting performance of contact predictors. (2/10)
15.06.2025 12:23
π 1
π 0
π¬ 1
π 0
Why does 3D human-object reconstruction fail in the wild or get limited to a few object classes? A key missing piece is accurate 3D contact. InteractVLM (#CVPR2025) uses foundational models to infer contact on humans & objects, improving reconstruction from a single image. (1/10)
15.06.2025 12:23
π 5
π 2
π¬ 1
π 0
β¨ Happy to be recognised again as an Outstanding Reviewer for #CVPR2025!
12.05.2025 06:46
π 2
π 0
π¬ 0
π 0
Thanks to the workshop organizers: @yixinchen.bsky.social, Baoxiong Jia, @yaoyaofeng.bsky.social, @songyoupeng.bsky.social, Chuhang Zou, @saidwivedi.in, Yixin Zhu, Siyuan Huang! π
And the challenge organizers: Xiongkun Linghu, Tai Wang, Jingli Lin, Xiaojian Ma
03.04.2025 08:29
π 1
π 0
π¬ 0
π 0
π’ Excited to announce the 5th Workshop on 3D Scene Understanding for Vision, Graphics & Robotics at #CVPR2025! Weβll dive into multimodal 3D scene understanding & reasoning with amazing speakers and challenges.
@cvprconference.bsky.social
More Details: scene-understanding.com.
03.04.2025 08:29
π 3
π 2
π¬ 1
π 0
I've been using GitHub's Lists feature for over a year, and it's seriously underrated! β
It lets you assign labels to all your starred repos, making it super easy to find projects later based on specific fields or topics. No more endless scrolling!
Link to my list: github.com/saidwivedi?t...
12.03.2025 08:37
π 3
π 0
π¬ 0
π 0
Thanks for sharing :) @chrisoffner3d.bsky.social can you also please add me to the list? I work on 3D human avatar.
09.01.2025 00:56
π 3
π 0
π¬ 0
π 0
[M2L 2024] Transformers - Lucas Beyer
YouTube video by Mediterranean Machine Learning (M2L) summer school
One of the best tutorials for understanding Transformers!
π½οΈ Watch here: www.youtube.com/watch?v=bMXq...
Big thanks to @giffmana.ai for this excellent content! π
08.12.2024 09:58
π 55
π 8
π¬ 0
π 0
Would love to be in the list π
24.11.2024 09:18
π 1
π 0
π¬ 1
π 0
Writing a good scientific paper
For those who missed this post on the-network-that-is-not-to-be-named, I made public my "secrets" for writing a good CVPR paper (or any scientific paper). I've compiled these tips of many years. It's long but hopefully it helps people write better papers. perceiving-systems.blog/en/post/writ...
20.11.2024 10:18
π 260
π 64
π¬ 4
π 8