Anton Obukhov's Avatar

Anton Obukhov

@obukhov.ai

Research Scientist in Computer Vision and Generative AI

1,089
Followers
182
Following
42
Posts
22.11.2024
Joined
Posts Following

Latest posts by Anton Obukhov @obukhov.ai

Likely not too consistent, as most image based methods

16.12.2025 17:26 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

🌐 Web: hf.co/spaces/prs-e...
πŸ“• Paper: hf.co/papers/2512....
πŸ™ Code: github.com/prs-eth/ster...
πŸ€— Demo: hf.co/spaces/tosha...
πŸ€— Weights: hf.co/prs-eth/ster...

By ETHZ (Tjark, @bingxinke.bsky.social,
, Konrad), University of Bologna (Fabio, @mattpoggi.bsky.social), HUAWEI Bayer Lab (@obukhov.ai)

16.12.2025 16:49 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail

Introducing StereoSpace -- our new end-to-end method for turning photos into stereo images without explicit geometry or depth maps. This makes it especially robust with thin structures and transparencies. Try the demo below

16.12.2025 16:49 πŸ‘ 9 πŸ” 4 πŸ’¬ 2 πŸ“Œ 0

Other resources:
🌎 Website: huggingface.co/spaces/huawe...
πŸŽ“ Paper: arxiv.org/abs/2512.05000
πŸ€— Model: huggingface.co/huawei-bayer...
πŸ™ Code: github.com/huawei-bayer...

Team: Daniyar Zakarin*, Thiemo Wandel*, Anton Obukhov, Dengxin Dai.
*Work done during internships at HUAWEI Bayer Lab

11.12.2025 12:21 πŸ‘ 2 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail

Introducing WindowSeat - our new method for removing reflections from photos taken through windows, on planes, in malls, offices, and other glass-filled environments.

Try it with your own photos in this demo: huggingface.co/spaces/tosha...

11.12.2025 12:21 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 1

Original announcement of Marigold Depth (CVPR 2024 Oral, Best Paper Award Candidate): x.com/AntonObukhov...

15.05.2025 16:23 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Team: Bingxin Ke (@bingxinke.bsky.social), Kevin Qu (@kevinqu.bsky.social), Tianfu Wang ( ), Nando Metzger (@nandometzger.bsky.social), Shengyu Huang, Bo Li, Anton Obukhov (@obukhov.ai), Konrad Schindler.
We thank @hf.co for their sustained support.

15.05.2025 16:23 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

A lot of cutting-edge research in related areas comes from ETH ZΓΌrich’s PRS lab, led by Prof. Konrad Schindler. Follow the lab for updates:
πŸ™ github.com/prs-eth
πŸ€— huggingface.co/prs-eth
🌐 prs.igp.ethz.ch
πŸŽ“ Open ETHZ BSc/MSc projects:
prs.igp.ethz.ch/education/op...

15.05.2025 16:23 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

πŸ€— Demo (Depth): huggingface.co/spaces/prs-e...
πŸ€— Demo (Normals): huggingface.co/spaces/prs-e...
πŸ€— Demo (Albedo): huggingface.co/spaces/prs-e...
πŸ€— Models: huggingface.co/collections/...

15.05.2025 16:23 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

🌎 Website: marigoldcomputervision.github.io
🧨 Tutorial: huggingface.co/docs/diffuse...
πŸ“— Paper: arxiv.org/abs/2505.09358
πŸ™ Training code: github.com/prs-eth/Mari...

15.05.2025 16:23 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Post image

New modalities include surface normals and intrinsic decompositions like albedo, material properties (roughness, metallicity), and lighting decompositions. Marigold proves to be an efficient fine-tuning protocol that generalizes across image analysis tasks.

15.05.2025 16:23 πŸ‘ 6 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Video thumbnail

Big Marigold update!
Last year, we showed how to turn Stable Diffusion 2 into a SOTA depth estimator with a few synthetic samples and 2–3 days on just 1 GPU.
Today's release features:
🏎️ 1-step inference
πŸ”’ New modalities
🫣 High resolution
🧨 Diffusers support
πŸ•ΉοΈ New demos
πŸ§ΆπŸ‘‡

15.05.2025 16:23 πŸ‘ 44 πŸ” 8 πŸ’¬ 1 πŸ“Œ 3

🍸🍸The TRICKY25 challenge: "Monocular Depth from Images of Specular and Transparent Surfaces" is live! 🍸🍸 Hosted at the 3rd TRICKY workshop #ICCV2025, with exciting speakers! @obukhov.ai @taiyasaki.bsky.social

Site: sites.google.com/view/iccv25t...
Codalab: codalab.lisn.upsaclay.fr/competitions...

14.05.2025 09:18 πŸ‘ 1 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0

@mattpoggi.bsky.social will probably know!

01.05.2025 14:13 πŸ‘ 1 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

The workshop report is now released! arxiv.org/abs/2504.17787

25.04.2025 15:24 πŸ‘ 2 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Preview
Research Intern - Foundation Models for Computer Vision - Huawei Research Center ZΓΌrich If you are enthusiastic in shaping Huawei’s European Research Institute together with a multicultural team of leading researchers, this is the right opportunity for you!

Huawei Research Center ZΓΌrich is looking for a Research Scientist intern to work with me on advancing foundation models for computer vision, focusing on enhancing computational photography features in mobile phones. Λ™βœ§Λ–Β°πŸ“Έβ‹†ο½‘Λš

careers.huaweirc.ch/jobs/5702605...

23.03.2025 14:59 πŸ‘ 9 πŸ” 2 πŸ’¬ 0 πŸ“Œ 0

Look at them stripes! A principled super-resolution drop by colleagues from PRS-ETH! Interactive demo with gradio-dualvision down in the post

14.03.2025 14:31 πŸ‘ 7 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

bsky.app/profile/obuk...

28.02.2025 10:26 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Video thumbnail

RollingDepth rolls into Nashville for #CVPR2025! 🎸

28.02.2025 10:26 πŸ‘ 4 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Previous posts: bsky.app/profile/obuk...

04.02.2025 15:57 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

MDEC Challenge update! The 4th Monocular Depth Estimation Workshop at #CVPR2025 will be accepting submissions in two phases:
πŸš€ Dev phase: Feb 1 - Mar 1
🎯 Final phase: Mar 1 - Mar 21
Website: jspenmar.github.io/MDEC/
🌐 Codalab: codalab.lisn.upsaclay.fr/competitions...

Bring your best depth!

04.02.2025 15:57 πŸ‘ 7 πŸ” 4 πŸ’¬ 2 πŸ“Œ 0

Announcement: bsky.app/profile/obuk...

31.01.2025 19:23 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

Update about the 4th Monocular Depth Estimation Workshop at #CVPR2025:
πŸŽ‰ Website is LIVE: jspenmar.github.io/MDEC/
πŸŽ‰ Keynotes: Peter Wonka, Yiyi Liao, and Konrad Schindler
πŸŽ‰ Challenge updates: new prediction types, baselines & metrics

31.01.2025 19:23 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 2

Not finished of course, but it has been demonstrated that such massive compute might be excessive. Video makes sense, but top labs have already delivered astounding video generators with comparable resources.

25.01.2025 14:59 πŸ‘ 0 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

What's the next frontier after LLMs, that will demand nuclear-powered GPU clusters? No agents or AGI please

25.01.2025 14:11 πŸ‘ 2 πŸ” 0 πŸ’¬ 3 πŸ“Œ 0

Stay tuned for more updates and resources!

MDEC Team: Matteo Poggi (@mattpoggi.bsky.social), Fabio Tosi, Ripudaman Singh Arora, Anton Obukhov (@obukhov.ai), Jaime Spencer, Chris Russell (@cruss.bsky.social), Simon Hadfield, Richard Bowden.

21.12.2024 15:52 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Post image

The 4th Monocular Depth Estimation Challenge (MDEC) is coming to #CVPR2025, and I’m excited to join the org team! After 2024’s breakthroughs in monodepth driven by generative model advances in transformers and diffusion, this year's focus is on OOD generalization and evaluation.

21.12.2024 15:52 πŸ‘ 22 πŸ” 3 πŸ’¬ 1 πŸ“Œ 1

Monocular depth meets depth completionπŸš€ Check out our latest work where we modified Marigold to a zero-shot depth completion tool. Everything without retraining🌼 (This paper, for once, contains geese instead of catsπŸ˜„ keep an eye open)

19.12.2024 12:14 πŸ‘ 20 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0

Team ETH ZΓΌrich: Massimiliano Viola, Kevin Qu (@kevinqu.bsky.social), Nando Metzger (@nandometzger.bsky.social), Bingxin Ke (@bingxinke.bsky.social), Alexander Becker, Konrad Schindler, and Anton Obukhov (@obukhov.ai). We thank @hf.co and @gradio-hf.bsky.social for their continuous support.

19.12.2024 01:51 πŸ‘ 3 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Preview
Marigold Depth Completion - a Hugging Face Space by prs-eth Discover amazing ML apps made by the community

🌎 Website: marigolddepthcompletion.github.io
πŸ€— Demo: huggingface.co/spaces/prs-e...
πŸ“• Paper: arxiv.org/abs/2412.13389
πŸ‘Ύ Code: github.com/prs-eth/mari...

19.12.2024 01:51 πŸ‘ 4 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0