Likely not too consistent, as most image based methods
Likely not too consistent, as most image based methods
π Web: hf.co/spaces/prs-e...
π Paper: hf.co/papers/2512....
π Code: github.com/prs-eth/ster...
π€ Demo: hf.co/spaces/tosha...
π€ Weights: hf.co/prs-eth/ster...
By ETHZ (Tjark, @bingxinke.bsky.social,
, Konrad), University of Bologna (Fabio, @mattpoggi.bsky.social), HUAWEI Bayer Lab (@obukhov.ai)
Introducing StereoSpace -- our new end-to-end method for turning photos into stereo images without explicit geometry or depth maps. This makes it especially robust with thin structures and transparencies. Try the demo below
Other resources:
π Website: huggingface.co/spaces/huawe...
π Paper: arxiv.org/abs/2512.05000
π€ Model: huggingface.co/huawei-bayer...
π Code: github.com/huawei-bayer...
Team: Daniyar Zakarin*, Thiemo Wandel*, Anton Obukhov, Dengxin Dai.
*Work done during internships at HUAWEI Bayer Lab
Introducing WindowSeat - our new method for removing reflections from photos taken through windows, on planes, in malls, offices, and other glass-filled environments.
Try it with your own photos in this demo: huggingface.co/spaces/tosha...
Original announcement of Marigold Depth (CVPR 2024 Oral, Best Paper Award Candidate): x.com/AntonObukhov...
Team: Bingxin Ke (@bingxinke.bsky.social), Kevin Qu (@kevinqu.bsky.social), Tianfu Wang ( ), Nando Metzger (@nandometzger.bsky.social), Shengyu Huang, Bo Li, Anton Obukhov (@obukhov.ai), Konrad Schindler.
We thank @hf.co for their sustained support.
A lot of cutting-edge research in related areas comes from ETH ZΓΌrichβs PRS lab, led by Prof. Konrad Schindler. Follow the lab for updates:
π github.com/prs-eth
π€ huggingface.co/prs-eth
π prs.igp.ethz.ch
π Open ETHZ BSc/MSc projects:
prs.igp.ethz.ch/education/op...
π€ Demo (Depth): huggingface.co/spaces/prs-e...
π€ Demo (Normals): huggingface.co/spaces/prs-e...
π€ Demo (Albedo): huggingface.co/spaces/prs-e...
π€ Models: huggingface.co/collections/...
π Website: marigoldcomputervision.github.io
𧨠Tutorial: huggingface.co/docs/diffuse...
π Paper: arxiv.org/abs/2505.09358
π Training code: github.com/prs-eth/Mari...
New modalities include surface normals and intrinsic decompositions like albedo, material properties (roughness, metallicity), and lighting decompositions. Marigold proves to be an efficient fine-tuning protocol that generalizes across image analysis tasks.
Big Marigold update!
Last year, we showed how to turn Stable Diffusion 2 into a SOTA depth estimator with a few synthetic samples and 2β3 days on just 1 GPU.
Today's release features:
ποΈ 1-step inference
π’ New modalities
π«£ High resolution
𧨠Diffusers support
πΉοΈ New demos
π§Άπ
πΈπΈThe TRICKY25 challenge: "Monocular Depth from Images of Specular and Transparent Surfaces" is live! πΈπΈ Hosted at the 3rd TRICKY workshop #ICCV2025, with exciting speakers! @obukhov.ai @taiyasaki.bsky.social
Site: sites.google.com/view/iccv25t...
Codalab: codalab.lisn.upsaclay.fr/competitions...
@mattpoggi.bsky.social will probably know!
The workshop report is now released! arxiv.org/abs/2504.17787
Huawei Research Center ZΓΌrich is looking for a Research Scientist intern to work with me on advancing foundation models for computer vision, focusing on enhancing computational photography features in mobile phones. Λβ§ΛΒ°πΈβqΛ
careers.huaweirc.ch/jobs/5702605...
Look at them stripes! A principled super-resolution drop by colleagues from PRS-ETH! Interactive demo with gradio-dualvision down in the post
bsky.app/profile/obuk...
RollingDepth rolls into Nashville for #CVPR2025! πΈ
Previous posts: bsky.app/profile/obuk...
MDEC Challenge update! The 4th Monocular Depth Estimation Workshop at #CVPR2025 will be accepting submissions in two phases:
π Dev phase: Feb 1 - Mar 1
π― Final phase: Mar 1 - Mar 21
Website: jspenmar.github.io/MDEC/
π Codalab: codalab.lisn.upsaclay.fr/competitions...
Bring your best depth!
Announcement: bsky.app/profile/obuk...
Update about the 4th Monocular Depth Estimation Workshop at #CVPR2025:
π Website is LIVE: jspenmar.github.io/MDEC/
π Keynotes: Peter Wonka, Yiyi Liao, and Konrad Schindler
π Challenge updates: new prediction types, baselines & metrics
Not finished of course, but it has been demonstrated that such massive compute might be excessive. Video makes sense, but top labs have already delivered astounding video generators with comparable resources.
What's the next frontier after LLMs, that will demand nuclear-powered GPU clusters? No agents or AGI please
Stay tuned for more updates and resources!
MDEC Team: Matteo Poggi (@mattpoggi.bsky.social), Fabio Tosi, Ripudaman Singh Arora, Anton Obukhov (@obukhov.ai), Jaime Spencer, Chris Russell (@cruss.bsky.social), Simon Hadfield, Richard Bowden.
The 4th Monocular Depth Estimation Challenge (MDEC) is coming to #CVPR2025, and Iβm excited to join the org team! After 2024βs breakthroughs in monodepth driven by generative model advances in transformers and diffusion, this year's focus is on OOD generalization and evaluation.
Monocular depth meets depth completionπ Check out our latest work where we modified Marigold to a zero-shot depth completion tool. Everything without retrainingπΌ (This paper, for once, contains geese instead of catsπ keep an eye open)
Team ETH ZΓΌrich: Massimiliano Viola, Kevin Qu (@kevinqu.bsky.social), Nando Metzger (@nandometzger.bsky.social), Bingxin Ke (@bingxinke.bsky.social), Alexander Becker, Konrad Schindler, and Anton Obukhov (@obukhov.ai). We thank @hf.co and @gradio-hf.bsky.social for their continuous support.
π Website: marigolddepthcompletion.github.io
π€ Demo: huggingface.co/spaces/prs-e...
π Paper: arxiv.org/abs/2412.13389
πΎ Code: github.com/prs-eth/mari...