Likely not too consistent, as most image based methods
Posts by Anton Obukhov
๐ Web: hf.co/spaces/prs-e...
๐ Paper: hf.co/papers/2512....
๐ Code: github.com/prs-eth/ster...
๐ค Demo: hf.co/spaces/tosha...
๐ค Weights: hf.co/prs-eth/ster...
By ETHZ (Tjark, @bingxinke.bsky.social,
, Konrad), University of Bologna (Fabio, @mattpoggi.bsky.social), HUAWEI Bayer Lab (@obukhov.ai)
Introducing StereoSpace -- our new end-to-end method for turning photos into stereo images without explicit geometry or depth maps. This makes it especially robust with thin structures and transparencies. Try the demo below
Other resources:
๐ Website: huggingface.co/spaces/huawe...
๐ Paper: arxiv.org/abs/2512.05000
๐ค Model: huggingface.co/huawei-bayer...
๐ Code: github.com/huawei-bayer...
Team: Daniyar Zakarin*, Thiemo Wandel*, Anton Obukhov, Dengxin Dai.
*Work done during internships at HUAWEI Bayer Lab
Introducing WindowSeat - our new method for removing reflections from photos taken through windows, on planes, in malls, offices, and other glass-filled environments.
Try it with your own photos in this demo: huggingface.co/spaces/tosha...
Original announcement of Marigold Depth (CVPR 2024 Oral, Best Paper Award Candidate): x.com/AntonObukhov...
Team: Bingxin Ke (@bingxinke.bsky.social), Kevin Qu (@kevinqu.bsky.social), Tianfu Wang ( ), Nando Metzger (@nandometzger.bsky.social), Shengyu Huang, Bo Li, Anton Obukhov (@obukhov.ai), Konrad Schindler.
We thank @hf.co for their sustained support.
A lot of cutting-edge research in related areas comes from ETH Zรผrichโs PRS lab, led by Prof. Konrad Schindler. Follow the lab for updates:
๐ github.com/prs-eth
๐ค huggingface.co/prs-eth
๐ prs.igp.ethz.ch
๐ Open ETHZ BSc/MSc projects:
prs.igp.ethz.ch/education/op...
๐ค Demo (Depth): huggingface.co/spaces/prs-e...
๐ค Demo (Normals): huggingface.co/spaces/prs-e...
๐ค Demo (Albedo): huggingface.co/spaces/prs-e...
๐ค Models: huggingface.co/collections/...
๐ Website: marigoldcomputervision.github.io
๐งจ Tutorial: huggingface.co/docs/diffuse...
๐ Paper: arxiv.org/abs/2505.09358
๐ Training code: github.com/prs-eth/Mari...
New modalities include surface normals and intrinsic decompositions like albedo, material properties (roughness, metallicity), and lighting decompositions. Marigold proves to be an efficient fine-tuning protocol that generalizes across image analysis tasks.
Big Marigold update!
Last year, we showed how to turn Stable Diffusion 2 into a SOTA depth estimator with a few synthetic samples and 2โ3 days on just 1 GPU.
Today's release features:
๐๏ธ 1-step inference
๐ข New modalities
๐ซฃ High resolution
๐งจ Diffusers support
๐น๏ธ New demos
๐งถ๐
๐ธ๐ธThe TRICKY25 challenge: "Monocular Depth from Images of Specular and Transparent Surfaces" is live! ๐ธ๐ธ Hosted at the 3rd TRICKY workshop #ICCV2025, with exciting speakers! @obukhov.ai @taiyasaki.bsky.social
Site: sites.google.com/view/iccv25t...
Codalab: codalab.lisn.upsaclay.fr/competitions...
@mattpoggi.bsky.social will probably know!
The workshop report is now released! arxiv.org/abs/2504.17787
Huawei Research Center Zรผrich is looking for a Research Scientist intern to work with me on advancing foundation models for computer vision, focusing on enhancing computational photography features in mobile phones. หโงหยฐ๐ธโ๏ฝกห
careers.huaweirc.ch/jobs/5702605...
Look at them stripes! A principled super-resolution drop by colleagues from PRS-ETH! Interactive demo with gradio-dualvision down in the post
bsky.app/profile/obuk...
RollingDepth rolls into Nashville for #CVPR2025! ๐ธ
Previous posts: bsky.app/profile/obuk...
MDEC Challenge update! The 4th Monocular Depth Estimation Workshop at #CVPR2025 will be accepting submissions in two phases:
๐ Dev phase: Feb 1 - Mar 1
๐ฏ Final phase: Mar 1 - Mar 21
Website: jspenmar.github.io/MDEC/
๐ Codalab: codalab.lisn.upsaclay.fr/competitions...
Bring your best depth!
Announcement: bsky.app/profile/obuk...
Update about the 4th Monocular Depth Estimation Workshop at #CVPR2025:
๐ Website is LIVE: jspenmar.github.io/MDEC/
๐ Keynotes: Peter Wonka, Yiyi Liao, and Konrad Schindler
๐ Challenge updates: new prediction types, baselines & metrics
Not finished of course, but it has been demonstrated that such massive compute might be excessive. Video makes sense, but top labs have already delivered astounding video generators with comparable resources.
What's the next frontier after LLMs, that will demand nuclear-powered GPU clusters? No agents or AGI please
Stay tuned for more updates and resources!
MDEC Team: Matteo Poggi (@mattpoggi.bsky.social), Fabio Tosi, Ripudaman Singh Arora, Anton Obukhov (@obukhov.ai), Jaime Spencer, Chris Russell (@cruss.bsky.social), Simon Hadfield, Richard Bowden.
The 4th Monocular Depth Estimation Challenge (MDEC) is coming to #CVPR2025, and Iโm excited to join the org team! After 2024โs breakthroughs in monodepth driven by generative model advances in transformers and diffusion, this year's focus is on OOD generalization and evaluation.
Monocular depth meets depth completion๐ Check out our latest work where we modified Marigold to a zero-shot depth completion tool. Everything without retraining๐ผ (This paper, for once, contains geese instead of cats๐ keep an eye open)
Team ETH Zรผrich: Massimiliano Viola, Kevin Qu (@kevinqu.bsky.social), Nando Metzger (@nandometzger.bsky.social), Bingxin Ke (@bingxinke.bsky.social), Alexander Becker, Konrad Schindler, and Anton Obukhov (@obukhov.ai). We thank @hf.co and @gradio-hf.bsky.social for their continuous support.