
Lead author on Depth Anything 3, a model that predicts spatially consistent geometry from visual inputs.
How media typically covers Haotong Lin
Referenced in coverage
ByteDance released Depth Anything 3, a unified transformer-based model that predicts spatially consistent geometry from arbitrary visual inputs for monocular depth estimation, multi-view depth estimation, pose estimation, and 3D Gaussian generation using a single depth-ray representation.
“Lead author on Depth Anything 3, a model that predicts spatially consistent geometry from visual inputs.”