AI

Depth Pro: Sharp Monocular Metric Depth in Less Than a Second

1 Mins read

We present a foundation model for zero-shot metric monocular depth estimation. Our model, Depth Pro, synthesizes high-resolution depth maps with unparalleled sharpness and high-frequency details. The predictions are metric, with absolute scale, without relying on the availability of metadata such as camera intrinsics. And the model is fast, producing a 2.25-megapixel depth map in 0.3 seconds on a standard GPU. These characteristics are enabled by a number of technical contributions, including an efficient multi-scale vision transformer for dense prediction, a training protocol that combines real and synthetic datasets to achieve high metric accuracy alongside fine boundary tracing, dedicated evaluation metrics for boundary accuracy in estimated depth maps, and state-of-the-art focal length estimation from a single image. Extensive experiments analyze specific design choices and demonstrate that Depth Pro outperforms prior work along multiple dimensions.


Source link

Related posts
AI

Google AI Researchers Propose 'MODEL SWARMS': A Collaborative Search Algorithm to Flexibly Adapt Diverse LLM Experts to Wide-Ranging Purposes

2 Mins read
There is a need for flexible and efficient adaptation of large language models (LLMs) to various tasks. Existing approaches, such as mixture-of-experts…
AI

From ONNX to Static Embeddings: What Makes Sentence Transformers v3.2.0 a Game-Changer?

2 Mins read
There is a growing demand for embedding models that balance accuracy, efficiency, and versatility. Existing models often struggle to achieve this balance,…
AI

CodeJudge: An Machine Learning Framework that Leverages LLMs to Evaluate Code Generation Without the Need for Test Cases

2 Mins read
Artificial Intelligence is evolving significantly, and Large Language Models have shown a remarkable capacity to comprehend human-text inputs. Going beyond simple text…

 

 

Leave a Reply

Your email address will not be published. Required fields are marked *