AI

Depth Pro: Sharp Monocular Metric Depth in Less Than a Second

1 Mins read

We present a foundation model for zero-shot metric monocular depth estimation. Our model, Depth Pro, synthesizes high-resolution depth maps with unparalleled sharpness and high-frequency details. The predictions are metric, with absolute scale, without relying on the availability of metadata such as camera intrinsics. And the model is fast, producing a 2.25-megapixel depth map in 0.3 seconds on a standard GPU. These characteristics are enabled by a number of technical contributions, including an efficient multi-scale vision transformer for dense prediction, a training protocol that combines real and synthetic datasets to achieve high metric accuracy alongside fine boundary tracing, dedicated evaluation metrics for boundary accuracy in estimated depth maps, and state-of-the-art focal length estimation from a single image. Extensive experiments analyze specific design choices and demonstrate that Depth Pro outperforms prior work along multiple dimensions.


Source link

Related posts
AI

This AI Paper Propsoes an AI Framework to Prevent Adversarial Attacks on Mobile Vehicle-to-Microgrid Services

3 Mins read
Mobile Vehicle-to-Microgrid (V2M) services enable electric vehicles to supply or store energy for localized power grids, enhancing grid stability and flexibility. AI…
AI

IoT-LLM: An AI Framework that Integrates IoT Sensor Data with LLMs to Enhance their Perception and Reasoning Abilities in the Physical World

2 Mins read
Widely growing sectors, like Healthcare, logistics, and smart cities, are interconnected on devices that require task reasoning capabilities in the Internet of…
AI

Meissonic: A Non-Autoregressive Mask Image Modeling Text-to-Image Synthesis Model that can Generate High-Resolution Images

3 Mins read
Large Language Models (LLMs) have demonstrated remarkable progress in natural language processing tasks, inspiring researchers to explore similar approaches for text-to-image synthesis….

 

 

Leave a Reply

Your email address will not be published. Required fields are marked *