Monocular Depth Estimation Mapping with Depth Anything V2

SLAM
C++
ROS
Computer Vision
3D Reconstruction
Author

Suraj Kiron Nair

Published

January 1, 2025

This is a demonstration of the MDE mapping pipeline on a dataset collected at the NYU Tandon campus at Brooklyn Commons. The pipeline is using Depth Anything V2 to generate context based disparity. This is used to generate the unscaled depth image(bottom left), which is then scaled using the sparse depth of the Realsense D455(top left). The Depth Anything V2 Vit-s model is optimized to run on Nvidia Jestson devices using Nvidia TensorRT to achieve 60fps inference. The resulting system is capable of accurately mapping outdoor environments in real time at 30hz. The system is using OpenVins for localization.

References:

Reactive Collision Avoidance for Safe Agile Navigation