Underwater intervention is an important capability in several marine domains, with numerous industrial, scientific, and defense applications. However, existing perception systems used during intervention operations rely on data from optical cameras, which limits capabilities in poor visibility or lighting conditions. Prior work has examined opti-acoustic fusion methods, which use sonar data to resolve the depth ambiguity of the camera data while using camera data to resolve the elevation angle ambiguity of the sonar data. However, existing methods cannot achieve dense 3D reconstructions in real-time, and few studies have reported results from applying these methods in a turbid environment. In this work, we propose the opti-acoustic fusion method Sonar-MASt3R, which uses MASt3R to extract dense correspondences from optical camera data in real-time and pairs it with geometric cues from an acoustic 3D reconstruction to ensure robustness in turbid conditions. Experimental results using data recorded from an "opti-acoustic eye-in-hand" configuration across turbidity values ranging from 0.5 to 12 NTU highlight this method's improved robustness to turbidity relative to baseline methods.
@inproceedings{phung2025oasis,
author = {Phung, Amy and Camilli, Richard},
title = {Sonar-MASt3R: Real-Time Opti-Acoustic Fusion in Turbid, Unstructured Environments},
booktitle = {2026 IEEE International Conference on Robotics and Automation (ICRA)},
year = {2026},
}