Real-time underwater stereofusion

Matija Rossi, Petar Trslić, Satja Sivčev, James Riordan, Daniel Toal, Gerard Dooly

Research output: Contribution to journalArticlepeer-review

Abstract

Many current and future applications of underwater robotics require real-time sensing and interpretation of the environment. As the vast majority of robots are equipped with cameras, computer vision is playing an increasingly important role it this field. This paper presents the implementation and experimental results of underwater StereoFusion, an algorithm for real-time 3D dense reconstruction and camera tracking. Unlike KinectFusion on which it is based, StereoFusion relies on a stereo camera as its main sensor. The algorithm uses the depth map obtained from the stereo camera to incrementally build a volumetric 3D model of the environment, while simultaneously using the model for camera tracking. It has been successfully tested both in a lake and in the ocean, using two different state-of-the-art underwater Remotely Operated Vehicles (ROVs). Ongoing work focuses on applying the same algorithm to acoustic sensors, and on the implementation of a vision based monocular system with the same capabilities.

Original languageEnglish
Article number3936
JournalSensors
Volume18
Issue number11
DOIs
Publication statusPublished - 2018

Keywords

  • 3D
  • Camera
  • Fusion
  • GPU
  • Real-time
  • ROV
  • Stereo
  • Tracking
  • Underwater
  • Vision

Fingerprint

Dive into the research topics of 'Real-time underwater stereofusion'. Together they form a unique fingerprint.

Cite this