3-D motion estimation by integrating visual cues in 2-D multi-modal opti-acoustic stereo sequences

Research output: Contribution to journalArticle

13 Citations (Scopus)

Abstract

Object reconstruction and target-based positioning are among critical capabilities in deploying submersible platforms for a range of underwater applications, e.g., search and inspection missions. Optical cameras provide high-resolution and target details, but their utility becomes constrained by the visibility range. In comparison, high-frequency (MHz) 2-D sonar imaging systems introduced to the commercial market in recent years can image targets at distances of tens of meters in highly turbid waters. Where fair visibility permits optical imaging at reasonable quality, the integration with 2-D sonar data can enable better performance compared to deploying either system alone, and thus enabling automated operation in a wider range of conditions. We investigate the estimation of 3-D motion by exploiting the visual cues in optical and sonar video for vision-based navigation and 3-D positioning of submersible platforms. The application of structure from motion paradigm in this multi-modal imaging scenario also enables the 3-D reconstruction of scene features. Our method does not require establishing multi-modal association between corresponding optical and sonar features, but rather the tracking of features in the sonar and optical motion sequences independently. In addition to improving the motion estimation accuracy, another advantage of the proposed method includes overcoming the inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity and dual interpretation of motion relative to planar scenes. We discuss how our solution can also provide an effective strategy to address the complex opti-acoustic stereo matching problem. Experiment with synthetic and real data demonstrate the advantages of our technical contribution.

Original languageEnglish
Pages (from-to)928-941
Number of pages14
JournalComputer Vision and Image Understanding
Volume114
Issue number8
DOIs
StatePublished - Aug 1 2010

Fingerprint

Sonar
Motion estimation
Acoustics
Visibility
Imaging techniques
Imaging systems
Navigation
Inspection
Cameras
Water
Experiments

Keywords

  • 3-D motion estimation
  • Multi-modal stereo imaging
  • Sonar and opical stereo imaging
  • Structure from Motion

ASJC Scopus subject areas

  • Software
  • Computer Vision and Pattern Recognition
  • Signal Processing

Cite this

3-D motion estimation by integrating visual cues in 2-D multi-modal opti-acoustic stereo sequences. / Negahdaripour, Shahriar; Taatian, Ali.

In: Computer Vision and Image Understanding, Vol. 114, No. 8, 01.08.2010, p. 928-941.

Research output: Contribution to journalArticle

@article{ec2d4c24b03f4227ada6b9058d5d27ca,
title = "3-D motion estimation by integrating visual cues in 2-D multi-modal opti-acoustic stereo sequences",
abstract = "Object reconstruction and target-based positioning are among critical capabilities in deploying submersible platforms for a range of underwater applications, e.g., search and inspection missions. Optical cameras provide high-resolution and target details, but their utility becomes constrained by the visibility range. In comparison, high-frequency (MHz) 2-D sonar imaging systems introduced to the commercial market in recent years can image targets at distances of tens of meters in highly turbid waters. Where fair visibility permits optical imaging at reasonable quality, the integration with 2-D sonar data can enable better performance compared to deploying either system alone, and thus enabling automated operation in a wider range of conditions. We investigate the estimation of 3-D motion by exploiting the visual cues in optical and sonar video for vision-based navigation and 3-D positioning of submersible platforms. The application of structure from motion paradigm in this multi-modal imaging scenario also enables the 3-D reconstruction of scene features. Our method does not require establishing multi-modal association between corresponding optical and sonar features, but rather the tracking of features in the sonar and optical motion sequences independently. In addition to improving the motion estimation accuracy, another advantage of the proposed method includes overcoming the inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity and dual interpretation of motion relative to planar scenes. We discuss how our solution can also provide an effective strategy to address the complex opti-acoustic stereo matching problem. Experiment with synthetic and real data demonstrate the advantages of our technical contribution.",
keywords = "3-D motion estimation, Multi-modal stereo imaging, Sonar and opical stereo imaging, Structure from Motion",
author = "Shahriar Negahdaripour and Ali Taatian",
year = "2010",
month = "8",
day = "1",
doi = "10.1016/j.cviu.2010.04.005",
language = "English",
volume = "114",
pages = "928--941",
journal = "Computer Vision and Image Understanding",
issn = "1077-3142",
publisher = "Academic Press Inc.",
number = "8",

}

TY - JOUR

T1 - 3-D motion estimation by integrating visual cues in 2-D multi-modal opti-acoustic stereo sequences

AU - Negahdaripour, Shahriar

AU - Taatian, Ali

PY - 2010/8/1

Y1 - 2010/8/1

N2 - Object reconstruction and target-based positioning are among critical capabilities in deploying submersible platforms for a range of underwater applications, e.g., search and inspection missions. Optical cameras provide high-resolution and target details, but their utility becomes constrained by the visibility range. In comparison, high-frequency (MHz) 2-D sonar imaging systems introduced to the commercial market in recent years can image targets at distances of tens of meters in highly turbid waters. Where fair visibility permits optical imaging at reasonable quality, the integration with 2-D sonar data can enable better performance compared to deploying either system alone, and thus enabling automated operation in a wider range of conditions. We investigate the estimation of 3-D motion by exploiting the visual cues in optical and sonar video for vision-based navigation and 3-D positioning of submersible platforms. The application of structure from motion paradigm in this multi-modal imaging scenario also enables the 3-D reconstruction of scene features. Our method does not require establishing multi-modal association between corresponding optical and sonar features, but rather the tracking of features in the sonar and optical motion sequences independently. In addition to improving the motion estimation accuracy, another advantage of the proposed method includes overcoming the inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity and dual interpretation of motion relative to planar scenes. We discuss how our solution can also provide an effective strategy to address the complex opti-acoustic stereo matching problem. Experiment with synthetic and real data demonstrate the advantages of our technical contribution.

AB - Object reconstruction and target-based positioning are among critical capabilities in deploying submersible platforms for a range of underwater applications, e.g., search and inspection missions. Optical cameras provide high-resolution and target details, but their utility becomes constrained by the visibility range. In comparison, high-frequency (MHz) 2-D sonar imaging systems introduced to the commercial market in recent years can image targets at distances of tens of meters in highly turbid waters. Where fair visibility permits optical imaging at reasonable quality, the integration with 2-D sonar data can enable better performance compared to deploying either system alone, and thus enabling automated operation in a wider range of conditions. We investigate the estimation of 3-D motion by exploiting the visual cues in optical and sonar video for vision-based navigation and 3-D positioning of submersible platforms. The application of structure from motion paradigm in this multi-modal imaging scenario also enables the 3-D reconstruction of scene features. Our method does not require establishing multi-modal association between corresponding optical and sonar features, but rather the tracking of features in the sonar and optical motion sequences independently. In addition to improving the motion estimation accuracy, another advantage of the proposed method includes overcoming the inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity and dual interpretation of motion relative to planar scenes. We discuss how our solution can also provide an effective strategy to address the complex opti-acoustic stereo matching problem. Experiment with synthetic and real data demonstrate the advantages of our technical contribution.

KW - 3-D motion estimation

KW - Multi-modal stereo imaging

KW - Sonar and opical stereo imaging

KW - Structure from Motion

UR - http://www.scopus.com/inward/record.url?scp=77953962662&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=77953962662&partnerID=8YFLogxK

U2 - 10.1016/j.cviu.2010.04.005

DO - 10.1016/j.cviu.2010.04.005

M3 - Article

AN - SCOPUS:77953962662

VL - 114

SP - 928

EP - 941

JO - Computer Vision and Image Understanding

JF - Computer Vision and Image Understanding

SN - 1077-3142

IS - 8

ER -