Horizontal Attention Based Generation Module for Unsupervised Domain Adaptive Stereo Matching

Research output: Contribution to journalArticlepeer-review

4 Scopus citations

Abstract

The emergence of convolutional neural networks (CNNs) has led to significant advancements in various computer vision tasks. Among them, stereo matching is one of the most popular research areas that enables the reconstruction of 3D information, which is difficult to obtain with only a monocular camera. However, CNNs have their limitations, particularly their susceptibility to domain shift. The CNN-based stereo matching networks suffered from performance degradation under domain changes. Moreover, obtaining a significant amount of real-world ground truth data is laborious and costly when compared to acquiring synthetic data. In this letter, we propose an end-to-end framework that utilizes image-to-image translation to overcome the domain gap in stereo matching. Specifically, we suggest a horizontal attentive generation (HAG) module that incorporates the epipolar constraints when generating target-stylized left-right views. By employing a horizontal attention mechanism during generation, our method can address the issues related to small receptive field by aggregating more information of each view without using the entire feature map. Therefore, our network can maintain consistencies between each view during image generation, making it more robust for different datasets.

Original languageEnglish
Pages (from-to)6779-6786
Number of pages8
JournalIEEE Robotics and Automation Letters
Volume8
Issue number10
DOIs
StatePublished - 1 Oct 2023

Bibliographical note

Publisher Copyright:
© 2016 IEEE.

Keywords

  • Deep learning for visual perception
  • computer vision for automation

Fingerprint

Dive into the research topics of 'Horizontal Attention Based Generation Module for Unsupervised Domain Adaptive Stereo Matching'. Together they form a unique fingerprint.

Cite this