Publication | Open Access
SelfVIO: Self-supervised deep monocular Visual–Inertial Odometry and depth estimation
95
Citations
82
References
2022
Year
Machine VisionMachine LearningImage AnalysisDepth Map EstimationInertial Measurement UnitEngineering3D Pose EstimationField Robotics3D VisionExtended RealityDepth EstimationOdometryDepth MapRobot LearningSensor FusionDeep LearningMulti-view GeometryComputer Vision
In the last decade, numerous supervised deep learning approaches have been proposed for visual-inertial odometry (VIO) and depth map estimation, which require large amounts of labelled data. To overcome the data limitation, self-supervised learning has emerged as a promising alternative that exploits constraints such as geometric and photometric consistency in the scene. In this study, we present a novel self-supervised deep learning-based VIO and depth map recovery approach (SelfVIO) using adversarial training and self-adaptive visual-inertial sensor fusion. SelfVIO learns the joint estimation of 6 degrees-of-freedom (6-DoF) ego-motion and a depth map of the scene from unlabelled monocular RGB image sequences and inertial measurement unit (IMU) readings. The proposed approach is able to perform VIO without requiring IMU intrinsic parameters and/or extrinsic calibration between IMU and the camera. We provide comprehensive quantitative and qualitative evaluations of the proposed framework and compare its performance with state-of-the-art VIO, VO, and visual simultaneous localization and mapping (VSLAM) approaches on the KITTI, EuRoC and Cityscapes datasets. Detailed comparisons prove that SelfVIO outperforms state-of-the-art VIO approaches in terms of pose estimation and depth recovery, making it a promising approach among existing methods in the literature.
| Year | Citations | |
|---|---|---|
Page 1
Page 1