Login Paper Search My Schedule Paper Index Help

My ICIP 2021 Schedule

Note: Your custom schedule will not be saved unless you create a new account or login to an existing account.
  1. Create a login based on your email (takes less than one minute)
  2. Perform 'Paper Search'
  3. Select papers that you desire to save in your personalized schedule
  4. Click on 'My Schedule' to see the current list of selected papers
  5. Click on 'Printable Version' to create a separate window suitable for printing (the header and menu will appear, but will not actually print)

Paper Detail

Paper IDSS-3DPU.7
Paper Title UNSUPERVISED LEARNING OF VISUAL ODOMETRY USING DIRECT MOTION MODELING
Authors Silviu Andrei, Illinois Institute of Technology / amazon.com, United States; Gady Agam, Illinois Institute of Technology, United States
SessionSS-3DPU: Special Session: 3D Visual Perception and Understanding
LocationArea B
Session Time:Tuesday, 21 September, 15:30 - 17:00
Presentation Time:Tuesday, 21 September, 15:30 - 17:00
Presentation Poster
Topic Applications of Machine Learning: Machine Learning for 3D Image and Video Processing
IEEE Xplore Open Preview  Click here to view in IEEE Xplore
Abstract Data for supervised learning of ego-motion and depth from video is scarce and expensive to produce. Subsequently, recent work has focused on unsupervised learning methods and achieved remarkable results. Many unsupervised approaches rely on single-view predicted depth and so ignore motion information. Some unsupervised methods incorporate motion information indirectly by designing the depth prediction network as an RNN. However, none of the existing methods make direct use of multiple frames when predicting depth, which are readily available in videos. In this work, we show that it is possible to achieve superior pose prediction results by modeling motion more directly. Our method uses a novel learning-based formulation for depth propagation and refinement which warps predicted depth maps forward from the current frame onto the next frame where it serves as a prior for predicting the next frame’s depth map. Code will be made available upon acceptance.