资源论文Aligning 3D Models to RGB-D Images of Cluttered Scenes

Aligning 3D Models to RGB-D Images of Cluttered Scenes

2019-12-19 | |  74 |   41 |   0

Abstract

The goal of this work is to represent objects in an RGB-D scene with corresponding 3D models from a library. We approach this problem by fifirst detecting and segmenting object instances in the scene and then using a convolutional neural network (CNN) to predict the pose of the object. This CNN is trained using pixel surface normals in images containing renderings of synthetic objects. When tested on real data, our method outperforms alternative algorithms trained on real data. We then use this coarse pose estimate along with the inferred pixel support to align a small number of prototypical models to the data, and place into the scene the model that fifits best. We observe a 48% relative improvement in performance at the task of 3D detection over the current state-of-the-art [34], while being an order of magnitude faster.

上一篇:Propagated Image Filtering

下一篇:Learning an Efficient Model of Hand Shape Variation from Depth Images

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...