Authors
Saining Xie, Jiatao Gu, Demi Guo, Charles R Qi, Leonidas Guibas, Or Litany
Publication date
2020/8/23
Conference
European Conference on Computer Vision
Pages
574-591
Publisher
Springer, Cham
Description
Arguably one of the top success stories of deep learning is transfer learning. The finding that pre-training a network on a rich source set (e.g., ImageNet) can help boost performance once fine-tuned on a usually much smaller target set, has been instrumental to many applications in language and vision. Yet, very little is known about its usefulness in 3D point cloud understanding. We see this as an opportunity considering the effort required for annotating data in 3D. In this work, we aim at facilitating research on 3D representation learning. Different from previous works, we focus on high-level scene understanding tasks. To this end, we select a suit of diverse datasets and tasks to measure the effect of unsupervised pre-training on a large source set of 3D scenes. Our findings are extremely encouraging: using a unified triplet of architecture, source dataset, and contrastive loss for pre-training, we achieve …
Total citations
20202021202220232024664133211144
Scholar articles
S Xie, J Gu, D Guo, CR Qi, L Guibas, O Litany - Computer Vision–ECCV 2020: 16th European …, 2020