Prototypical VoteNet for Few-Shot 3D Point Cloud Object Detection
Shizhen Zhao
Xiaojuan Qi
[Paper]
[GitHub]

Abstract

Most existing 3D point cloud object detection approaches heavily rely on large amounts of labeled training data. However, the labeling process is costly and time-consuming. This paper considers few-shot 3D point cloud object detection, where only a few annotated samples of novel classes are needed with abundant samples of base classes. To this end, we propose Prototypical VoteNet to recognize and localize novel instances, which incorporates two new modules: Prototypical Vote Module (PVM) and Prototypical Head Module (PHM). Specifically, as the 3D basic geometric structures can be shared among categories, PVM is designed to leverage class-agnostic geometric prototypes, which are learned from base classes, to refine local features of novel categories. Then PHM is proposed to utilize class prototypes to enhance the global feature of each object, facilitating subsequent object localization and classification, which is trained by the episodic training strategy. To evaluate the model in this new setting, we contribute two new benchmark datasets, FS-ScanNet and FS-SUNRGBD. We conduct extensive experiments to demonstrate the effectiveness of Prototypical VoteNet, and our proposed method shows significant and consistent improvements compared to baselines on two benchmark datasets.


Code


 [GitHub]


Paper and Supplementary Material

Zhao,Shizhen and Qi, Xiaojuan.
Prototypical VoteNet for Few-Shot 3D Point Cloud Object Detection.
In Advances in Neural Information Processing Systems, 2022.
(hosted on ArXiv)


[Bibtex]


Acknowledgements

This work has been supported by Hong Kong Research Grant Council - Early Career Scheme (Grant No. 27209621), HKU Startup Fund, and HKU Seed Fund for Basic Research. Part of the described research work is conducted in the JC STEM Lab of Robotics for Soft Materials funded by The Hong Kong Jockey Club Charities Trust.


This template was originally made by Phillip Isola and Richard Zhang for a colorful ECCV project; the code can be found here.