PRIN/SPRIN: On Extracting Point-Wise Rotation Invariant Features

  • Yang You
  • , Yujing Lou
  • , Ruoxi Shi
  • , Qi Liu
  • , Yu Wing Tai
  • , Lizhuang Ma
  • , Weiming Wang*
  • , Cewu Lu*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

20 Scopus citations

Abstract

Point cloud analysis without pose priors is very challenging in real applications, as the orientations of point clouds are often unknown. In this paper, we propose a brand new point-set learning framework PRIN, namely, Point-wise Rotation Invariant Network, focusing on rotation invariant feature extraction in point clouds analysis. We construct spherical signals by Density Aware Adaptive Sampling to deal with distorted point distributions in spherical space. Spherical Voxel Convolution and Point Re-sampling are proposed to extract rotation invariant features for each point. In addition, we extend PRIN to a sparse version called SPRIN, which directly operates on sparse point clouds. Both PRIN and SPRIN can be applied to tasks ranging from object classification, part segmentation, to 3D feature matching and label alignment. Results show that, on the dataset with randomly rotated point clouds, SPRIN demonstrates better performance than state-of-the-art methods without any data augmentation. We also provide thorough theoretical proof and analysis for point-wise rotation invariance achieved by our methods. The code to reproduce our results will be made publicly available.

Original languageEnglish
Pages (from-to)9489-9502
Number of pages14
JournalIEEE Transactions on Pattern Analysis and Machine Intelligence
Volume44
Issue number12
DOIs
StatePublished - 1 Dec 2022
Externally publishedYes

Keywords

  • Point cloud
  • feature learning
  • object analysis
  • rotation invariance

Fingerprint

Dive into the research topics of 'PRIN/SPRIN: On Extracting Point-Wise Rotation Invariant Features'. Together they form a unique fingerprint.

Cite this