PyTorch code for pre-training 3D point cloud transformers
Top 54.2% on sourcepulse
Point-BERT offers a novel pre-training approach for 3D point cloud Transformers, inspired by BERT's masked language modeling. It enables robust feature learning for downstream tasks like classification, few-shot learning, and part segmentation on point cloud data.
How It Works
Point-BERT employs a Masked Point Modeling (MPM) strategy. Input point clouds are divided into local patches, and a discrete Variational AutoEncoder (dVAE) tokenizes these patches into discrete tokens. During pre-training, random patches are masked, and the Transformer backbone learns to reconstruct the original tokens for these masked regions. This approach effectively captures local geometric information and semantic context within point clouds.
Quick Start & Requirements
pip install -r requirements.txt
and build custom PyTorch extensions for Chamfer Distance, PointNet++, and kNN using install.sh
.Highlighted Details
Maintenance & Community
The project is associated with CVPR 2022 and has contributions from multiple researchers at Tsinghua University. Community interaction channels are not explicitly mentioned in the README.
Licensing & Compatibility
Limitations & Caveats
The setup requires compiling custom PyTorch extensions, which can be complex and may encounter compatibility issues with different CUDA or PyTorch versions. The README indicates specific versions for Python and PyTorch, suggesting potential fragility with newer releases.
4 months ago
Inactive