HyperAI
HyperAI
Home
Console
Docs
News
Papers
Tutorials
Datasets
Wiki
SOTA
LLM Models
GPU Leaderboard
Events
Search
About
Terms of Service
Privacy Policy
English
HyperAI
HyperAI
Toggle Sidebar
Search the site…
⌘
K
Command Palette
Search for a command to run...
Console
Home
SOTA
Few-Shot 3D Point Cloud Classification
Few Shot 3D Point Cloud Classification On 4
Few Shot 3D Point Cloud Classification On 4
Metrics
Overall Accuracy
Standard Deviation
Results
Performance results of various models on this benchmark
Columns
Model Name
Overall Accuracy
Standard Deviation
Paper Title
ReCon++
96.5
3.0
ShapeLLM: Universal 3D Object Understanding for Embodied Interaction
Point-JEPA
96.4
2.7
Point-JEPA: A Joint Embedding Predictive Architecture for Self-Supervised Learning on Point Cloud
3D-JEPA
96.3
2.4
3D-JEPA: A Joint Embedding Predictive Architecture for 3D Self-Supervised Representation Learning
PointGPT
96.1
2.8
PointGPT: Auto-regressively Generative Pre-training from Point Clouds
PCP-MAE
95.9
2.7
PCP-MAE: Learning to Predict Centers for Point Masked Autoencoders
Point-RAE
95.8
3.0
Regress Before Construct: Regress Autoencoder for Point Cloud Self-supervised Learning
Point-FEMAE
95.8
-
Towards Compact 3D Representations via Point Feature Enhancement Masked Autoencoders
ReCon
95.8
3.0
Contrast with Reconstruct: Contrastive 3D Representation Learning Guided by Generative Pretraining
point2vec
95.8
3.1
Point2Vec for Self-Supervised Representation Learning on Point Clouds
ACT
95.6
2.8
Autoencoders as Cross-Modal Teachers: Can Pretrained 2D Image Transformers Help 3D Representation Learning?
OTMae3D
95.6
2.6
-
I2P-MAE
95.5
3.0
Learning 3D Representations from 2D Pre-trained Models via Image-to-Point Masked Autoencoders
IDPT
95.4
-
Instance-aware Dynamic Prompt Tuning for Pre-trained Point Cloud Models
Point-LGMask
95.1
3.4
Point-LGMask: Local and Global Contexts Embedding for Point Cloud Pre-training with Multi-Ratio Masking
Point-MAE
95.0
3.0
Masked Autoencoders for Point Cloud Self-supervised Learning
Point-M2AE
95.0
3.0
Point-M2AE: Multi-scale Masked Autoencoders for Hierarchical Point Cloud Pre-training
MaskPoint
93.4
3.5
Masked Discrimination for Self-Supervised Learning on Point Clouds
Point-BERT
92.7
5.1
Point-BERT: Pre-training 3D Point Cloud Transformers with Masked Point Modeling
CrossMoCo
91.0
3.4
CrossMoCo: Multi-modal Momentum Contrastive Learning for Point Cloud
OcCo+PointNet
89.7
1.5
Unsupervised Point Cloud Pre-Training via Occlusion Completion
0 of 31 row(s) selected.
Previous
Next