HyperAI

Learning With Noisy Labels On Cifar 10N Worst

Metrics

Accuracy (mean)

Results

Performance results of various models on this benchmark

Model Name
Accuracy (mean)
Paper TitleRepository
Co-Teaching83.83Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels
ELR83.58Early-Learning Regularization Prevents Memorization of Noisy Labels
GNL86.99Partial Label Supervision for Agnostic Generative Noisy Label Learning
T-Revision80.48Are Anchor Points Really Indispensable in Label-Noise Learning?
JoCoR83.37Combating noisy labels by agreement: A joint training method with co-regularization
PSSCL95.12PSSCL: A progressive sample selection framework with contrastive loss designed for noisy labels
PGDF93.65Sample Prior Guided Robust Model Learning to Suppress Noisy Labels
CE77.69--
Negative-LS82.99Understanding Generalized Label Smoothing when Learning with Noisy Labels-
Forward-T79.79Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach
Co-Teaching+83.26How does Disagreement Help Generalization against Label Corruption?
F-div82.53When Optimizing $f$-divergence is Robust with Label Noise
ProMix96.16ProMix: Combating Label Noise via Maximizing Clean Sample Utility
GCE80.66Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels
Divide-Mix92.56DivideMix: Learning with Noisy Labels as Semi-supervised Learning
CAL85.36Clusterability as an Alternative to Anchor Points When Learning with Noisy Labels
VolMinNet80.53Provably End-to-end Label-Noise Learning without Anchor Points
CORES83.60Learning with Instance-Dependent Label Noise: A Sample Sieve Approach
ELR+91.09Early-Learning Regularization Prevents Memorization of Noisy Labels
Peer Loss82.53Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates
0 of 25 row(s) selected.