Learning With Noisy Labels On Cifar 100N

평가 지표

Accuracy (mean)

평가 결과

이 벤치마크에서 각 모델의 성능 결과

모델 이름
Accuracy (mean)
Paper TitleRepository
Peer Loss57.59Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates-
ILL65.84Imprecise Label Learning: A Unified Framework for Learning with Various Imprecise Label Configurations-
F-div57.10When Optimizing $f$-divergence is Robust with Label Noise-
PGDF74.08Sample Prior Guided Robust Model Learning to Suppress Noisy Labels-
T-Revision51.55Are Anchor Points Really Indispensable in Label-Noise Learning?-
VolMinNet57.80Provably End-to-end Label-Noise Learning without Anchor Points-
CE55.50--
Backward-T57.14Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach-
Co-Teaching60.37Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels-
Positive-LS55.84Does label smoothing mitigate label noise?-
CAL61.73Clusterability as an Alternative to Anchor Points When Learning with Noisy Labels-
ELR+66.72Early-Learning Regularization Prevents Memorization of Noisy Labels-
Divide-Mix71.13DivideMix: Learning with Noisy Labels as Semi-supervised Learning-
Negative-LS58.59To Smooth or Not? When Label Smoothing Meets Noisy Labels-
CORES*55.72Learning with Instance-Dependent Label Noise: A Sample Sieve Approach-
ELR58.94Early-Learning Regularization Prevents Memorization of Noisy Labels-
Forward-T57.01Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach-
JoCoR59.97Combating noisy labels by agreement: A joint training method with co-regularization-
CORES61.15Learning with Instance-Dependent Label Noise: A Sample Sieve Approach-
GCE56.73Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels-
0 of 24 row(s) selected.
Learning With Noisy Labels On Cifar 100N | SOTA | HyperAI초신경