HyperAI

Learning With Noisy Labels On Cifar 10N 3

Métriques

Accuracy (mean)

Résultats

Résultats de performance de divers modèles sur ce benchmark

Nom du modèle
Accuracy (mean)
Paper TitleRepository
GCE87.58Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels
F-div89.55When Optimizing $f$-divergence is Robust with Label Noise
CAL90.74Clusterability as an Alternative to Anchor Points When Learning with Noisy Labels
Backward-T86.86Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach
CORES89.79Learning with Instance-Dependent Label Noise: A Sample Sieve Approach
JoCoR90.11Combating noisy labels by agreement: A joint training method with co-regularization
Positive-LS89.82Does label smoothing mitigate label noise?-
ELR91.41Early-Learning Regularization Prevents Memorization of Noisy Labels
VolMinNet88.19Provably End-to-end Label-Noise Learning without Anchor Points
SOP+95.39Robust Training under Label Noise by Over-parameterization
Co-Teaching90.15Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels
Co-Teaching+89.54How does Disagreement Help Generalization against Label Corruption?
Forward-T87.04Making Deep Neural Networks Robust to Label Noise: a Loss Correction Approach
PSSCL96.49PSSCL: A progressive sample selection framework with contrastive loss designed for noisy labels
GNL91.83Partial Label Supervision for Agnostic Generative Noisy Label Learning
CORES*94.74Learning with Instance-Dependent Label Noise: A Sample Sieve Approach
ILL95.13Imprecise Label Learning: A Unified Framework for Learning with Various Imprecise Label Configurations
CE85.16--
Negative-LS90.13To Smooth or Not? When Label Smoothing Meets Noisy Labels
Peer Loss88.57Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates
0 of 23 row(s) selected.