HyperAI

Machine Translation On Iwslt2014 German

Métriques

BLEU score

Résultats

Résultats de performance de divers modèles sur ce benchmark

Nom du modèle
BLEU score
Paper TitleRepository
Transformer34.44Attention Is All You Need
Rfa-Gate-arccos34.4Random Feature Attention-
TaLK Convolutions35.5Time-aware Large Kernel Convolutions
Bi-SimCut38.37Bi-SimCut: A Simple Strategy for Boosting Neural Machine Translation
Mask Attention Network (small)36.3Mask Attention Networks: Rethinking and Strengthen Transformer
Cutoff+Knee37.78Wide-minima Density Hypothesis and the Explore-Exploit Learning Rate Schedule
CNAT31.15Non-Autoregressive Translation by Learning Target Categorical Codes
Minimum Risk Training [Edunov2017]32.84Classical Structured Prediction Losses for Sequence to Sequence Learning
BiBERT38.61BERT, mBERT, or BiBERT? A Study on Contextualized Embeddings for Neural Machine Translation
LightConv34.8Pay Less Attention with Lightweight and Dynamic Convolutions
TransformerBase + AutoDropout35.8AutoDropout: Learning Dropout Patterns to Regularize Deep Networks
DynamicConv35.2Pay Less Attention with Lightweight and Dynamic Convolutions
Transformer + R-Drop37.25R-Drop: Regularized Dropout for Neural Networks
Transformer35.1385Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks-
Back-Translation Finetuning28.83Tag-less Back-Translation-
Local Joint Self-attention35.7Joint Source-Target Self Attention with Locality Constraints
Transformer + R-Drop + Cutoff37.90R-Drop: Regularized Dropout for Neural Networks
Data Diversification37.2Data Diversification: A Simple Strategy For Neural Machine Translation
UniDrop36.88UniDrop: A Simple yet Effective Technique to Improve Transformer without Extra Cost-
Actor-Critic [Bahdanau2017]28.53An Actor-Critic Algorithm for Sequence Prediction
0 of 34 row(s) selected.