HyperAI

Machine Translation On Iwslt2014 German

المقاييس

BLEU score

النتائج

نتائج أداء النماذج المختلفة على هذا المعيار القياسي

اسم النموذج
BLEU score
Paper TitleRepository
Transformer34.44Attention Is All You Need
Rfa-Gate-arccos34.4Random Feature Attention-
TaLK Convolutions35.5Time-aware Large Kernel Convolutions
Bi-SimCut38.37Bi-SimCut: A Simple Strategy for Boosting Neural Machine Translation
Mask Attention Network (small)36.3Mask Attention Networks: Rethinking and Strengthen Transformer
Cutoff+Knee37.78Wide-minima Density Hypothesis and the Explore-Exploit Learning Rate Schedule
CNAT31.15Non-Autoregressive Translation by Learning Target Categorical Codes
Minimum Risk Training [Edunov2017]32.84Classical Structured Prediction Losses for Sequence to Sequence Learning
BiBERT38.61BERT, mBERT, or BiBERT? A Study on Contextualized Embeddings for Neural Machine Translation
LightConv34.8Pay Less Attention with Lightweight and Dynamic Convolutions
TransformerBase + AutoDropout35.8AutoDropout: Learning Dropout Patterns to Regularize Deep Networks
DynamicConv35.2Pay Less Attention with Lightweight and Dynamic Convolutions
Transformer + R-Drop37.25R-Drop: Regularized Dropout for Neural Networks
Transformer35.1385Guidelines for the Regularization of Gammas in Batch Normalization for Deep Residual Networks-
Back-Translation Finetuning28.83Tag-less Back-Translation-
Local Joint Self-attention35.7Joint Source-Target Self Attention with Locality Constraints
Transformer + R-Drop + Cutoff37.90R-Drop: Regularized Dropout for Neural Networks
Data Diversification37.2Data Diversification: A Simple Strategy For Neural Machine Translation
UniDrop36.88UniDrop: A Simple yet Effective Technique to Improve Transformer without Extra Cost-
Actor-Critic [Bahdanau2017]28.53An Actor-Critic Algorithm for Sequence Prediction
0 of 34 row(s) selected.