Grammatical Error Detection On Conll 2014 A1
Metrics
F0.5
Results
Performance results of various models on this benchmark
Comparison Table
Model Name | F0.5 |
---|---|
jointly-learning-to-label-sentences-and | 22.14 |
auxiliary-objectives-for-neural-error | 36.1 |
artificial-error-generation-with-machine | 21.87 |
semi-supervised-multitask-learning-for | 17.86 |
compositional-sequence-labeling-models-for | 34.3 |
compositional-sequence-labeling-models-for | 16.4 |
auxiliary-objectives-for-neural-error | 17.5 |
neural-quality-estimation-with-multiple | 54.3 |