HyperAI

Constituency Parsing On Penn Treebank

Metrics

F1 score

Results

Performance results of various models on this benchmark

Model Name
F1 score
Paper TitleRepository
Model combination94.66Improving Neural Parsing by Disentangling Model Combination and Reranking Effects-
Label Attention Layer + HPSG + XLNet96.38Rethinking Self-Attention: Towards Interpretability in Neural Parsing
RNN Grammar93.3Recurrent Neural Network Grammars
Tetra Tagging95.44Tetra-Tagging: Word-Synchronous Parsing with Linear-Time Inference
Head-Driven Phrase Structure Grammar Parsing (Joint) + XLNet96.33Head-Driven Phrase Structure Grammar Parsing on Penn Treebank
Hashing + XLNet96.43To be Continuous, or to be Discrete, Those are Bits of Questions
NFC + BERT-large95.92Investigating Non-local Features for Neural Constituency Parsing
Stack-only RNNG93.6What Do Recurrent Neural Network Grammars Learn About Syntax?
CRF Parser + RoBERTa96.32Fast and Accurate Neural CRF Constituency Parsing
Transformer92.7Attention Is All You Need
Self-attentive encoder + ELMo95.13Constituency Parsing with a Self-Attentive Encoder
SAPar + XLNet96.40Improving Constituency Parsing with Span Attention
Semi-supervised LSTM-LM93.8--
CNN Large + fine-tune95.6Cloze-driven Pretraining of Self-attention Networks-
N-ary semi-markov + BERT-large95.92N-ary Constituent Tree Parsing with Recursive Semi-Markov Model
Head-Driven Phrase Structure Grammar Parsing (Joint) + BERT95.84Head-Driven Phrase Structure Grammar Parsing on Penn Treebank
Parse fusion92.6--
LSTM Encoder-Decoder + LSTM-LM94.47Direct Output Connection for a High-Rank Language Model
Attach-Juxtapose Parser + XLNet96.34Strongly Incremental Constituency Parsing with Graph Neural Networks
Self-training92.1Effective Self-Training for Parsing
0 of 27 row(s) selected.