8 papers with code • 1 benchmarks • 1 datasets
Inducing a constituency-based phrase structure grammar.
When a larger constituent ends, all of the smaller constituents that are nested within it must also be closed.
Ranked #6 on Constituency Grammar Induction on PTB
On language modeling, unsupervised RNNGs perform as well their supervised counterparts on benchmarks in English and Chinese.
Ranked #5 on Constituency Grammar Induction on PTB (Max F1 (WSJ) metric)
We study a formalization of the grammar induction problem that models sentences as being generated by a compound probabilistic context-free grammar.
Ranked #3 on Constituency Grammar Induction on PTB
We introduce the deep inside-outside recursive autoencoder (DIORA), a fully-unsupervised method for discovering syntax that simultaneously learns representations for constituents within the induced tree.
Ranked #2 on Constituency Grammar Induction on PTB
In this work, we propose a novel generative model that jointly learns discrete syntactic structure and continuous word representations in an unsupervised fashion by cascading an invertible neural network with a structured generative prior.
Ranked #7 on Constituency Grammar Induction on PTB
In this paper, We propose a novel neural language model, called the Parsing-Reading-Predict Networks (PRPN), that can simultaneously induce the syntactic structure from unannotated sentences and leverage the inferred structure to learn a better language model.
Ranked #9 on Constituency Grammar Induction on PTB
In this work, we present a new parameterization form of PCFGs based on tensor decomposition, which has at most quadratic computational complexity in the symbol number and therefore allows us to use a much larger number of symbols.
Ranked #1 on Constituency Grammar Induction on PTB