3260 papers • 126 benchmarks • 313 datasets
Inducing a constituency-based phrase structure grammar.
(Image credit: Papersgraph)
These leaderboards are used to track progress in constituency-grammar-induction-8
Use these libraries to find constituency-grammar-induction-8 models and implementations
No subtasks available.
The novel recurrent architecture, ordered neurons LSTM (ON-LSTM), achieves good performance on four different tasks: language modeling, unsupervised parsing, targeted syntactic evaluation, and logical inference.
A formalization of the grammar induction problem that models sentences as being generated by a compound probabilistic context free grammar, which is modulated by a per-sentence continuous latent variable, which induces marginal dependencies beyond the traditional context-free assumptions.
This work uses tensor rank decomposition (aka. CPD) to decrease inference computational complexities for a subset of FGGs subsuming HMMs and PCFGs, and conducts experiments on HMM language modeling and unsupervised PCFG parsing, showing better performance.
A novel neural language model, called the Parsing-Reading-Predict Networks (PRPN), that can simultaneously induce the syntactic structure from unannotated sentences and leverage the inferred structure to learn a better language model is proposed.
An inference network parameterized as a neural CRF constituency parser is developed to maximize the evidence lower bound and apply amortized variational inference to unsupervised learning of RNNGs.
DIORA is introduced, a fully-unsupervised method for discovering syntax that simultaneously learns representations for constituents within the induced tree that outperforms previously reported results for unsupervised binary constituency parsing on the benchmark WSJ dataset.
A novel generative model is proposed that jointly learns discrete syntactic structure and continuous word representations in an unsupervised fashion by cascading an invertible neural network with a structured generative prior so long as the prior is well-behaved.
This work studies visually grounded grammar induction and learns a constituency parser from both unlabeled text and its visual groundings, and shows that using an extension of probabilistic context-free grammar model, it can do fully-differentiable end-to-end visually grounded learning.
A new parameterization form of PCFGs based on tensor decomposition is presented, which has at most quadratic computational complexity in the symbol number and therefore allows us to use a much larger number of symbols.
This paper proposes an approach to parameterize L-PCFGs without making implausible independence assumptions, which directly models bilexical dependencies and meanwhile reduces both learning and representation complexities of L- PCFGs.
Adding a benchmark result helps the community track progress.