Shared Logistic Normal Distributions for Soft Parameter Tying in Unsupervised Grammar Induction

1 Jun 2009  ·  Shay Cohen, Noah A. Smith ·

We present a family of priors over probabilistic grammar weights, called the shared logistic normal distribution. This family extends the partitioned logistic normal distribution, enabling factored covariance between the probabilities of different derivation events in the probabilistic grammar, providing a new way to encode prior knowledge about an unknown grammar. We describe a variational EM algorithm for learning a probabilistic grammar based on this family of priors. We then experiment with unsupervised dependency grammar induction and show significant improvements using our model for both monolingual learning and bilingual learning with a non-parallel, multilingual corpus.

PDF Abstract

Datasets


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Unsupervised Dependency Parsing Penn Treebank Shared Logistic Normal DMV UAS 41.4 # 4

Methods


No methods listed for this paper. Add relevant methods here