Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

26
Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group

Transcript of Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Page 1: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Posterior Regularization for Structured Latent Variable Models

Li ZhonghuaI2R SMT Reading Group

Page 2: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Outline

• Motivation and Introduction• Posterior Regularization• Application• Implementation• Some Related Frameworks

Page 3: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Motivation and Introduction

Prior Knowledge

We posses a wealth of prior knowledge about most NLP tasks.

Page 4: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Motivation and Introduction --Prior Knowledge

Page 5: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Motivation and Introduction --Prior Knowledge

Page 6: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Motivation and Introduction

Leveraging Prior Knowledge Possible approaches and their limitations

Page 7: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Motivation and Introduction --Limited Approach

Bayesian Approach : Encode prior knowledge with a prior on parameters

Limitation: Our prior knowledge is not about parameters!Parameters are difficult to interpret; hard to get desired effect.

Page 8: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Augmenting Model : Encode prior knowledge with additional variables and dependencies.

Motivation and Introduction --Limited Approach

limitation: may make exact inference intractable

Page 9: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Posterior Regularization

• A declarative language for specifying prior knowledge

-- Constraint Features & Expectations

• Methods for learning with knowledge in this language

-- EM style learning algorithm

Page 10: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Posterior Regularization

Page 11: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Posterior Regularization

Original Objective :

Page 12: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Posterior RegularizationEM style learning algorithm

Page 13: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Posterior Regularization

Computing the Posterior Regularizer

Page 14: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Application

Statistical Word AlignmentsIBM Model 1 and HMM

Page 15: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Application

One feature for each source word m, that counts how many times it is aligned to a target word in the alignment y.

Page 16: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Application

Define feature for each target-source position pair i,j . The feature takes the value zero in expectation if a word pair i ,j is aligned with equal probability in both directions.

Page 17: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Application

Learning Tractable Word Alignment Models with Complex Constraints CL10

Page 18: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Application

• Six language pairs• both types of constraints improve over the

HMM in terms of both precision and recall• improve over the HMM by 10% to 15%• S-HMM performs slightly better than B-HMM• S-HMM performs better than B-HMM in 10

out of 12 cases• improve over IBM M4 9 times out of 12

Page 19: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Application

Page 20: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Implementation

• http://code.google.com/p/pr-toolkit/

Page 21: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Some Related Frameworks

Page 22: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Some Related Frameworks

Page 23: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Some Related Frameworks

Page 24: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Some Related Frameworks

Page 25: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

Some Related Frameworks

Page 26: Posterior Regularization for Structured Latent Variable Models Li Zhonghua I2R SMT Reading Group.

more info: http://sideinfo.wikkii.com many of my slides get from there

Thanks!