Maximum Entropy Model (I) LING 572 Advanced Statistical Methods for - - PowerPoint PPT Presentation

maximum entropy model i
SMART_READER_LITE
LIVE PREVIEW

Maximum Entropy Model (I) LING 572 Advanced Statistical Methods for - - PowerPoint PPT Presentation

Maximum Entropy Model (I) LING 572 Advanced Statistical Methods for NLP January 28, 2020 1 MaxEnt in NLP The maximum entropy principle has a long history. The MaxEnt algorithm was introduced to the NLP field by Berger et. al. (1996).


slide-1
SLIDE 1

Maximum Entropy Model (I)

LING 572 Advanced Statistical Methods for NLP January 28, 2020

1

slide-2
SLIDE 2

MaxEnt in NLP

  • The maximum entropy principle has a long history.
  • The MaxEnt algorithm was introduced to the NLP field by Berger et. al.

(1996).

  • Used in many NLP tasks: Tagging, Parsing, PP attachment, …

2

slide-3
SLIDE 3

Readings & Comments

  • Several readings:
  • (Berger, 1996), (Ratnaparkhi, 1997)
  • (Klein & Manning, 2003): Tutorial
  • Note: Some of these are very ‘dense’
  • Don’t spend huge amount of time on every detail
  • Take a first pass before class, review after lecture
  • Going forward:
  • Techniques more complex
  • Goal: Understand basic model, concepts
  • Training is complex; we’ll discuss, but not implement

3

slide-4
SLIDE 4

Notation

4

Input Output Pair Berger et al 1996

x y (x, y)

Ratnaparkhi 1997

b a x

Ratnaparkhi 1996

h t (h, t)

Klein and Manning 2003

d c (d, c)

We use this one

slide-5
SLIDE 5

Outline

  • Overview
  • The Maximum Entropy Principle
  • Modeling**
  • Decoding
  • Training**
  • Case study: POS tagging

5

slide-6
SLIDE 6

Overview

6

slide-7
SLIDE 7

Joint vs. Conditional models

  • Given training data {(x,y)}, we want to build a model to predict y for new x’s. For each model, we

need to estimate the parameters µ.

  • Joint (aka generative) models estimate P(x,y) by maximizing the likelihood: P(X,Y|µ)
  • Ex: n-gram models, HMM, Naïve Bayes, PCFG
  • Choosing weights is trivial: just use relative frequencies.
  • Conditional (aka discriminative) models estimate P(y | x) by maximizing the conditional likelihood:

P(Y | X, µ)

  • Ex: MaxEnt, SVM, CRF, etc.
  • Computing weights is more complex.

7

slide-8
SLIDE 8

Naïve Bayes Model

8

C

f1 f2 fn

Assumption: each fi is conditionally independent from fj given C.

slide-9
SLIDE 9

The conditional independence assumption

fm and fn are conditionally independent given c: P(fm | c, fn) = P(fm | c) Counter-examples in the text classification task:

  • P(“Manchester” | entertainment) !=

P(“Manchester” | entertainment, “Oscar”) Q: How to deal with correlated features? A: Many models, including MaxEnt, do not assume that features are conditionally independent.

9

slide-10
SLIDE 10

Naïve Bayes highlights

  • Choose

c* = arg maxc P(c) ∏k P(fk | c)

  • Two types of model parameters:
  • Class prior: P(c)
  • Conditional probability: P(fk | c)
  • The number of model parameters:

|C|+|CV|

10

slide-11
SLIDE 11

P(f | c) in NB

11

f1 f2 … fj

c1 P(f1 |c1) P(f2 |c1) … P(fj | c1) c2 P(f1 |c2) … … … … … ci P(f1 |ci) … … P(fj | ci) Each cell is a weight for a particular (class, feat) pair.

slide-12
SLIDE 12

Weights in NB and MaxEnt

  • In NB
  • P(f | y) are probabilities (i.e., in [0,1])
  • P(f | y) are multiplied at test time
  • In MaxEnt
  • the weights are real numbers: they can be negative.
  • the weighted features are added at test time

12

slide-13
SLIDE 13

Highlights of MaxEnt

13

Training: to estimate Testing: to calculate P(y | x) fj(x,y) is a feature function, which normally corresponds to a (feature, class) pair.

slide-14
SLIDE 14

Main questions

  • What is the maximum entropy principle?
  • What is a feature function?
  • Modeling: Why does P(y|x) have the form?
  • Training: How do we estimate λj ?

14

slide-15
SLIDE 15

Outline

  • Overview
  • The Maximum Entropy Principle
  • Modeling**
  • Decoding
  • Training*
  • Case study

15

slide-16
SLIDE 16

Maximum Entropy Principle

16

slide-17
SLIDE 17

Maximum Entropy Principle

  • Intuitively, model all that is known, and assume as little as possible about

what is unknown.

  • Related to Occam’s razor and other similar justifications for scientific

inquiry

  • Also: Laplace’s Principle of Insufficient Reason: when one has no

information to distinguish between the probability of two events, the best strategy is to consider them equally likely.

17

slide-18
SLIDE 18

Maximum Entropy

  • Why maximum entropy?
  • Maximize entropy = Minimize commitment
  • Model all that is known and assume nothing about what is

unknown.

  • Model all that is known: satisfy a set of constraints that must hold
  • Assume nothing about what is unknown: 


choose the most “uniform” distribution ➔ choose the one with maximum entropy

18

slide-19
SLIDE 19

Ex1: Coin-flip example
 (Klein & Manning, 2003)

  • Toss a coin: p(H)=p1, p(T)=p2.
  • Constraint: p1 + p2 = 1
  • Question: what’s p(x)? That is, what is the value of p1?
  • Answer: choose the p that maximizes



 


H(p) = − ∑

x

p(x)log p(x)

19

slide-20
SLIDE 20

Ex2: An MT example
 (Berger et. al., 1996)

20

Possible translation for the word “in” is: {dans, en, à, au cours de, pendant} Constraint: Intuitive answer:

slide-21
SLIDE 21

An MT example (cont)

21

Constraints: Intuitive answer:

slide-22
SLIDE 22

An MT example (cont)

22

Constraints: Intuitive answer: ??

slide-23
SLIDE 23

Ex3: POS tagging
 (Klein and Manning, 2003)

23

slide-24
SLIDE 24

Ex3 (cont)

24

slide-25
SLIDE 25

Ex4: Overlapping features
 (Klein and Manning, 2003)

25

p1 p2 p3 p4

slide-26
SLIDE 26

Ex4 (cont)

26

p1 p2

slide-27
SLIDE 27

Ex4 (cont)

27

p1

slide-28
SLIDE 28

The MaxEnt Principle summary

  • Goal: Among all the distributions that satisfy the constraints,

choose the one, p*, that maximizes H(p).
 
 
 
 


  • Q1: How to represent constraints?
  • Q2: How to find such distributions?

p* = arg max

p∈P H(p)

28