Slides2
Transcription
Slides2
COMS 4705 (Fall 2011)
Machine Translation Part II
1
Recap: The Noisy Channel Model
• Goal: translation system from French to English
• Have a model p(e | f ) which estimates conditional probability of any
English sentence e given the French sentence f . Use the training corpus to
set the parameters.
• A Noisy Channel Model has two components:
p(e)
p(f | e)
the language model
the translation model
• Giving:
p(e)p(f | e)
p(e, f )
P
=
p(e | f ) =
p(f )
e p(e)p(f | e)
and
argmaxe p(e | f ) = argmaxe p(e)p(f | e)
2
Roadmap for the Next Few Lectures
• Lecture 1 (today): IBM Models 1 and 2
• Lecture 2: phrase-based models
• Lecture 3: Syntax in statistical machine translation
3
Overview
• IBM Model 1
• IBM Model 2
• EM Training of Models 1 and 2
• Some examples of training Models 1 and 2
• Decoding
4
IBM Model 1: Alignments
• How do we model p(f | e)?
• English sentence e has l words e1 . . . el ,
French sentence f has m words f1 . . . fm .
• An alignment a identifies which English word each French
word originated from
• Formally, an alignment a is {a1 , . . . am }, where each aj ∈
{0 . . . l}.
• There are (l + 1)m possible alignments.
5
IBM Model 1: Alignments
• e.g., l = 6, m = 7
e = And the program has been implemented
f = Le programme a ete mis en application
• One alignment is
{2, 3, 4, 5, 6, 6, 6}
• Another (bad!) alignment is
{1, 1, 1, 1, 1, 1, 1}
6
Alignments in the IBM Models
• We’ll define models for p(a | e) and p(f | a, e),
giving
p(f , a | e) = p(a | e)p(f | a, e)
• Also,
p(f | e) =
X
p(a | e)p(f | a, e)
a∈A
where A is the set of all possible alignments
7
A By-Product: Most Likely Alignments
• Once we have a model p(f , a | e) = p(a | e)p(f | a, e) we can
also calculate
p(f , a | e)
p(a | f , e) = P
a∈A p(f , a | e)
for any alignment a
• For a given f , e pair, we can also compute the most likely
alignment,
a∗ = arg max
p(a | f , e)
a
• Nowadays, the original IBM models are rarely (if ever) used
for translation, but they are used for recovering alignments
8
An Example Alignment
French:
le conseil a rendu son avis , et nous devons à présent adopter un
nouvel avis sur la base de la première position .
English:
the council has stated its position , and now , on the basis of the first
position , we again have to give our opinion .
Alignment:
the/le council/conseil has/à stated/rendu its/son position/avis ,/,
and/et now/présent ,/NULL on/sur the/le basis/base of/de the/la
first/première position/position ,/NULL we/nous again/NULL
have/devons to/a give/adopter our/nouvel opinion/avis ./.
9
IBM Model 1: Alignments
• In IBM model 1 all allignments a are equally likely:
1
p(a | e) = C ×
(l + 1)m
where C = prob(length(f ) = m) is a constant.
• This is a major simplifying assumption, but it gets things
started...
10
IBM Model 1: Translation Probabilities
• Next step: come up with an estimate for
p(f | a, e)
• In model 1, this is:
p(f | a, e) =
m
Y
j=1
11
t(fj | eaj )
• e.g., l = 6, m = 7
e = And the program has been implemented
f = Le programme a ete mis en application
• a = {2, 3, 4, 5, 6, 6, 6}
p(f | a, e) = t(Le | the) ×
t(programme | program) ×
t(a | has) ×
t(ete | been) ×
t(mis | implemented) ×
t(en | implemented) ×
t(application | implemented)
12
IBM Model 1: The Generative Process
To generate a French string f from an English string e:
• Step 1: Pick the length of f (all lengths equally probable,
probability C)
• Step 2: Pick an alignment a with probability
1
(l+1)m
• Step 3: Pick the French words with probability
p(f | a, e) =
m
Y
t(fj | eaj )
j=1
The final result:
m
Y
C
p(f , a | e) = p(a | e) × p(f | a, e) =
t(fj | eaj )
m
(l + 1) j=1
13
An Example
• I have the following training examples
the dog ⇒ le chien
the cat ⇒ le chat
• Need to find estimates for:
p(le | the)
p(chien | the)
p(chat | the)
p(le | dog)
p(chien | dog)
p(chat | dog)
p(le | cat)
p(chien | cat)
p(chat | cat)
• As a result, each (ei , fi ) pair will have a most likely alignment.
14
An Example Lexical Entry
English
position
position
position
position
position
position
French
position
situation
mesure
vue
point
attitude
Probability
0.756715
0.0547918
0.0281663
0.0169303
0.0124795
0.0108907
. . . de la situation au niveau des négociations de l ’ ompi . . .
. . . of the current position in the wipo negotiations . . .
nous ne sommes pas en mesure de décider , . . .
we are not in a position to decide , . . .
. . . le point de vue de la commission face à ce problème complexe .
. . . the commission ’s position on this complex problem .
. . . cette attitude laxiste et irresponsable .
. . . this irresponsibly lax position .
15
Overview
• IBM Model 1
• IBM Model 2
• EM Training of Models 1 and 2
• Some examples of training Models 1 and 2
• Decoding
16
IBM Model 2
• Only difference: we now introduce alignment or distortion parameters
q(i | j, l, m) = Probability that j’th French word is connected
to i’th English word, given sentence lengths of
e and f are l and m respectively
• Define
p(a | e, l, m) =
m
Y
q(aj | j, l, m)
j=1
where a = {a1 , . . . am }
• Gives
p(f , a | e, l, m) =
m
Y
j=1
17
q(aj | j, l, m)t(fj | eaj )
• Note: Model 1 is a special case of Model 2, where q(i | j, l, m) =
all i, j.
18
1
l+1
for
An Example
l
m
=
=
6
7
e
= And the program has been implemented
f
= Le programme a ete mis en application
a
= {2, 3, 4, 5, 6, 6, 6}
p(a | e, 6, 7) = q(2 | 1, 6, 7) ×
q(3 | 2, 6, 7) ×
q(4 | 3, 6, 7) ×
q(5 | 4, 6, 7) ×
q(6 | 5, 6, 7) ×
q(6 | 6, 6, 7) ×
q(6 | 7, 6, 7)
19
p(f | a, e) = t(Le | the) ×
t(programme | program) ×
t(a | has) ×
t(ete | been) ×
t(mis | implemented) ×
t(en | implemented) ×
t(application | implemented)
20
IBM Model 2: The Generative Process
To generate a French string f from an English string e:
• Step 1: Pick the length of f (all lengths equally probable, probability C)
• Step 2: Pick an alignment a = {a1 , a2 . . . am } with probability
m
Y
q(aj | j, l, m)
j=1
• Step 3: Pick the French words with probability
p(f | a, e) =
m
Y
t(fj | eaj )
j=1
The final result:
p(f , a | e) = p(a | e)p(f | a, e) = C
m
Y
j=1
21
q(aj | j, l, m)t(fj | eaj )
Overview
• IBM Model 1
• IBM Model 2
• EM Training of Models 1 and 2
• Some examples of training Models 1 and 2
• Decoding
22
Overview
• IBM Model 1
• IBM Model 2
• EM Training of Models 1 and 2
• Some examples of training Models 1 and 2
• Decoding
23
An Example of Training Models 1 and 2
Example will use following translations:
e[1] = the dog
f[1] = le chien
e[2] = the cat
f[2] = le chat
e[3] = the bus
f[3] = l’
autobus
NB: I won’t use a NULL word e0
24
e
the
the
the
the
the
dog
dog
dog
dog
Initial (random) parameters: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
25
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
t(f | e)
0.23
0.2
0.11
0.25
0.21
0.2
0.16
0.33
0.12
0.18
0.26
0.28
0.19
0.24
0.03
0.22
0.05
0.26
0.19
0.27
Alignment probabilities:
i
1
2
1
2
1
2
1
2
1
2
1
2
j
1
1
2
2
1
1
2
2
1
1
2
2
k
0
0
0
0
1
1
1
1
2
2
2
2
a(i,j,k)
0.526423237959726
0.473576762040274
0.552517995605817
0.447482004394183
0.466532602066533
0.533467397933467
0.356364544422507
0.643635455577493
0.571950438336247
0.428049561663753
0.439081311724508
0.560918688275492
26
e
the
the
the
the
the
dog
dog
dog
dog
Expected counts: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
tcount(e, f )
0.99295584002626
0.552517995605817
0.356364544422507
0.571950438336247
0.439081311724508
0.473576762040274
0.447482004394183
0
0
0
0.533467397933467
0
0.643635455577493
0
0
0
0
0
0.428049561663753
0.560918688275492
27
e
the
the
the
the
the
dog
dog
dog
dog
Old and new parameters: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
28
old
0.23
0.2
0.11
0.25
0.21
0.2
0.16
0.33
0.12
0.18
0.26
0.28
0.19
0.24
0.03
0.22
0.05
0.26
0.19
0.27
new
0.34
0.19
0.12
0.2
0.15
0.51
0.49
0
0
0
0.45
0
0.55
0
0
0
0
0
0.43
0.57
e
the
the
the
the
the
dog
dog
dog
dog
dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
0.23
0.2
0.11
0.25
0.21
0.2
0.16
0.33
0.12
0.18
0.26
0.28
0.19
0.24
0.03
0.22
0.05
0.26
0.19
0.27
0.34
0.19
0.12
0.2
0.15
0.51
0.49
0
0
0
0.45
0
0.55
0
0
0
0
0
0.43
0.57
0.46
0.15
0.1
0.17
0.12
0.46
0.54
0
0
0
0.41
0
0.59
0
0
0
0
0
0.47
0.53
0.56
0.12
0.08
0.15
0.1
0.39
0.61
0
0
0
0.36
0
0.64
0
0
0
0
0
0.47
0.53
29
0.64
0.09
0.06
0.13
0.08
0.33
0.67
0
0
0
0.3
0
0.7
0
0
0
0
0
0.47
0.53
0.71
0.06
0.04
0.11
0.07
0.28
0.72
0
0
0
0.26
0
0.74
0
0
0
0
0
0.48
0.52
e
the
the
the
the
the
dog
dog
dog
dog
After 20 iterations: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
0.94
0
0
0.03
0.02
0.06
0.94
0
0
0
0.06
0
0.94
0
0
0
0
0
0.49
0.51
30
e
the
the
the
the
the
dog
dog
dog
dog
Model 2 has several local maxima – good one: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
31
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
t(f | e)
0.67
0
0
0.33
0
0
1
0
0
0
0
0
1
0
0
0
0
0
0
1
e
the
the
the
the
the
dog
dog
dog
dog
Model 2 has several local maxima – bad one: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
32
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
t(f | e)
0
0.4
0.3
0
0.3
0.5
0.5
0
0
0
0.5
0
0.5
0
0
0
0
0
0.5
0.5
e
the
the
the
the
the
dog
dog
dog
dog
another bad one: dog
cat
cat
cat
cat
cat
bus
bus
bus
bus
bus
f
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
le
chien
chat
l’
autobus
t(f | e)
0
0.33
0.33
0
0.33
1
0
0
0
0
1
0
0
0
0
0
0
0
1
0
33
• Alignment parameters for good solution:
q(i = 1 | j = 1, l = 2, m = 2) =
1
q(i = 2 | j = 1, l = 2, m = 2) =
0
q(i = 1 | j = 2, l = 2, m = 2) =
0
q(i = 2 | j = 2, l = 2, m = 2) =
1
log probability = −1.91
• Alignment parameters for first bad solution:
q(i = 1 | j = 1, l = 2, m = 2) =
0
q(i = 2 | j = 1, l = 2, m = 2) =
1
q(i = 1 | j = 2, l = 2, m = 2) =
0
q(i = 2 | j = 2, l = 2, m = 2) =
1
log probability = −4.16
34
• Alignment parameters for second bad solution:
q(i = 1 | j = 1, l = 2, m = 2) =
q(i = 2 | j = 1, l = 2, m = 2) =
0
1
q(i = 1 | j = 2, l = 2, m = 2) =
1
q(i = 2 | j = 2, l = 2, m = 2) =
0
log probability = −3.30
35
Improving the Convergence Properties of Model 2
• Out of 100 random starts, only 60 converged to the best
local maxima
• Model 1 converges to the same, global maximum every time
(see the Brown et. al paper)
• Method in IBM paper: run Model 1 to estimate t parameters,
then use these as the initial parameters for Model 2
• In 100 tests using this method, Model 2 converged to the
correct point every time.
36
Overview
• IBM Model 1
• IBM Model 2
• EM Training of Models 1 and 2
• Some examples of training Models 1 and 2
• Decoding
37