Model-Based Aligner Combination Using Dual Decomposition

John DeNero Google Research [email protected]

Abstract Unsupervised word alignment is most often modeled as a Markov process that generates a sentence f conditioned on its translation e. A similar model generating e from f will make different alignment predictions. Statistical machine translation systems combine the predictions of two directional models, typically using heuristic combination procedures like grow-diag-final. This paper presents a graphical model that embeds two directional aligners into a single model. Inference can be performed via dual decomposition, which reuses the efficient inference algorithms of the directional models. Our bidirectional model enforces a one-to-one phrase constraint while accounting for the uncertainty in the underlying directional models. The resulting alignments improve upon baseline combination heuristics in word-level and phrase-level evaluations.

1

Introduction

Word alignment is the task of identifying corresponding words in sentence pairs. The standard approach to word alignment employs directional Markov models that align the words of a sentence f to those of its translation e, such as IBM Model 4 (Brown et al., 1993) or the HMM-based alignment model (Vogel et al., 1996). Machine translation systems typically combine the predictions of two directional models, one which aligns f to e and the other e to f (Och et al., 1999). Combination can reduce errors and relax the one-to-many structural restriction of directional models. Common combination methods include the union or intersection of directional alignments, as

Klaus Macherey Google Research [email protected]

well as heuristic interpolations between the union and intersection like grow-diag-final (Koehn et al., 2003). This paper presents a model-based alternative to aligner combination. Inference in a probabilistic model resolves the conflicting predictions of two directional models, while taking into account each model’s uncertainty over its output. This result is achieved by embedding two directional HMM-based alignment models into a larger bidirectional graphical model. The full model structure and potentials allow the two embedded directional models to disagree to some extent, but reward agreement. Moreover, the bidirectional model enforces a one-to-one phrase alignment structure, similar to the output of phrase alignment models (Marcu and Wong, 2002; DeNero et al., 2008), unsupervised inversion transduction grammar (ITG) models (Blunsom et al., 2009), and supervised ITG models (Haghighi et al., 2009; DeNero and Klein, 2010). Inference in our combined model is not tractable because of numerous edge cycles in the model graph. However, we can employ dual decomposition as an approximate inference technique (Rush et al., 2010). In this approach, we iteratively apply the same efficient sequence algorithms for the underlying directional models, and thereby optimize a dual bound on the model objective. In cases where our algorithm converges, we have a certificate of optimality under the full model. Early stopping before convergence still yields useful outputs. Our model-based approach to aligner combination yields improvements in alignment quality and phrase extraction quality in Chinese-English experiments, relative to typical heuristic combinations methods applied to the predictions of independent directional models.

2

Model Definition

Our bidirectional model G = (V, D) is a globally normalized, undirected graphical model of the word alignment for a fixed sentence pair (e, f ). Each vertex in the vertex set V corresponds to a model variable Vi , and each undirected edge in the edge set D corresponds to a pair of variables (Vi , Vj ). Each vertex has an associated potential function ωi (vi ) that assigns a real-valued potential to each possible value vi of Vi .1 Likewise, each edge has an associated potential function µij (vi , vj ) that scores pairs of values. The probability under the model of any full assignment v to the model variables, indexed by V, factors over vertex and edge potentials. Y Y P(v) ∝ ωi (vi ) · µij (vi , vj ) vi ∈V

(vi ,vj )∈D

Our model contains two directional hidden Markov alignment models, which we review in Section 2.1, along with additional structure that that we introduce in Section 2.2. 2.1

HMM-Based Alignment Model

This section describes the classic hidden Markov model (HMM) based alignment model (Vogel et al., 1996). The model generates a sequence of words f conditioned on a word sequence e. We conventionally index the words of e by i and f by j. P(f |e) is defined in terms of a latent alignment vector a, where aj = i indicates that word position i of e aligns to word position j of f . X P(f |e) = P(f , a|e) a

P(f , a|e) =

|f | Y

D(aj |aj−1 )M(fj |eaj ) .

(1)

where c(i0 − i) is a learned distribution over signed distances, normalized over the possible transitions from i. The parameters of the conditional multinomial M and the transition model c can be learned from a sentence aligned corpus via the expectation maximization algorithm. The null parameter po is typically fixed.2 The highest probability word alignment vector under the model for a given sentence pair (e, f ) can be computed exactly using the standard Viterbi algorithm for HMMs in O(|e|2 · |f |) time. An alignment vector a can be converted trivially into a set of word alignment links A: Aa = {(i, j) : aj = i, i 6= 0} . Aa is constrained to be many-to-one from f to e; many positions j can align to the same i, but each j appears at most once. We have defined a directional model that generates f from e. An identically structured model can be defined that generates e from f . Let b be a vector of alignments where bi = j indicates that word position j of f aligns to word position i of e. Then, P(e, b|f ) is defined similarly to Equation 1, but with e and f swapped. We can distinguish the transition and emission distributions of the two models by subscripting them with their generative direction.

P(e, b|f ) =

D(aj = 0|aj−1 = i) = po D(aj = i0 6= 0|aj−1 = i) = (1 − po ) · c(i0 − i) , 1

Potentials in an undirected model play the same role as conditional probabilities in a directed model, but do not need to be locally normalized.

Df →e (bi |bi−1 )Mf →e (ei |fbi ) .

j=1

The vector b can be interpreted as a set of alignment links that is one-to-many: each value i appears at most once in the set.

j=1

In Equation 1 above, the emission model M is a learned multinomial distribution over word types. The transition model D is a multinomial over transition distances, which treats null alignments as a special case.

|e| Y

Ab = {(i, j) : bi = j, j 6= 0} . 2.2

A Bidirectional Alignment Model

We can combine two HMM-based directional alignment models by embedding them in a larger model 2 In experiments, we set po = 10−6 . Transitions from a nullaligned state aj−1 = 0 are also drawn from a fixed distribution, where D(aj = 0|aj−1 = 0) = 10−4 and for i0 ≥ 1,

D(aj = i0 |aj−1 = 0) ∝ 0.8

  |f | − i0 · |e| −j

.

With small po , the shape of this distribution has little effect on the alignment outcome.

c11

c21

c12

c22

c13

c23

a1

a2

a3

How

are

you

b1 你

b2 好

Figure 1: The structure of our graphical model for a simple sentence pair. The variables a are blue, b are red, and c are green.

(b) (b) (b) 12 {0, 1} is c13 Each model node for anc11element cij c∈ connected to aj and bi via coherence edges. These (a) (a) c11(a) to ensurec12that edges allow the model the threec13sets of variables, a, b, and c, together encode a coherent alignment analysis of the sentence pair. c21(b) c22(b)Figure 1 c23(b) depicts the graph structure of the model. c22(a) c21(a) c23(a) 2.3 Coherence Potentials

The potentials on coherence edges are not learned and do not express any patterns in the data. Instead, they are fixed functions that promote consistency bea1 directional a2 alignment avec3 tween the integer-valued tors a and b and the boolean-valued matrix c. How are Consider the assignment aj = i, where i you = 0 indicates that word fj is null-aligned, and i ≥ 1 indicates that fj aligns to ei . The coherence potential ensures the following relationship between the variable assignment aj = i and the variables ci0 j , for any i0 ∈ [1, |e|].

that includes all of the random variables of two di• If i = 0 (null-aligned), then all ci0 j = 0. rectional models, along with additional structure that promotes agreement c11(a) and resolves discrepancies. c12(a) c13(a) • If i > 0, then cij = 1. The original directional models include observed word sequences e and f , along with the two latent • ci0 j = 1 only if i0 ∈ {i − 1, i, i + 1}. alignment vectors a and b defined in Section 2.1. • Assigning ci0 j = 1 for i0 6= i incurs a cost e−α . Because the word(a)types and lengths (a) of e and f are (a) c21 c22 c23 always fixed by the observed sentence pair, we can Collectively, the list of cases above enforce an intudefine our model only over a and b, where the edge itive correspondence: an alignment aj = i ensures potentials between any aj , fj , and e are compiled that cij must be 1, adjacent neighbors may be 1 but (a) into a vertex potential function ωj on aj , defined incur a cost, and all other elements are 0. a1 e, and likewise fora2any bi . a3 This pattern of effects can be encoded in a potenin terms of f and tial function µ(c) for each coherence edge. These (a) ωj (i) = Me→f (fj |e i) edge potential functions takes an integer value i for How are you (b) some variable aj and a binary value k for some ci0 j . ωi (j) = Mf →e (ei |fj )  1 i=0∧k =0  The edge potentials between a and b encode the     transition model in Equation 1. 0 i=0∧k =1      (a) 1 i = i0 ∧ k = 1 µj−1,j (i, i0 ) = De→f (aj = i0 |aj−1 = i)  (c) µ(aj ,c 0 ) (i, k) = 0 (2) i = i0 ∧ k = 0 (b) i j  µi−1,i (j, j 0 ) = Df →e (bi = j 0 |bi−1 = j)  0   1 i 6= i ∧ k = 0    In addition, we include in our model a latent −α   e |i − i0 | = 1 ∧ k = 1   boolean matrix c that encodes the output of the com 0 |i − i0 | > 1 ∧ k = 1 bined aligners: Above, potentials of 0 effectively disallow certain c ∈ {0, 1}|e|×|f | . cases because a full assignment to (a, b, c) is scored This matrix encodes the alignment links proposed by the product of all model potentials. The poten(c) by the bidirectional model: tial function µ(bi ,c 0 ) (j, k) for a coherence edge beij Ac = {(i, j) : cij = 1} . tween b and c is defined similarly.

b1

b2

2.4

Model Properties

We interpret c as the final alignment produced by the model, ignoring a and b. In this way, we relax the one-to-many constraints of the directional models. However, all of the information about how words align is expressed by the vertex and edge potentials on a and b. The coherence edges and the link matrix c only serve to resolve conflicts between the directional models and communicate information between them. Because directional alignments are preserved intact as components of our model, extensions or refinements to the underlying directional Markov alignment model could be integrated cleanly into our model as well, including lexicalized transition models (He, 2007), extended conditioning contexts (Brunning et al., 2009), and external information (Shindo et al., 2010). For any assignment to (a, b, c) with non-zero probability, c must encode a one-to-one phrase alignment with a maximum phrase length of 3. That is, any word in either sentence can align to at most three words in the opposite sentence, and those words must be contiguous. This restriction is directly enforced by the edge potential in Equation 2.

the graph belongs to exactly one of these two subgraphs. The dual decomposition inference approach allows us to exploit this sub-graph structure (Rush et al., 2010). In particular, we can iteratively apply exact inference to the subgraph problems, adjusting their potentials to reflect the constraints of the full problem. The technique of dual decomposition has recently been shown to yield state-of-the-art performance in dependency parsing (Koo et al., 2010). 3.2

Dual Problem Formulation

To describe a dual decomposition inference procedure for our model, we first restate the inference problem under our graphical model in terms of the two overlapping subgraphs that admit tractable inference. Let c(a) be a copy of c associated with Ga , and c(b) with Gb . Also, let f (a, c(a) ) be the unnormalized log-probability of an assignment to Ga and g(b, c(b) ) be the unnormalized log-probability of an assignment to Gb . Finally, let I be the index set of all (i, j) for c. Then, the maximum likelihood assignment to our original model can be found by optimizing f (a, c(a) ) + g(b, c(b) )

max a,b,c(a) ,c(b)

3

Model Inference

In general, graphical models admit efficient, exact inference algorithms if they do not contain cycles. Unfortunately, our model contains numerous cycles. For every pair of indices (i, j) and (i0 , j 0 ), the following cycle exists in the graph: cij → bi → cij 0 → aj 0 →

(a)

3.1

Dual Decomposition

While the entire graphical model has loops, there are two overlapping subgraphs that are cycle-free. One subgraph Ga includes all of the vertices corresponding to variables a and c. The other subgraph Gb includes vertices for variables b and c. Every edge in

∀ (i, j) ∈ I .

The Lagrangian relaxation of this optimization problem is L(a, b, c(a) , c(b) , u) = X (a) (b) f (a, c(a) ) + g(b, c(b) ) + u(i, j)(ci,j − ci,j ) . (i,j)∈I

Hence, we can rewrite the original problem as max

ci0 j 0 → bi0 → ci0 j → aj → cij Additional cycles also exist in the graph through the edges between aj−1 and aj and between bi−1 and bi . The general phrase alignment problem under an arbitrary model is known to be NP-hard (DeNero and Klein, 2008).

(b)

such that: cij = cij

(3)

a,b,c(a) ,c(b)

min L(a, b, c(a) , c(b) , u) . u

We can form a dual problem that is an upper bound on the original optimization problem by swapping the order of min and max. In this case, the dual problem decomposes into two terms that are each local to an acyclic subgraph.    X (a) min  max f (a, c(a) ) + u(i, j)c  u

ij

a,c(a)

i,j

 + max g(b, c(b) ) − b,c(b)

 X i,j

(b) u(i, j)cij 

(4)

c11(b) c11(a)

c12(b) c12(a)

c21(b)

b1



c11(a)

c12(a)

c13(a)

c23(b)

b2



c21(a)

c22(a)

c23(a)

a1

a2

a3

How

are

you

c13(a) c22(b)

c22(a)

c21(a)

c13(b)

c23(a)

a1

a2

a3

How

are

you

Figure 2: Our combined model decomposes into two acyclic models that each contain a copy of c.

c13(a)

c23(a)

a3

The decomposed model is depicted in Figure 2. As in previous work, we solve for the dual variable u by repeatedly performing inference in the two decoupled maximization problems. 3.3

Sub-Graph Inference

We now address the problem of evaluating Equation 4 for fixed u. Consider the first line of Equation 4, which includes variables a and c(a) .   X (a) max f (a, c(a) ) + u(i, j)cij  (5) a,c(a)

you

i,j

Because the graph Ga is tree-structured, Equation 5 can be evaluated in polynomial time. In fact, we can make a stronger claim: we can reuse the Viterbi inference algorithm for linear chain graphical models that applies to the embedded directional HMM models. That is, we can cast the optimization of Equation 5 as   |f | Y max  De→f (aj |aj−1 ) · M0j (aj = i) . a

j=1

In the original HMM-based aligner, the vertex potentials correspond to bilexical probabilities. Those quantities appear in f (a, c(a) ), and therefore will be a part of M0j (·) above. The additional terms of Equation 5 can also be factored into the vertex potentials of this linear chain model, because the optimal

Figure 3: The tree-structured subgraph Ga can be mapped to an equivalent chain-structured model by optimizing over ci0 j for aj = i.

choice of each cij can be determined from aj and the model parameters. If aj = i, then cij = 1 according to our edge potential defined in Equation 2. Hence, setting aj = i requires the inclusion of the corre(a) sponding vertex potential ωj (i), as well as u(i, j). For i0 6= i, either ci0 j = 0, which contributes nothing to Equation 5, or ci0 j = 1, which contributes u(i0 , j)−α, according to our edge potential between aj and ci0 j . Thus, we can capture the net effect of assigning aj and then optimally assigning all ci0 j in a single potential M0j (aj = i) =  (a)



ωj (i) + expu(i, j) +

X

max(0, u(i, j 0 ) − α)

j 0 :|j 0 −j|=1

Note that Equation 5 and f are sums of terms in log space, while Viterbi inference for linear chains assumes a product of terms in probability space, which introduces the exponentiation above. Defining this potential allows us to collapse the source-side sub-graph inference problem defined by Equation 5, into a simple linear chain model that only includes potential functions M0j and µ(a) . Hence, we can use a highly optimized linear chain inference implementation rather than a solver for general tree-structured graphical models. Figure 3 depicts this transformation. An equivalent approach allows us to evaluate the

Algorithm 1 Dual decomposition inference algorithm for the bidirectional model for t = 1 to max iterations do r ← 1t . Learning rate P (a) (a) (a) c ← arg max f (a, c ) + i,j u(i, j)cij P (b) c(b) ← arg max g(b, c(b) ) − i,j u(i, j)cij if c(a) = c(b) then return c(a) . Converged u ← u + r · (c(b) − c(a) ) . Dual update return combine(c(a) , c(b) ) . Stop early

second line of Equation 4 for fixed u:  max g(b, c(b) ) +

b,c(b)

3.4

 X

(b) u(i, j)cij  .

(6)

i,j

Dual Decomposition Algorithm

Now that we have the means to efficiently evaluate Equation 4 for fixed u, we can define the full dual decomposition algorithm for our model, which searches for a u that optimizes Equation 4. We can iteratively search for such a u via sub-gradient descent. We use a learning rate 1t that decays with the number of iterations t. The full dual decomposition optimization procedure appears in Algorithm 1. If Algorithm 1 converges, then we have found a u such that the value of c(a) that optimizes Equation 5 is identical to the value of c(b) that optimizes Equation 6. Hence, it is also a solution to our original optimization problem: Equation 3. Since the dual problem is an upper bound on the original problem, this solution must be optimal for Equation 3. 3.5

Convergence and Early Stopping

Our dual decomposition algorithm provides an inference method that is exact upon convergence.3 When Algorithm 1 does not converge, the two alignments c(a) and c(b) can still be used. While these alignments may differ, they will likely be more similar than the alignments of independent aligners. These alignments will still need to be combined procedurally (e.g., taking their union), but because 3

This certificate of optimality is not provided by other approximate inference algorithms, such as belief propagation, sampling, or simulated annealing.

they are more similar, the importance of the combination procedure is reduced. We analyze the behavior of early stopping experimentally in Section 5. 3.6

Inference Properties

Because we set a maximum number of iterations n in the dual decomposition algorithm, and each iteration only involves optimization in a sequence model, our entire inference procedure is only a constant multiple n more computationally expensive than evaluating the original directional aligners. Moreover, the value of u is specific to a sentence pair. Therefore, our approach does not require any additional communication overhead relative to the independent directional models in a distributed aligner implementation. Memory requirements are virtually identical to the baseline: only u must be stored for each sentence pair as it is being processed, but can then be immediately discarded once alignments are inferred. Other approaches to generating one-to-one phrase alignments are generally more expensive. In particular, an ITG model requires O(|e|3 · |f |3 ) time, whereas our algorithm requires only O(n · (|f ||e|2 + |e||f |2 )) . Moreover, our approach allows Markov distortion potentials, while standard ITG models are restricted to only hierarchical distortion.

4

Related Work

Alignment combination normally involves selecting some A from the output of two directional models. Common approaches include forming the union or intersection of the directional sets. A∪ = Aa ∪ A b A∩ = Aa ∩ Ab . More complex combiners, such as the grow-diagfinal heuristic (Koehn et al., 2003), produce alignment link sets that include all of A∩ and some subset of A∪ based on the relationship of multiple links (Och et al., 1999). In addition, supervised word alignment models often use the output of directional unsupervised aligners as features or pruning signals. In the case

that a supervised model is restricted to proposing alignment links that appear in the output of a directional aligner, these models can be interpreted as a combination technique (Deng and Zhou, 2009). Such a model-based approach differs from ours in that it requires a supervised dataset and treats the directional aligners’ output as fixed. Combination is also related to agreement-based learning (Liang et al., 2006). This approach to jointly learning two directional alignment models yields state-of-the-art unsupervised performance. Our method is complementary to agreement-based learning, as it applies to Viterbi inference under the model rather than computing expectations. In fact, we employ agreement-based training to estimate the parameters of the directional aligners in our experiments. A parallel idea that closely relates to our bidirectional model is posterior regularization, which has also been applied to the word alignment problem (Grac¸a et al., 2008). One form of posterior regularization stipulates that the posterior probability of alignments from two models must agree, and enforces this agreement through an iterative procedure similar to Algorithm 1. This approach also yields state-of-the-art unsupervised alignment performance on some datasets, along with improvements in end-to-end translation quality (Ganchev et al., 2008). Our method differs from this posterior regularization work in two ways. First, we iterate over Viterbi predictions rather than posteriors. More importantly, we have changed the output space of the model to be a one-to-one phrase alignment via the coherence edge potential functions. Another similar line of work applies belief propagation to factor graphs that enforce a one-to-one word alignment (Cromi`eres and Kurohashi, 2009). The details of our models differ: we employ distance-based distortion, while they add structural correspondence terms based on syntactic parse trees. Also, our model training is identical to the HMMbased baseline training, while they employ belief propagation for both training and Viterbi inference. Although differing in both model and inference, our work and theirs both find improvements from defining graphical models for alignment that do not admit exact polynomial-time inference algorithms.

Aligner Intersection Union Model |A∩ | |A∪ | Baseline 5,554 10,998 Bidirectional 7,620 10,262

Agreement |A∩ |/|A∪ | 50.5% 74.3%

Table 1: The bidirectional model’s dual decomposition algorithm substantially increases the overlap between the predictions of the directional models, measured by the number of links in their intersection.

5

Experimental Results

We evaluated our bidirectional model by comparing its output to the annotations of a hand-aligned corpus. In this way, we can show that the bidirectional model improves alignment quality and enables the extraction of more correct phrase pairs. 5.1

Data Conditions

We evaluated alignment quality on a hand-aligned portion of the NIST 2002 Chinese-English test set (Ayan and Dorr, 2006). We trained the model on a portion of FBIS data that has been used previously for alignment model evaluation (Ayan and Dorr, 2006; Haghighi et al., 2009; DeNero and Klein, 2010). We conducted our evaluation on the first 150 sentences of the dataset, following previous work. This portion of the dataset is commonly used to train supervised models. We trained the parameters of the directional models using the agreement training variant of the expectation maximization algorithm (Liang et al., 2006). Agreement-trained IBM Model 1 was used to initialize the parameters of the HMM-based alignment models (Brown et al., 1993). Both IBM Model 1 and the HMM alignment models were trained for 5 iterations on a 6.2 million word parallel corpus of FBIS newswire. This training regimen on this data set has provided state-of-the-art unsupervised results that outperform IBM Model 4 (Haghighi et al., 2009). 5.2

Convergence Analysis

With n = 250 maximum iterations, our dual decomposition inference algorithm only converges 6.2% of the time, perhaps largely due to the fact that the two directional models have different one-to-many structural constraints. However, the dual decompo-

Model Baseline

Bidirectional

Combiner union intersect grow-diag union intersect grow-diag

Prec 57.6 86.2 60.1 63.3 77.5 65.6

Rec 80.0 62.7 78.8 81.5 75.1 80.6

AER 33.4 27.2 32.1 29.1 23.6 28.0

posed alignment set A to the sure set S and possible set P in the annotation, where S ⊆ P. |A ∩ P| |A| |A ∩ S| Rec(A, S) = |S| |A ∩ S| + |A ∩ P| AER(A, S, P) = 1 − |A| + |S| Prec(A, P) =

Table 2: Alignment error rate results for the bidirectional model versus the baseline directional models. “growdiag” denotes the grow-diag-final heuristic.

Model Baseline

Bidirectional

Combiner union intersect grow-diag union intersect grow-diag

Prec 75.1 64.3 68.3 63.2 57.1 60.2

Rec 33.5 43.4 37.5 44.9 53.6 47.4

F1 46.3 51.8 48.4 52.5 55.3 53.0

Table 3: Phrase pair extraction accuracy for phrase pairs up to length 5. “grow-diag” denotes the grow-diag-final heuristic.

sition algorithm does promote agreement between the two models. We can measure the agreement between models as the fraction of alignment links in the union A∪ that also appear in the intersection A∩ of the two directional models. Table 1 shows a 47% relative increase in the fraction of links that both models agree on by running dual decomposition (bidirectional), relative to independent directional inference (baseline). Improving convergence rates represents an important area of future work. 5.3

Alignment Error Evaluation

To evaluate alignment error of the baseline directional aligners, we must apply a combination procedure such as union or intersection to Aa and Ab . Likewise, in order to evaluate alignment error for our combined model in cases where the inference algorithm does not converge, we must apply combination to c(a) and c(b) . In cases where the algorithm does converge, c(a) = c(b) and so no further combination is necessary. We evaluate alignments relative to hand-aligned data using two metrics. First, we measure alignment error rate (AER), which compares the pro-

AER results for Chinese-English are reported in Table 2. The bidirectional model improves both precision and recall relative to all heuristic combination techniques, including grow-diag-final (Koehn et al., 2003). Intersected alignments, which are one-to-one phrase alignments, achieve the best AER. Second, we measure phrase extraction accuracy. Extraction-based evaluations of alignment better coincide with the role of word aligners in machine translation systems (Ayan and Dorr, 2006). Let R5 (S, P) be the set of phrases up to length 5 extracted from the sure link set S and possible link set P. Possible links are both included and excluded from phrase pairs during extraction, as in DeNero and Klein (2010). Null aligned words are never included in phrase pairs for evaluation. Phrase extraction precision, recall, and F1 for R5 (A, A) are reported in Table 3. Correct phrase pair recall increases from 43.4% to 53.6% (a 23.5% relative increase) for the bidirectional model, relative to the best baseline. Finally, we evaluated our bidirectional model in a large-scale end-to-end phrase-based machine translation system from Chinese to English, based on the alignment template approach (Och and Ney, 2004). The translation model weights were tuned for both the baseline and bidirectional alignments using lattice-based minimum error rate training (Kumar et al., 2009). In both cases, union alignments outperformed other combination heuristics. Bidirectional alignments yielded a modest improvement of 0.2% BLEU4 on a single-reference evaluation set of sentences sampled from the web (Papineni et al., 2002). 4

BLEU improved from 29.59% to 29.82% after training IBM Model 1 for 3 iterations and training the HMM-based alignment model for 3 iterations. During training, link posteriors were symmetrized by pointwise linear interpolation.

As our model only provides small improvements in alignment precision and recall for the union combiner, the magnitude of the BLEU improvement is not surprising.

6

Conclusion

We have presented a graphical model that combines two classical HMM-based alignment models. Our bidirectional model, which requires no additional learning and no supervised data, can be applied using dual decomposition with only a constant factor additional computation relative to independent directional inference. The resulting predictions improve the precision and recall of both alignment links and extraced phrase pairs in Chinese-English experiments. The best results follow from combination via intersection. Because our technique is defined declaratively in terms of a graphical model, it can be extended in a straightforward manner, for instance with additional potentials on c or improvements to the component directional models. We also look forward to discovering the best way to take advantage of these new alignments in downstream applications like machine translation, supervised word alignment, bilingual parsing (Burkett et al., 2010), part-of-speech tag induction (Naseem et al., 2009), or cross-lingual model projection (Smith and Eisner, 2009; Das and Petrov, 2011).

References Necip Fazil Ayan and Bonnie J. Dorr. 2006. Going beyond AER: An extensive analysis of word alignments and their impact on MT. In Proceedings of the Association for Computational Linguistics. Phil Blunsom, Trevor Cohn, Chris Dyer, and Miles Osborne. 2009. A Gibbs sampler for phrasal synchronous grammar induction. In Proceedings of the Association for Computational Linguistics. Peter F. Brown, Stephen A. Della Pietra, Vincent J. Della Pietra, and Robert L. Mercer. 1993. The mathematics of statistical machine translation: Parameter estimation. Computational Linguistics. Jamie Brunning, Adria de Gispert, and William Byrne. 2009. Context-dependent alignment models for statistical machine translation. In Proceedings of the North American Chapter of the Association for Computational Linguistics.

David Burkett, John Blitzer, and Dan Klein. 2010. Joint parsing and alignment with weakly synchronized grammars. In Proceedings of the North American Association for Computational Linguistics and IJCNLP. Fabien Cromi`eres and Sadao Kurohashi. 2009. An alignment algorithm using belief propagation and a structure-based distortion model. In Proceedings of the European Chapter of the Association for Computational Linguistics and IJCNLP. Dipanjan Das and Slav Petrov. 2011. Unsupervised partof-speech tagging with bilingual graph-based projections. In Proceedings of the Association for Computational Linguistics. John DeNero and Dan Klein. 2008. The complexity of phrase alignment problems. In Proceedings of the Association for Computational Linguistics. John DeNero and Dan Klein. 2010. Discriminative modeling of extraction sets for machine translation. In Proceedings of the Association for Computational Linguistics. John DeNero, Alexandre Bouchard-Cˆot´e, and Dan Klein. 2008. Sampling alignment structure under a Bayesian translation model. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Yonggang Deng and Bowen Zhou. 2009. Optimizing word alignment combination for phrase table training. In Proceedings of the Association for Computational Linguistics. Kuzman Ganchev, Joao Grac¸a, and Ben Taskar. 2008. Better alignments = better translations? In Proceedings of the Association for Computational Linguistics. Joao Grac¸a, Kuzman Ganchev, and Ben Taskar. 2008. Expectation maximization and posterior constraints. In Proceedings of Neural Information Processing Systems. Aria Haghighi, John Blitzer, John DeNero, and Dan Klein. 2009. Better word alignments with supervised ITG models. In Proceedings of the Association for Computational Linguistics. Xiaodong He. 2007. Using word-dependent transition models in HMM-based word alignment for statistical machine. In ACL Workshop on Statistical Machine Translation. Philipp Koehn, Franz Josef Och, and Daniel Marcu. 2003. Statistical phrase-based translation. In Proceedings of the North American Chapter of the Association for Computational Linguistics. Terry Koo, Alexander M. Rush, Michael Collins, Tommi Jaakkola, and David Sontag. 2010. Dual decomposition for parsing with non-projective head automata. In Proceedings of the Conference on Empirical Methods in Natural Language Processing.

Shankar Kumar, Wolfgang Macherey, Chris Dyer, and Franz Josef Och. 2009. Efficient minimum error rate training and minimum bayes-risk decoding for translation hypergraphs and lattices. In Proceedings of the Association for Computational Linguistics. Percy Liang, Ben Taskar, and Dan Klein. 2006. Alignment by agreement. In Proceedings of the North American Chapter of the Association for Computational Linguistics. Daniel Marcu and William Wong. 2002. A phrase-based, joint probability model for statistical machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Tahira Naseem, Benjamin Snyder, Jacob Eisenstein, and Regina Barzilay. 2009. Multilingual part-of-speech tagging: Two unsupervised approaches. Journal of Artificial Intelligence Research. Franz Josef Och and Hermann Ney. 2004. The alignment template approach to statistical machine translation. Computational Linguistics. Franz Josef Och, Christopher Tillman, and Hermann Ney. 1999. Improved alignment models for statistical machine translation. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Kishore Papineni, Salim Roukos, Todd Ward, and WeiJing Zhu. 2002. BLEU: A method for automatic evaluation of machine translation. In Proceedings of the Association for Computational Linguistics. Alexander M. Rush, David Sontag, Michael Collins, and Tommi Jaakkola. 2010. On dual decomposition and linear programming relaxations for natural language processing. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Hiroyuki Shindo, Akinori Fujino, and Masaaki Nagata. 2010. Word alignment with synonym regularization. In Proceedings of the Association for Computational Linguistics. David A. Smith and Jason Eisner. 2009. Parser adaptation and projection with quasi-synchronous grammar features. In Proceedings of the Conference on Empirical Methods in Natural Language Processing. Stephan Vogel, Hermann Ney, and Christoph Tillmann. 1996. HMM-based word alignment in statistical translation. In Proceedings of the Conference on Computational linguistics.

Model-Based Aligner Combination Using Dual ... - John DeNero

2 Model Definition ... aligned state αj−1 = 0 are also drawn from a fixed distribution, where D(αj = 0|αj−1 ... define our model only over a and b, where the edge.

609KB Sizes 2 Downloads 241 Views

Recommend Documents

Model Combination for Machine Translation - John DeNero
System combination procedures, on the other hand, generate ..... call sentence-level combination, chooses among the .... In Proceedings of the Conference on.

Unsupervised Translation Sense Clustering - John DeNero
large monolingual and parallel corpora using ..... Their analysis shows that of the wide range of met- rics, only ..... Data-driven semantic anal- ... Mining. Lin Sun and Anna Korhonen. 2011. Hierarchical verb clustering using graph factorization.

Unsupervised Translation Sense Clustering - John DeNero
the synonym groups of WordNet R (Miller, 1995).1 ..... Their analysis shows that of the wide range of met- rics, only BCubed ..... Data-driven semantic anal-.

A Feature-Rich Constituent Context Model for ... - John DeNero
and label x, we use the following feature templates: ... 3 Training. In the EM algorithm for estimating CCM parame- ters, the E-Step computes posteriors over ...

Inducing Sentence Structure from Parallel Corpora for ... - John DeNero
proach, which we call Structure Induction for Re- ordering (STIR) ... a clear application and extrinsic evaluation method- ology. ..... to the problem of inducing probabilistic context-free .... Proceedings of the Conference on Empirical Methods.

Activity Recognition Using a Combination of ... - ee.washington.edu
Aug 29, 2008 - work was supported in part by the Army Research Office under PECASE Grant. W911NF-05-1-0491 and MURI Grant W 911 NF 0710287. This paper was ... Z. Zhang is with Microsoft Research, Microsoft Corporation, Redmond, WA. 98052 USA (e-mail:

Evaluating Treatment Protocols using Data Combination
Nov 22, 2012 - endpoints of the study included death and myocardial infarction (heart attack), and ... endpoints included evaluation of angina and quality of life.

a winning combination
The principles of building an effective hybrid monetization strategy . . . . . . . . . . . . .12. A framework for segmenting users . ... Read this paper to learn: B What an effective hybrid monetization model looks like ... earn approximately 117% mo

Combination pocket article.
'Be' it known that I, STEFAN MACHUGA, a subject ofthe King of Hungary, residing at. Denver, in the county of Denver and State of Colorado, have invented ...

Gas combination oven
Sep 17, 1987 - system, a conventional gas-?red snorkel convection .... ture and other control data. A display 46 shows ..... An alternate gas distribution system.

DOSIMETRY OF 85Kr and 133Xe USING DUAL ...
To assess compliance with regulatory limits of radiation exposure listed in the Code of ... radioactive cloud was reported by Berger (1974) and Kocher (1980).

Minimizing Noise on Dual GSM Channels Using Adaptive Filters - IJRIT
threshold, transmitter power , synchronization scheme, error correction, processing gain, even the number of sun spots, all have effect on evaluating jamming. 2.

Minimizing Noise on Dual GSM Channels Using Adaptive Filters - IJRIT
Jamming makes itself known at the physical layer of the network, more commonly known as the MAC (Media Access Control) layer[4]. The increased noise floor ...

Space-Time Coding in Mobile Satellite Communications Using Dual ...
Space-Time Coding in Mobile Satellite Communications Using Dual-Polarized Channel.pdf. Space-Time Coding in Mobile Satellite Communications Using ...

A4 - Folded Circuit Synthesis: Logic Simplification Using Dual ... - kaist
Abstract— Dual edge-triggered flip-flop (DETFF) captures data at both clock ..... 7, i.e. decreasing area saving of folded circuit over SETFF implementation.

A4 - Folded Circuit Synthesis: Logic Simplification Using Dual ... - kaist
except that cares should be taken in timing analysis [2], clock gating [3], [4], and duty ... receives input data from sequencing elements. A. Motivational Example.

Combination Skyline Queries
replaced by the combinations when a dominance check is invoked. In order to keep the combinations for each rule encountered, we use a matrix M with MBRs as rows and cardinalities columns. As the search order shown in Fig. 5, pattern p1 comes before p

levonorgestrel / ethinylestradiol, ethinylestradiol (combination pack ...
Feb 8, 2018 - 30 Churchill Place ○ Canary Wharf ○ London E14 5EU ○ United Kingdom. An agency of the European Union. Telephone +44 (0)20 3660 6000 Facsimile +44 (0)20 3660 5525. Send a question via our website www.ema.europa.eu/contact. © Europ

a winning combination Services
Read this paper to learn: B What an effective hybrid monetization model ... B How AdMob's in-app purchase house ads can help you use hybrid monetization with minimal effort. Key findings: ... model to look at non-game apps, ads revenue could make up