z-logo
open-access-imgOpen Access
Large-scale Word Alignment Using Soft Dependency Cohesion Constraints
Author(s) -
Zhiguo Wang,
Chengqing Zong
Publication year - 2013
Publication title -
transactions of the association for computational linguistics
Language(s) - English
Resource type - Journals
ISSN - 2307-387X
DOI - 10.1162/tacl_a_00228
Subject(s) - computer science , cohesion (chemistry) , constraint (computer aided design) , artificial intelligence , generative grammar , dependency (uml) , discriminative model , natural language processing , word (group theory) , feature (linguistics) , disjoint sets , linguistics , mathematics , chemistry , philosophy , geometry , organic chemistry , combinatorics
Dependency cohesion refers to the observation that phrases dominated by disjoint dependency subtrees in the source language generally do not overlap in the target language. It has been verified to be a useful constraint for word alignment. However, previous work either treats this as a hard constraint or uses it as a feature in discriminative models, which is ineffective for large-scale tasks. In this paper, we take dependency cohesion as a soft constraint, and integrate it into a generative model for large-scale word alignment experiments. We also propose an approximate EM algorithm and a Gibbs sampling algorithm to estimate model parameters in an unsupervised manner. Experiments on large-scale Chinese-English translation tasks demonstrate that our model achieves improvements in both alignment quality and translation quality.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom