2014 IntegratingAnUnsupervisedTransl
- (Durrani et al., 2014) ⇒ Nadir Durrani, Hassan Sajjad, Hieu Hoang, and Philipp Koehn. (2014). “Integrating An Unsupervised Transliteration Model Into Statistical Machine Translation.” In: Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2014).
Subject Headings: OOV Word; Transliteration System, OOV Detection System, OOV Transliteration System.
Notes
Cited By
- Google Scholar: ~ 99 Citations.
Quotes
Abstract
We investigate three methods for integrating an unsupervised transliteration model into an end-to-end SMT system. We induce a transliteration model from parallel data and use it to translate OOV words. Our approach is fully unsupervised and language independent. In the methods to integrate transliterations, we observed improvements from 0.23-0.75 ($\Delta$ 0.41) BLEU points across 7 language pairs. We also show that our mined transliteration corpora provide better rule coverage and translation quality compared to the gold standard transliteration corpora.
1. Introduction
All machine translation (MT) systems suffer from the existence of out-of-vocabulary (OOV) words, irrespective of the amount of data available for training. OOV words are mostly named entities, technical terms or foreign words that can be translated to the target language using transliteration.
Much work (Al-Onaizan and Knight, 2002; Zhao et al., 2007; Kashani et al., 2007; Habash, 2009) has been done on transliterating named entities and OOVs, and transliteration has been shown to improve MT quality. Transliteration has also shown to be useful for translating closely related language pairs (Durrani et al., 2010; Nakov and Tiedemann, 2012), and for disambiguation (Hermjakob et al., 2008; Azab et al., 2013). However, despite its utility, a transliteration module does not exist in the commonly used MT toolkits, such as Moses (Koehn et al., 2007). One of the main reasons is that the training data, a corpus of transliteration pairs, required to build a transliteration system, is not readily available for many language pairs. Even if such a training data is available, mechanisms to integrate transliterated words into MT pipelines are unavailable in these toolkits. Generally, a supervised transliteration system is trained separately outside of an MT pipeline, and a naive approach, to replace OOV words with their 1-best transliterations in the post/pre-processing step of decoding is commonly used.
In this work i) we use an unsupervised model based on Expectation Maximization (EM) to induce transliteration corpus from word aligned parallel data, which is then used to train a transliteration model, ii) we investigate three different methods for integrating transliteration during decoding, that we implemented within the Moses toolkit. To the best of our knowledge, our work is the foremost attempt to integrate unsupervised transliteration model into SMT.
This paper is organized as follows. Section 2 describes the unsupervised transliteration mining system, which automatically mines transliteration pairs from the same word-aligned parallel corpus as used for training the MT system. Section 3 describes the transliteration model that is trained using the automatically extracted pairs. Section 4 presents three methods for incorporating transliteration into the MT pipeline, namely: i) replacing OOVs with the 1-best transliteration in a post-decoding step, ii) selecting the best transliteration from the list of $n$-best transliterations using transliteration and language model features in a post-decoding step, iii) providing a transliteration phrase-table to the decoder on the fly where it can consider all features to select the best transliteration of OOV words. Section 5 presents results. Our integrations achieved an average improvement of $0.41$ BLEU points over a competitive baseline across 7 language pairs (Arabic, Bengali, Farsi, Hindi, Russian, Telugu and Urdu-into-English). An additional experiment showed that our system provides better rule coverage as opposed to another built from gold standard transliteration corpus and produces better translations.
2. Transliteration Mining
The main bottleneck in building a transliteration system is the lack of availability of transliteration training pairs. It is, however, fair to assume that any parallel data would contain a reasonable number of transliterated word pairs. Transliteration mining can be used to extract such word pairs from the parallel corpus. Most previous techniques on transliteration mining generally use supervised and semi-supervised methods (Sherif and Kondrak, 2007; Jiampojamarn et al., 2010; Darwish, 2010; Kahki et al., 2012). This constrains the mining solution to language pairs for which training data (seed data) is available. A few researchers proposed unsupervised approaches to mine transliterations (Lee and Choi, 1998; Sajjad et al., 2011; Lin et al., 2011). We adapted the work of Sajjad|Sajjad et al. (2012) as summarized below.
Model
The transliteration mining model is a mixture of two sub-models, namely: a transliteration and a non-transliteration sub-model. The idea is that the transliteration model would assign higher probabilities to transliteration pairs compared to the probabilities assigned by a non-transliteration model to the same pairs. Consider a word pair $(e, f)$, the transliteration model probability for the word pair is defined as follows:
$p_{t r}(e, f)=\sum_{a \in A l i g n(e, f)} \prod_{j=1}^{|a|} p\left(q_{j}\right)$
where $Align(e, f)$ is the set of all possible sequences of character alignments, $a$ is one alignment sequence and $q_j$ is a character alignment.
The non-transliteration model deals with the word pairs that have no character relationship between them. It is modeled by multiplying source and target character unigram models:
$p_{n t r}(e, f)=\prod_{i=1}^{|e|} p_{E}\left(e_{i}\right) \prod_{i=1}^{|f|} p_{F}\left(f_{i}\right)$
The transliteration mining model is defined as an interpolation of the transliteration sub-model and the non-transliteration sub-model:
$p(e, f)=(1-\lambda) p_{t r}(e, f)+\lambda p_{n t r}(e, f)$
$\lambda$ is the prior probability of non-transliteration.
The non-transliteration model does not change during training. We compute it in a pre-processing step. The transliteration model learns character alignment using expectation maximization (EM). See Sajjad et al. (2012) for more details.
3. Transliteration Model
Now that we have transliteration word pairs, we can learn a transliteration model. We segment the training corpus into characters and learn a phrase-based system over character pairs. The transliteration model assumes that source and target characters are generated monotonically[1]. Therefore we do not use any reordering models. We use 4 basic phrase-translation features (direct, inverse phrase-translation, and lexical weighting features), language model feature (built from the target-side of mined transliteration corpus), and word and phrase penalties. The feature weights are tuned [2] on a devset of 1000 transliteration pairs.
4. Integration to Machine Translation
We experimented with three methods for integrating transliterations, described below:
Method 1: involves replacing OOVs in the output with the 1-best transliteration. The success of Method 1 is solely contingent on the accuracy of the transliteration model. Also, it ignores context which may lead to incorrect transliteration. For example, the Arabic word بيل
transliterates to “Bill
” when followed by “Clinton
” and “Bell
” if preceded by “Alexander Graham
".
Method 2: provides n-best transliterations to a monotonic decoder that uses a monolingual language model and a transliteration phrase-translation table to rescore transliterations. We carry forward the 4 translation model features used in the transliteration system to build a transliteration phrase-table. We additionally use an LMOOV feature which counts the number of words in a hypothesis that are unknown to the language model. Smoothing methods such as KneserNey assign significant probability mass to unseen events, which may cause the decoder to make incorrect transliteration selection. The LM-OOV feature acts as a prior to penalize such hypotheses.
Method 3: Method 2 can not benefit from all in decoding features and phenomenon like reordering. It transliterates Urdu compound بحیرہ عرب
(Arabian Sea
) to “Sea Arabian
", if عرب
is an unknown word. In method 3, we feed the transliteration phrase-table directly into the first-pass decoding which allows reordering of UNK
words. We use the decoding-graph-backoff
option in Moses, that allows multiple translation phrase tables and back-off models. As in method 2, we also use the LM-OOV feature in method 3.[3]
5. Evaluation
Data
We experimented with 7 language pairs, namely: Arabic, Bengali, Farsi, Hindi, Russian, Telugu and Urdu-into-English. For Arabic [4] and Farsi, we used the TED talks data (Cettolo et al., 2012) made available for IWSLT-13, and we used the dev2010 set for tuning and the test2011 and test2012 sets for evaluation. For Indian languages we used the Indic multi-parallel corpus (Post et al., 2012), and we used the dev and test sets provided with the parallel corpus. For Russian, we used WMT-13 data (Bojar et al., 2013), and we used half of the news-test2012 for tuning and other half for testing. We also evaluated on the newstest2013 set. For all, we trained the language model using the monolingual WMT-13 data. See Table 1 for data statistics.
Lang | Traintm | Traintr</sun> | Dev | Test1 | Test2 |
---|---|---|---|---|---|
AR | 152K | 6795 | 887 | 1434 | 1704 |
BN | 24K | 1916 | 775 | 1000 | − |
FA | 79K | 4039 | 852 | 1185 | 1116 |
HI | 39K | 4719 | 1000 | 1000 | − |
RU | 2M | 302K | 1501 | 1502 | 3000 |
TE | 45K | 4924 | 1000 | 1000 | − |
UR | 87K | 9131 | 980 | 883 | − |
Baseline Settings
We trained a Moses system replicating the settings used in competition-grade systems (Durrani et al., 2013b; Birch et al., 2013): a maximum sentence length of 80, GDFA symmetrization of GIZA++ alignments (Och and Ney, 2003), an interpolated Kneser-Ney smoothed 5-gram language model with KenLM (Heafield, 2011) used at runtime, a 5-gram OSM (Durrani et al., 2013a), msd-bidirectional-fe lexicalized reordering, sparse lexical and domain features (Hasler et al., 2012), a distortion limit of 6, 100-best translation options, MBR decoding (Kumar and Byrne, 2004), Cube Pruning (Huang and Chiang, 2007), and the no-reordering-overpunctuation heuristic. We tuned with the k-best batch MIRA (Cherry and Foster, 2012).[5]
Transliteration Miner
The miner extracts transliterations from a word-aligned parallel corpus. We only used word pairs with 1-to-1 alignments.[6] Before feeding the list into the miner, we cleaned it by removing digits, symbols, word pairs where source or target is composed from less than 3 characters, and words containing foreign characters that do not belong to this scripts. We ran the miner with 10 iterations of EM. The number of transliteration pairs (types) extracted for each language pair is shown in Table 1 (Traintr).
Transliteration System
Before evaluating our integrations into the SMT system, we performed an intrinsic evaluation of the transliteration system that we built from the mined pairs. We formed test data for Arabic–English (1799 pairs), Hindi–English (2394 pairs) and Russian–English (1859 pairs) by concatenating the seed data and gold standard transliteration pairs both provided for the Shared Task on Transliteration mining (Kumaran et al., 2010). Table 2 shows precision and recall of the mined transliteration system (MTS).
AR | HI | RU | |
---|---|---|---|
Precision (1-best Accuracy) | 20.0% | 25.3% | 46.1% |
Recall (100-best Accuracy) | 80.2% | 79.3% | 87.5% |
The precision (1-best accuracy) of the transliteration model is quite low. This is because the transliteration corpus is noisy and contains imperfect transliteration pairs. For example, the miner extracted the pair (أستراليا
, Australasia
), while the correct transliteration is “Australia
". We can improve the precision by tightening the mining threshold probability. However, our end goal is to improve end-to-end MT and not the transliteration system. We observed that recall is more important than precision for overall MT quality. We provide an empirical justification for this when discussing the final experiments.
MT Experiments
Table 3 gives a comprehensive evaluation of the three methods of integration discussed in Section 4 along with the number[7] of OOV words (types) in different tests. We report BLEU gains (Papineni et al., 2002) obtained by each method. Method 1 (M1), that replaces OOV words with 1-best transliteration gave an average improvement of + 0.13. This result can be attributed to the low precision of the transliteration system (Table 2). Method 2 (M2), that transliterates OOVs in second pass monotonic decoding, gave an average improvement of + 0.39. Slightly higher gains were obtained using Method 3 (M3), that integrates transliteration phrase-table inside decoder on the fly. However, the efficacy of M3 in comparison to M2 is not as apparent, as M2 produced better results than M3 in half of the cases.
Lang | Test | B0 | M1 | M2 | M3 | OOV |
---|---|---|---|---|---|---|
AR | iwslt11 | 26.75 | +0.12 | +0.36 | +0.25 | 587 |
iwslt12 | 29.03 | +0.10 | +0.30 | +0.27 | 682 | |
BN | jhu12 | 16.29 | +0.12 | +0.42 | +0.46 | 1239 |
FA | iwslt11 | 20.85 | +0.10 | +0.40 | +0.31 | 559 |
iwslt12 | 16.26 | +0.04 | +0.20 | +0.26 | 400 | |
HI | jhu12 | 15.64 | +0.21 | +0.35 | +0.47 | 1629 |
RU | wmt12 | 33.95 | +0.24 | +0.55 | +0.49 | 434 |
wmt13 | 25.98 | +0.25 | +0.40 | +0.23 | 799 | |
TE | jhu12 | 11.04 | -0.09 | +0.40 | +0.75 | 2343 |
UR | jhu12 | 23.25 | +0.24 | +0.54 | +0.60 | 827 |
Avg | 21.9 | +0.13 | +0.39 | +0.41 | 950 |
In an effort to test whether improving transliteration precision would improve end-to-end SMT results, we carried out another experiment. Instead of building a transliteration system from mined corpus, we built it using the gold standard corpus (for Arabic, Hindi and Russian), that we also used previously to do an intrinsic evaluation. We then replaced our mined transliteration systems with the gold standard transliteration systems, in the best performing SMT systems for these languages. Table 4 shows a comparison of performances. Although the differences are small, systems using mined transliteration system (MTS) outperformed its counterpart that uses gold standard transliteration system (GTS), except in Hindi–English where both systems were equal.
AR | HI | RU | |||
---|---|---|---|---|---|
iwslt11 | iwslt12 | jhu12 | wmt12 | iwslt13 | |
MTS | 27.11 | 29.33 | 16.11 | 34.50 | 26.38 |
GST | 26.99 | 29.20 | 16.11 | 34.33 | 26.22 |
In the error analysis we found that the GST system suffered from sparsity and did not provide enough coverage of rules to produce right transliterations. For example, Arabic drops the determiner آل
(al
), but such additions were not observed in gold transliteration pairs. Arabic word جيجابيكسل
(Gigapixel
) is therefore transliterated to “'algegabksl
". Similarly the GST system learned no transliteration pairs to account for the rule "b→p" and therefore erroneously transliterated سبريلوك
(Spurlock
) to “Sbrlok
". Similar observations were true for the case of Russian–English. The rules "a→u" and "y→ε" were not observed in the gold set, and hence харрикейнэ
(hurricane
) was transliterated to “herricane” and
Талботу
(Talbot
) to “Talboty”. This shows that better recall obtained from the mined pairs led to overall improvement.
6. Conclusion
We incorporated unsupervised transliteration mining model into standard MT pipeline to automatically transliterate OOV words without needing additional resources. We evaluated three methods for integrating transliterations on 7 language pairs and showed improvements ranging from 0.23-0.75 (Δ 0.41) BLEU points. We also showed that our mined transliteration corpus provide better recall and overall translation quality compared to the gold standard transliteration corpus. The unsupervised transliteration miner and its integration to SMT has been made available to the research community via the Moses toolkit.
Acknowledgments
We wish to thank the anonymous reviewers and Kareem Darwish for their valuable feedback on an earlier draft of this paper. The research leading to these results has received funding from the European Union Seventh Framework Programme (FP7/2007-2013) under grant agreement no 287658. This publication only reflects the authors' views.
Footnotes
- ↑ Mining algorithm also makes this assumption.
- ↑ Tuning data is subtracted from the training corpus while tuning to avoid over-fitting. After the weights are tuned, we add it back, retrain GIZA, and estimate new models.
- ↑ Method 3 is desirable in cases where the decoder can translate or transliterate a word. For example Hindi word
सीमा
can be translated to “Border
” and also transliterated to name “Seema
". Identifying such candidates that can be translated or transliterated is a challenge. Machine learning techniques (Goldwasser and Roth, 2008; Kirschenbaum and Wintner, 2009) and named entity recognizers (Klementiev and Roth, 2006; Hermjakob et al., 2008) have been used for this purpose. Though, we only focus on OOV words, method 3 can be used if such a classifier/NE tagger is available. - ↑ Arabic and Urdu are segmented using MADA (Habash and Sadat, 2006) and UWS (Durrani and Hussain, 2010).
- ↑ Retuning the transliteration features was not helpful, default weights are used.
- ↑ M-N/1-N alignments are less likely to be transliterations.
- ↑ Note that not all OOVs can be transliterated. This number is therefore an upper bound what can be transliterated.
References
BibTeX
@inproceedings{2014_IntegratingAnUnsupervisedTransl, author = {Nadir Durrani and Hassan Sajjad and Hieu Hoang and Philipp Koehn}, editor = {Gosse Bouma and Yannick Parmentier}, title = {Integrating an Unsupervised Transliteration Model into Statistical Machine Translation}, booktitle = {Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics (EACL 2014)}, pages = {148--153}, publisher = {The Association for Computer Linguistics}, year = {2014}, url = {https://doi.org/10.3115/v1/e14-4029}, doi = {10.3115/v1/e14-4029}, }
Author | volume | Date Value | title | type | journal | titleUrl | doi | note | year | |
---|---|---|---|---|---|---|---|---|---|---|
2014 IntegratingAnUnsupervisedTransl | Philipp Koehn Hieu Hoang Nadir Durrani Hassan Sajjad | Integrating An Unsupervised Transliteration Model Into Statistical Machine Translation | 2014 |