Automated Evaluation of Machine Translation Using SVMs Clint Sbisa

EECS Undergraduate Student Northwestern University

[email protected]

ABSTRACT

1.

Machine translation is a challenging field in which evaluation of translation quality is still best done by hand. Automated approaches to this would be better due to the enormous amounts of text available, as manual evaluation takes much more time than an automated approach would. There have been metrics developed that reported high correlations to human judgment such as BLEU, NIST, and METEOR. In this paper, using a SVM1 to learn the classification of automated translation using several different features extracted between the candidate and reference translations is proposed.

Evaluation of machine translation output is a significant part of improving machine translation systems. In order to identify problems in machine translation systems, the output is evaluated by a expert in order to determine whether that translation could be considered “good” or “bad.” However, human evaluation uses large amounts of time and manpower that is not reusable. Thus, an automated approach to this would be significantly better for the development of machine translation systems.

The data was extracted from one of the corpora in OPUS to obtain reference translations and the other sentence was fed through Google Translate to obtain the candidate translation. For the feature set, several metrics, including n-gram co-occurrences and word error rate, are used. After manual evaluation of the candidate and reference translations, the SVM was trained on the test data set to split the data set into good and bad translations. The trained classifier reported about 68% accuracy, which did not outperform other na¨ıve classifiers that depended on a single feature, which ranged from 50% to 70%.

Categories and Subject Descriptors I.2.7 [Artificial Intelligence]: Natural Language Processing—Machine translation

General Terms Experimentation, Languages

Keywords Classification, evaluation, machine translation, support vector machines 1

Support vector machine

INTRODUCTION

So how does one determine the quality of a translation? Given a reference sentence that is considered a good translation, one can compare the candidate translation being evaluated to that reference sentence. Thus, a basis of an automated evaluation system is a metric that quantifies the similarity between the candidate and reference translations. If it correlates highly with human judgment, then the metric can be considered a good measure of translation quality. However, this may prove to be as difficult as machine translation itself, although those approaches does not delve into language itself, as it simply uses numeric features that are extracted from the differences between the candidate and reference translations. Instead of other approaches that use a single metric to correlate to human judgments, a classifier utilizing several metrics measuring similiarity between two target texts will be used. A SVM will be used with the feature set of the metrics extracted from the candidate and reference translations to train a classifier to determine the translation quality of the candidate translation. If this SVM with the particular feature set is a good measure of translation quality, then it will be able to outperform other classifiers when predicting whether a translation is “good” or “bad.”

2.

RELATED WORK

Many attempts at automating evaluation of machine translation has been made. BLEU was one of the first metrics that reported high correlation with human judgment. [4] It uses a modified n-gram precision metric, matching both shorter and longer segments of words between the candidate and reference translations. The reason for using a wide range for n was to measure both adequacy and fluency of translations. Shorter n-grams measured the adequacy of the translation, while longer n-grams measured the fluency of the translation. The metric also incorporated features such as sentence length to penalize translations that may

otherwise be considered a good translation.

• n-gram co-occurrence rate for 1 < n < 5: This is a popular metric for machine translation, made popular by BLEU. [4] The matching n-grams will give us a good measure of similarity between the translations– lower n-grams for adequacy, and higher n-grams for fluency.

NIST is another metric, based on the BLEU metric, except it takes into account the content of the n-grams. [2] Unlike BLEU, which weighs all n-grams equally, the NIST metric takes the particular informativeness of each n-gram and weighs those n-grams accordingly. Thus, the n-grams that appear less frequently would be scored higher, as opposed to the n-grams that appear very frequently. It also reduced the brevity penalty used by BLEU, in order to reduce the impact of variations in length on the overall score. Compared to BLEU, the NIST metric correlated to human judgment better for adequacy, but only scored better on one corpus for fluency. METEOR attempts to address the problems with the BLEU, and the derived NIST, metrics such as the lack of recall and lack of explicit word-matching between the translations. [1] METEOR attempted to alleviate those problems with its own metric, which takes into account both precision and recall. As opposed to precision, recall takes into account the amount of n-grams in the reference translation, while precision takes into account the amount of n-grams in the candidate translation. It first generates an alignment between the two target sentences. This mapping between unigrams takes into account longer segments shared between the sentences and chunks them together. The number of chunks is used in computing a penalty for the translation– penalizing translations that do not share many adjacent words. This metric reported even higher correlations with human judgment than BLEU and NIST. There has yet to be a system with an even higher correlation with human judgment that utilizes multiple features in a classifier. That is the purpose of this paper– to see the if this approach is practical compared to the previously mentioned methods of evaluating translation quality.

3. 3.1

METHOD Data

About 200 sentence pairs were extracted from one of the corpora located at OPUS, a multilingual parallel corpus that uses text extracted from various open source projects. [5] To obtain the candidate translation being evaluated, a sentence in the pair was translated into the target language using the Google Translate site. The candidate translation was then evaluated with respect to the reference translation to the two questions: “Are those two sentences similar in meaning?” and “Does the candidate translation make sense?” The answers to the two questions were then generalized to “yes” if the both questions were answered affirmatively or “no” if either question was answered negatively. Those evaluations served as the basis for human judgments in the classification.

3.2

Features

In order to gather features to be used by the classifier, several scripts were used to extract metrics from the sentence pairs. For the feature set of the classifier, the following metrics were extracted.

• Relative difference in string lengths: This is fairly good metric for determining the information loss or redundant gain between two translations. A translation that is overly short may be considered high quality due to matching words, so this feature punishes significant differences in length between the two translations. • Longest common word subsequence: Similar translations are more likely to have longer common word subsequences. This is a extension to n-gram co-occurrences, as that is used up to only n = 5. This will reward pairs that have long common sequences, which is a likely indication of fluency. • Word accuracy rate: The complement of the word error rate, which measures the “distance” between two sentences through insertions, deletions, and substitutions. While it originated in speech recognition, it has also been applied to machine translation evaluation with some success. Those eight metrics were mostly extracted from their use in other classifiers. [1] [2] [4] With those features, the relevant pieces from the metrics with high correlation to human judgment are used, so hypothetically this set of features should make the classifier perform with high accuracy compared to other potential feature sets. It is a feature set with good cover without delving into language-specific features such as importance of particular words, synonyms, and so forth.

3.3

Classification

For classification of translation quality, the SVMlight package [3] was used. The metrics extracted from the translations were used for the features of the support vector machine, and the quantifiers for human judgment were adjusted to positive (+1) for “yes” and negative (-1) for “no.” For the parameters of the SVM, the defaults were used, except for using a polynomial kernel for the kernel. To avoid data bias, leave-one-out validation2 was used. This avoids training data “weighted” towards particular types of data, as it utilizes many slices of data in learning the classification.

3.4

Naïve Classifiers

Other naive classifiers were also used in this experiment as a benchmark for the proposed classification system. A simple classifier for each of the eight features will also be trained in order to see if combining several metrics improves the accuracy of the classification compared to classifiers using a single metric. A decision tree algorithm will be run with each metric to report the na¨ıve classification accuracy for single features. 2

K-fold validation with K=number of samples

Classifier Support vector machine with all features Unigram co-occurrence rate Bigram co-occurrence rate Trigram co-occurrence rate 4-gram co-occurrence rate 5-gram co-occurrence rate Relative difference in string lengths Longest common word subsequence Word accuracy rate

Accuracy 68% 62% 56% 64% 62% 63% 51% 70% 72%

Table 1: Comparison of classifiers

4.

EXPERIMENT

After the classifiers were trained and validated, the accuracy on the data set was reported. Those results are contained in Table 1. For the data gathered, the SVM classifier proposed did not perform optimally compared to the other classifications. It was able to report a higher accuracy than most of the ngram co-occurrence rate classifiers as well as the relative difference in string length classifier. However, it did not outperform the classifiers for longest common word subsequence and word accuracy rate. A potential explanation for the results is that the word accuracy rate and longest common word subsequence metrics represent the classification much more strongly than the other metrics, and the other metrics become “noise” to the classification, thereby reducing its accuracy. The use of other features in metrics tends to be for penalizing translations that score highly on the basic metric, but is not a good translation. In this case, the different metrics were used in their own right, and not used to penalize different metrics. In light of this, a classifier was run with a smaller feature set– consisting of the longest common word subsequence and word accuracy rate metrics. It performed nearly identically to the SVM with the full feature set, indicating that perhaps it does not take the other features as importantly as the two features used in this particular classifier, at least for this particular data set. Indeed, when a classifier with the feature set of all n-gram co-occurrence rates and the relative string length difference, it performs worse than the other SVM models.

5.

FUTURE WORK

A support vector machine approach to this problem is definitely practical, as it nearly matched the best na¨ıve classifier. There were a few problems with the approach, though. A larger corpus is needed in order to more reliably measure the accuracy of this classifier. While 200 is a sizable number, it is nowhere near the thousands of sentences that are used to evaluate other machine translation metrics such as BLEU, NIST, and METEOR, along with the rigid process of evaluating reference translations in order to prevent errors in the data. With access to the corpus used by those metrics, this approach could then be evaluated with a larger data set, and also could be compared to those metrics.

Another potential flaw is the feature set. More metrics could be tested– the ones used in this paper is by no means an exhaustive list of metrics that could be potentially used to evaluate the similarity between a candidate and reference translation. Also, the different feature sets tested in this paper seems to indicate that only two features were fully utilized to learn the classification. Thus, to improve the classification, more relevant feature sets should be used in order to reduce bias towards particular features– as is, it is a rather simple classification based on longest common word sequence and word accuracy rates. In any case, this is merely speculation– actual experiments will have to be conducted with different feature sets for the SVM to see the differences in classification accuracy with the feature set proposed in this paper. If those flaws are addressed in a future approach to this type of project, a better conclusion could be reached for the question of using this type of approach for evaluating machine translation. With that, the effects of the small data and feature set could be examined, and the potential use of SVMs in this field could be better reported.

6.

CONCLUSION

A SVM was used to learn the classification between “good” and “bad” translations given a feature set with several different metrics in order to automate the process of evaluating machine translations. With the data and feature set proposed in this paper, the SVM did not outperform the best of the na¨ıve classifiers, although it came close to matching the performance of the best classifier. However, as is, it is not very useful when the best classifier could just be used instead of the SVM. While this approach did not outperform the best na¨ıve classifiers, it showed the practicality of using support vector machines for automatically evaluating machine translation with respect to reference translations. With a good feature set, it could potentially outperform basic classifiers by utilizing several features and learning potential relations between the different metrics in the feature set. As mentioned in Section 5, different feature sets could improve the classification accuracy for machine translation quality.

7.

REFERENCES

[1] S. Banerjee and A. Lavie. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pages 65–72, 2005. [2] G. Doddington. Automatic evaluation of machine translation quality using n-gram co-occurence statistics. In Proceedings of 2nd Human Language Technologies Conference (HLT-02), pages 128–132, 2002. [3] T. Joachims. Making large-scale SVM learning practical. 1999. [4] K. Papineni, S. Roukos, T. Ward, and W. jing Zhu. BLEU: A method for automatic evaluation of machine translation. In ACL-2002: 40th Annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.

[5] J. Tiedemann. OPUS - an open source parallel corpus. In Proceedings of the 13th Nordic Conference on Computational Linguistics, University of Iceland, Reykjavik, 2003.

Automated Evaluation of Machine Translation Using ...

language itself, as it simply uses numeric features that are extracted from the differences between the candidate and ... It uses a modified n-gram precision metric, matching both shorter and longer segments of words between the candi- .... Making large-scale SVM learning practical. 1999. [4] K. Papineni, S. Roukos, T. Ward, ...

131KB Sizes 1 Downloads 270 Views

Recommend Documents

Automated Evaluation of Machine Translation Using ...
Automated Evaluation of Machine Translation Using SVMs. Clint Sbisa. EECS Undergraduate Student. Northwestern University [email protected].

machine translation using probabilistic synchronous ...
merged into one node. This specifies that an unlexicalized node cannot be unified with a non-head node, which ..... all its immediate children. The collected ETs are put into square boxes and the partitioning ...... As a unified approach, we augment

Improving Statistical Machine Translation Using ...
5http://www.fjoch.com/GIZA++.html. We select and annotate 33000 phrase pairs ran- ..... In AI '01: Proceedings of the 14th Biennial Conference of the Canadian ...

paper - Statistical Machine Translation
Jul 30, 2011 - used to generate the reordering reference data are generated in an ... group to analyze reordering errors for English to Japanese machine ...

Machine Translation Model using Inductive Logic ...
Rule based machine translation systems face different challenges in building the translation model in a form of transfer rules. Some of these problems require enormous human effort to state rules and their consistency. This is where different human l

Automatic Acquisition of Machine Translation ...
translation researches, from MT system mechanism to translation knowledge acquisition ...... The verb-object translation answer sets are built manually by English experts from Dept. of Foreign ... talk business ..... Iwasaki (1996) demonstrate how to

Machine Translation vs. Dictionary Term Translation - a ...
DTL method described above. 4.3 Example query translation. Figure 2 shows an example ... alone balloon round one rouad one revolution world earth universe world-wide internal ional base found ground de- ... one revolution go travel drive sail walk ru

Evaluation of an automated furrow irrigation system ...
crop (63.14 kg/ha/cm) was considerably higher than of conventional method (51.43 kg/ha/cm). Key words ... no need to go to the field at night or any other ...

Exploiting Similarities among Languages for Machine Translation
Sep 17, 2013 - ... world (such as. 1The code for training these models is available at .... CBOW is usually faster and for that reason, we used it in the following ...

The RWTH Machine Translation System
Jun 19, 2006 - We present the statistical machine translation system used by RWTH in the second TC-STAR evaluation. We give a short overview of the system as .... tactically and semantically meaningful sentence-like units, which pass all ...

Model Combination for Machine Translation - Semantic Scholar
ing component models, enabling us to com- bine systems with heterogenous structure. Un- like most system combination techniques, we reuse the search space ...

Exploiting Similarities among Languages for Machine Translation
Sep 17, 2013 - translations given GT as the training data for learn- ing the Translation Matrix. The subsequent 1K words in the source language and their ...

Model Combination for Machine Translation - John DeNero
System combination procedures, on the other hand, generate ..... call sentence-level combination, chooses among the .... In Proceedings of the Conference on.

Development and evaluation of automated systems for ...
Jul 22, 2005 - patterns of chromomsomes, will be quite a useful tool in clinical ... that high-resolution display and visualization of chromosome band patterns with .... and cross-section sequence graph (CSSG) to split touching chromosomes.

Machine Translation Oriented Syntactic Normalization ...
syntactic normalization can also improve the performance of machine ... improvement in MT performance. .... These identification rules were implemented in Perl.

Development of a Machine Vision Application for Automated Tool ...
measuring and classifying cutting tools wear, in order to provide a good ...... already under monitoring, resulting in better performance of the whole system.

Automated Detection of Engagement using Video-Based Estimation of ...
Abstract—We explored how computer vision techniques can be used to detect ... supervised learning for detection of concurrent and retrospective self-reported engagement. ...... [49] P. Ekman and W. V. Friesen, Facial Action Coding System: A ... [On

Statistical Machine Translation of Spontaneous Speech ...
In statistical machine translation, we are given a source lan- guage sentence fJ .... We use a dynamic programming beam search algorithm to generate the ...

Machine Translation of English Noun Phrases into Arabic
Machine Translation of English Noun Phrases into Arabic. KHALED SHAALAN. Computer Science Department,. Faculty of Computers and Information, Cairo Univ.,. 5 Tharwat St., Orman, Giza, Egypt [email protected]. AHMED RAFEA. Computer Science Depa

Statistical Machine Translation of Spontaneous Speech ...
an alignment A is a mapping from source sentence posi- tions to target sentence positions ... alized Iterative Scaling (GIS) algorithm. Alternatively, one can train ...

The Impact of Machine Translation Quality on Human Post-Editing
Apr 26, 2014 - Center for Speech and Language Processing. The Johns Hopkins University ... 2 http://www.casmacat.eu/index.php?n=Main.Downloads. 38 ...

Machine Translation of Arabic Interrogative Sentence ...
Cost: Computer aids to translation can reduce the cost per word of a translation. In addition, the use of MT can result in improvements in quality, particularly in ...

Automated Detection of Engagement using Video-Based Estimation of ...
Abstract—We explored how computer vision techniques can be used to detect engagement while ... supervised learning for detection of concurrent and retrospective self-reported engagement. ...... [Online]. Available: http://msdn.microsoft.com/en-us/l