Journal of Computational Information Systems 9: 20 (2013) 8351–8359 Available at http://www.Jofcis.com

Pivot Probability Induction for Statistical Machine Translation with Topic Similarity ⋆ Yanzhou HUANG 1,2 , Xiaodong SHI 1,2 , Jinsong SU 1,3,∗, Guimin HUANG 4 1 Fujian

Key Lab of the Brain-like Intelligent Systems, Xiamen University, Xiamen 361005, China

2 Department

of Cognitive Science, School of Information Science and Technology, Xiamen University, Xiamen 361005, China 3 School

4 Research

Yidong CHEN 1,2 ,

of Software, Xiamen University, Xiamen 361005, China

Center on Data Science and Social Computing, Guilin University of Electronic Technology, Guilin 541004, China

Abstract Previous works employ the pivot language approach to conduct statistical machine translation when encountering with limited amount of bilingual corpus. Conventional solutions based upon phrase-table combination overlook the semantic discrepancy between the source-pivot corpus and pivot-target corpus and consequently lead to probability estimation inaccuracy for the induced translation rules. In this paper, the latent topic structure of the document-level training data is learned automatically and each phrase translation rule is assigned to a topic distribution. Furthermore, the phrase probability induction is carried out on the basis of the topic similarity, allowing the translation system to consider the semantic relatedness among different rules. Using BLEU as a metric of translation accuracy, we find out our system is capable of achieving a absolute improvement in in-domain test compared with the baseline system. Keywords: Statistical Machine Translation; Topic Similarity; Pivot Phrase; Translation Model

1

Introduction

Statistical machine translation (SMT) is one of the branches of natural language processing and its target is to convert the text from one language to antoher automatically while preserving the semantic information. In order to circumvent the language resource limitations, previous works employ the pivot language approaches to conduct statistical machine translation, including phrase-table combination [1–3], transfer method [2, 3], synthetic method [2, 3] and hybrid method ⋆

Project supported by the Key Technologies R&D Program of China (Grant No. 2012BAH14F03), the Natural Science Foundation of China (Grant No. 61005052), the Fundamental Research Funds for the Central Universities (Grant No. 2010121068), the Natural Science Foundation of Fujian Province, China (Grant No. 2010J01351). ∗ Corresponding author. Email address: [email protected] (Jinsong SU).

1553–9105 / Copyright © 2013 Binary Information Press DOI: 10.12733/jcis8450 October 15, 2013

8352

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

[3]. Considering these methods, their main contributions are to present alternative choices for reducing data sparseness problems and thus improve the performance of the SMT system. Our studies are mainly inspired by the works investigated in [1], which takes advantage of the identical pivot language phrase to construct the source-target (ST) translation model. In this method, the translation probability induction may become inaccurate if the semantic tendency between the source-pivot (SP) and pivot-target (PT) corpus are different. In order to explain these issues more clearly, an example for the forward phrase translation probability induction is sketched out in Fig. 1. RECREATION

SPORT

balle (ball-1)

0.2

bola (ball-1)

RECREATION

SPORT 0.2

Induc on

1.0

balle (ball-1)

ball 0.8

danza (ball-2)

French

bola (ball-1)

0.8

French

danza (ball-2)

English Spanish

Spanish

(A)

(B)

Fig. 1: An example for the forward phrase translation probability induction from French to Spanish The above illustration displays the French-Spanish probability induction using English as the pivot language and the same senses existing in different language are marked in same color, purple and green in this case. In addition, the numbers above the connections represent the forward phrase translation probabilities. “ball-1” and “ball-2” reflect different meanings of the English word “ball” and their concrete definitions are: (1) ball-1: any object in the shape of a sphere. (2) ball-2: a large formal occasion where people dance. In the French-English side, the forward phrase translation probability of rule ⟨balle(ball-1)|||ball⟩ in (A) is 1.0 in that the bilingual corpus is derived from the SPORT news. As to the EnglishSpanish side, the corpus, on the other hand, focus more on RECREATION than SPORT, such that “ball” is more likely to be translated into ⟨danza(ball-2)⟩ than ⟨bola(ball-1)⟩. As a result, when we estimate the translation probability from French to Spanish, as shown in (B), the French word “balle” will tend to has a higher probability to be translated into “danza” than “bola”. In fact, such probability estimation is incorrect because the probability induction should consistent with the semantic relatedness. To overcome this problem, an effective method is to utilize the context information to measure the semantic similarity of the rule pairs. Recently, there have been some works on semantic application of SMT. Chen et al. [4] take advantage of the vector space model to compute the sense similarity and the resulting scores are appended as additional features in the translation model. Xiao et al. [5] use Latent Dirichlet Allocation (LDA) [6] topic model to discover the topic distribution of the translation rules for the purpose of deducing topic-related hypothesis. Su et al. [7] utilize the in-domain monolingual documents which can be obtained more easily than bilingual documents to learn the topic information for translation model adaption. In this paper, we focus on the semantic level research at the stage of pivot probability induction.

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

8353

The main contribution of our method is to consider the semantic relatedness among different rules so as to maintain the semantic consistency for the ST translation model. Our solution does not need extra linguistics knowledge labeled by human investment and the semantic information of the rules are learned automatically.

2 2.1

Using Topic Information for Probability Induction Topic assignment for translation rule

For ease of the domain partiality in SP and PT bilingual corpora, our approach transforms the induction based on pivot phrase translation probability into the induction based on topic similarity of the rules. Without loss of generality, suppose that we are given bilingual ∑m document∪m level ∑ training ∪ sets consisting of SP and PT language pairs which are denoted as ∑ SP = i=1 di and nP T = nj=1 dj respectively, where d denotes a document in document set . Thus the ∑ ∑n ∑(m+n) ∑m ∪ ∑n combination of m = SP ALL SP and P T can be denoted as P T . Intuitively, different rules occurring in similar contexts tend to have similar meanings and for a given translation rule, its semantic tendency is determined by all context in the training data. Here, we refer to the implement [5] which applies topic information in the hierarchical phrase-based SMT system and also define the context of the collected instance I for translation rule r as the whole document. Based on the above analysis, the first step is to capture the latent topic distribution Z = z1 z2 · · · zk of the training documents. However, in a concrete application, the SP and PT corpora can be retrieved from various sources independently and thus we cannot guarantee their word spaces are consistent if the topic models are trained separately. In that case, the topic similarity computation of the rule pair cannot be applied directly since the procedure is carried out based on the presupposition that computed vectors share identical semantics space. Our first response towards the challenge is to establish a topic projection mechanism between these two word space whereas the fact is that the results may inaccurate since the procedure is based on the word alignment which contains certain noisy interference. As a result, in order to force the consistency ∑(m+n) of the word space between SP and PT corpora, we extract all pivot language documents in ALL to train the document-level topic distribution, even though it will take up extra time consuming. Therefore, after collecting all I of rule r, its ith topic probability P (zi |r) can be computed as ∑ count (I, D) P (zi |dp ) ∑ D∈

P (zi |r) = ∑

∗ ∑

zi ∈Z D∈

count (I, D) P (zi |dp )



(1)



∑n ∑ ∑ where D is a bilingual document in the given bilingual corpus ∗ , which can be m P T in SP or th terms of the given language pair. P (zi |dp ) represents ∑ the probability of the i topic of the pivot language document dp in D. Obviously, dp ∈ D ∈ ∗ . In addition, function count(I, D) denotes the frequency of instance I in D. Alg. 1 illustrates the main steps of the topic assignment for the translation rule. For line 1–2, as mentioned above, it is executed to force the topical word space consistency between SP and PT bilingual corpora, and the topic distribution for the pivot language document dp is obtained after Gibbs sampling in topic model training. Accordingly, line 3–13 is implemented to calculate all different topic probability for a given rule r by means of exploring all different

8354

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

documents. ∑k Line 14–16 is designed to complete the normalization for each extracted rule, which makes i=1 P (zi |r) = 1. Therefore, after completing the procedure described in Alg. 1, each extracted rule is assigned to a topic distribution derived from the related documents. Algorithm 1 Topic Assignment for Translation Rule ∑ ∑n Input: SP corpus m SP and PT corpus PT Output: SP and PT translation model with topic distribution ∑(m+n) ∑m ∪ ∑n 1: ALL = SP PT ; ∑(m+n) 2: Discover latent topic distribution for each pivot document dp in ALL ; ∑ 3: for each document pair D in corpus ∗ do 4: Extract all different translation rules as R and count their instance frequency; 5: for each rule r in R do 6: if the translation rule r not emerge before then 7: Initialize P (Z|r) = 0; 8: end if 9: for each topic zi in r do 10: P (zi |r) + = count (I, D) P (zi |dp ); 11: end for 12: end for 13: end for 14: for each topic zi in r do 15: P (zi |r) = ∑P (zPi(z|r)i |r) ; zi 16: end for 17: Return result;

2.2

Translation probability computation

After assigning the topic distribution for each rule in the translation model, the following steps focus on the translation probability computation. Analogously, our approach also contains phrase translation probability induction and lexical translation probability computation. For the purpose of maintaining the semantic consistency of the rules, we adopt the topic similarity to perform the phrase probability induction. Given the SP and PT translation model attached with topic distribution for all rules, each source phrase is taken into account to expand its possible translation in the target side. During the course of our implement, there are two possible cases from which the source phrase is unable to conduct a proper induction to the target side. On the one hand, the pivot phrase may not exist in the PT translation model due to the corpus resource limitation. That is to say, no available bridge can be utilized to establish the connection between source phrase and its target translation. On the other hand, all indexes of the source phrase align to NULL after alignment induction in spite of detecting the identical pivot phrase. By analyzing the actual cases of the problem, we find out the instances mainly center on those phrases containing punctuations and function words such as “the”, “an”, “well”, etc. An example of French-English-Spanish alignment induction about this situation is offered as follow. (1) French-English rule: ⟨,|||,the|||0-0⟩ (2) English-Spanish rule: ⟨,the|||el|||1-0⟩

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

8355

∩ (3) French-Spanish induction: ⟨,|||,the|||0-0⟩ ⟨,the|||el|||1-0⟩ ⇒ ⟨,|||el|||NULL⟩ In this example, the denotation “X1 −X2 ” means the word with index X1 in source side aligns to the word with index X2 in ∩ target side. For the French-Spanish induction, no available alignment intersection, namely (0-0) (1-0)=NULL, can be detected and the rule ⟨,|||el⟩ also displays a incorrect induction. Thus we believe this type of induction is illegal and they will not be taken into account in practice. Consequently, the forward and backward phrase translation probability under a proper phrase induction are defined in Eq. (2) and Eq. (3). ∑ sim (P (Z|rs¯p¯) , P (Z|rp¯t¯)) p¯ ¯ ϕ (t|¯ s) = ∑ ∑ (2) sim (P (Z|rs¯p¯) , P (Z|rp¯t¯)) t¯





sim (P (Z|rt¯p¯) , P (Z|rp¯s¯)) p¯ ¯ ϕ (¯ s|t) = ∑ ∑ sim (P (Z|rt¯p¯) , P (Z|rp¯s¯)) s¯

(3)



sim (x, y) = cos (x, y) =

⟨x, y⟩ ∥x∥ ∥y∥

(4)

Where s¯, p¯, t¯ and Z are the source phrase, pivot phrase, target phrase and topic distribution respectively. Also, P (Z|rs¯p¯) = P (Z|rp¯s¯) and P (Z|rp¯t¯) = P (Z|rt¯p¯). sim(x, y) is the similarity function which is defined as the cosine distance (see Eq. (4)). Equally, the denominator in Eq. (2) and Eq. (3) are to normalize the phrase translation probability both in forward or backward ∑ ∑ ¯ ¯ s|t) = 1. direction, namely making t¯ P (t|¯ s) = 1 and s¯ P (¯ In view of the lexical translation probability, we also use the phrase method proposed by Wu et al. [1], which is proved to achieve better performance than word method in phrase-based SMT system [8] since it strengthens the frequently aligned pairs and weakens the infrequently aligned pairs. Let’s go back to the former example discussed in Fig. 1, we can apply topic similarity to induce the phrase translation probabilities P (ball|balle) = 0.7538 and P (danza|balle) = 0.2462. Obviously, the induced probability is consistent with their semantics tendency.

3

Experiment

In this section, we conduct the translation from French to Spanish using English as the pivot language and what we are focus is to use topic similarity to improve the probability estimation for those bilingual corpora with different semantic tendency. In order to determine the robustness and effectiveness of our method, the presented model is experimented not only on single source but also on multiple sources data sets.

3.1

Experimental preparation

The Europarl corpus [9], defined as our single source data sets, collects the european parliament multilingual texts for statistical machine translation. In order to compare performance of the baseline system [1], we also select the same language mode (3-gram), development set and test

8356

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

set in the shared task of the SMT workshop [10]. Table 1 shows the details of the data information. Particularly, “FR”, “EN”, “ES” denotes “French”, “English”, “Spanish” respectively and “M” is short for million. Also, the in-domain test set implies the domain of the test sentences is consistent with the data sets in training set. On the contrary, the out-domain test sentences is derived from a totally new source of which the topic is nothing to do with the training set. Certainly, applying the translation model to out-domain test will confront with more challenges than in-domain test. Table 1: Data information of the single source experiment data

document

sentence

source word

target word

FR-EN train

3,424

1M

30.20M

27.20M

EN-ES train

3,465

1M

27.30M

28.40M

development

/

2,000

62,265

60,233

in-domain test

/

2,000

65,098

55,785

out-domain test

/

1,064

31,220

26,969

Table 2 shows the data information about the multiple sources derived from OPUS [11], JRC [12] and Europarl [9]. In addition, ECB, KDE4 and Subtitle are the sub-corpora of OPUS [11] and “K” is short for thousand. Also, each corpus employed here is released in the form of documentlevel files. We use SRILM toolkit [13] to train a 4-gram language model for target language ES and select 4000 FR-ES sentences pair, 2000 for development and 2000 for test, in accordance with the multiple sources data sets. In other words, each source sentence in our development and test set has one reference. Table 2: Data information of the multiple sources experiment data

source

document

sentence

source word

target word

ECB

953

135.20K

4.50M

4.0M

KDE4

853

68.80K

1.40M

1.20M

Subtitile

654

201.40K

1.70M

1.90M

JRC

5,649

200K

6.90M

6.40M

Europarl

707

200.70K

6.10M

5.50M

ECB

842

77K

2.20M

2.50M

KDE4

892

75.70K

1.40M

1.50M

Subtitile

636

201.30K

2.10M

1.70M

JRC

5,725

200K

6.20M

7.0M

Europarl

718

201.10K

5.30M

5.70M

Development

mix

/

2,000

63,145

60,739

Test

mix

/

2,000

63,736

61,651

FR-EN train

EN-ES train

To perform the phrase-based SMT, we choose MOSES [14] as the experimental decoder which is a open-source system used extensively in machine translation. During the stage of the decoding,

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

8357

we set the ttable-limit as 20, the stack-size as 100, and the initial language model weight as 0.50. Minimum-error-rate training [15, 16] is performed to tune the feature wights in log-linear model. The translation quality is evaluated by a well-established automatic measure: BLEU [17] score.

3.2

Experimental result

As to single source test, the primary task is to determine whether the topic information can improve the performance of the SMT system when comparing with the baseline system. Therefore, the topic number is assigned as 50 experimentally and the experiment results are described in Table 3. Table 3: Single source comparison using 50 topics method

development

in-domain

out-domain

baseline

44.15

33.72

16.73

44.93

+34.36

-16.71

our method

50 topics

According to the results, we find out the BLEU score in our method achieves a improvement of 0.64 over the baseline system in in-domain test. Besides, the out-domain test results show that the topic information has a negative impact on the performance of the system. Intuitively, the tested results are not enough to deduce a reliable conclusion since the uncertainty of the topic number. Thus we conduct a experiment to investigate the effect of different topic numbers on translation quality. Table 4: Effect of different topic numbers in single source experiment our method

development

in-domain

out-domain

20 topics

45.37

+34.45

-16.26

30 topics

45.32

+34.46

-16.42

40 topics

45.15

+34.39

-16.18

50 topics

44.93

+34.36

-16.71

60 topics

45.28

+34.35

-16.33

70 topics

45.25

++34.69

+17.16

80 topics

45.24

+34.39

-16.21

In Table 4, we can find that our method achieve a better result than the baseline system for all different topic number during the stage of the in-domain test. Specifically, the best performance can be obtained in topic 70 with BLEU score 34.69. According to the result, we can infer that using the topic information for pivot language induction has a positive impact on the in-domain test. On the other hand, the result of the out-domain test seems unstable and the major tendency illustrates that topic information cannot improve the baseline system. The main reason is that the topics are learned from the in-domain training set and therefore the probability distribution in our translation model tends to explain well for in-domain data, which is certainly incompatible with the out-domain test set. From another perspective, the invalidation of the topic information

8358

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

for out-domain test further proves the effectiveness of our topic information in in-domain test. In fact, we cannot guarantee a statistical model both work well on the in-domain and out-domain data sets since the model reach a better performance for in-domain test at the expense of its generalization ability. In order to determine the robustness and effectiveness of our model for in-domain test set, we also apply our method on the multiple sources experiment and the result is offered in Table 5. Table 5: multiple sources experiment result method

development

test

baseline

52.45

38.67

20 topics

53.23

+39.20

30 topics

53.32

+39.39

40 topics

53.61

+39.17

50 topics

53.40

+39.30

60 topics

53.57

+39.21

70 topics

53.48

++39.63

80 topics

53.30

+39.51

our method

Similarly, we can see that our method outperforms the baseline for all different topic numbers. The best topic number is 70 and its BLEU can reach 39.63 which has a improvement of 0.96 compared with the baseline system. Therefore, based on the former experiment results, we can conclude that using topic information for pivot language induction between SP and PT translation model can improve the SMT system in in-domain test.

4

Conclusion

In this paper, we present a novel method for pivot language induction by incorporating topic similarity information into the phrase probability estimation for phrase-based statistical machine translation system. Our approach firstly uses the topic model to discover the latent topic structure for the pivot language document and each synchronous pivot phrase rule is then attached with the topic distribution according to the document it emerged. Secondly, the probability induction is executed by means of computing the topic similarity for the rule pair instead of using simple phrase translation probability multiplication between SP and PT translation model. As to the in-domain test, our solution outperform the baseline system both in single source and multiple sources data sets for all different topic numbers and thus we can imply that the topic information is useful in probability induction. On the other hand, as the topic distribution extracted from the in-domain data sets is incompatible with the out-domain data sets, the topic information is not applicable in out-domain test and finally degrade the system performance. In the future, we will try to incorporating the topic adaption technique into the course of the probability induction. For example, the monolingual corpus can be retrieved more conveniently and efficiently than bilingual data sets and such available resources can be applied to expand the

Y. Huang et al. /Journal of Computational Information Systems 9: 20 (2013) 8351–8359

8359

topic coverage for our constructed model. In addition, how to detect a reasonable estimation of topic number for better translation in our method will become another study emphasis.

References [1]

H. Wu and H. F. Wang, Pivot Language Approach for Phrase-Based Statistical Machine Translation, in: Proc. ACL’07, 2007, pp. 856-863.

[2]

M. Paul, H. Yamamoto, E. Sumita and S. Nakamura, On the Importance of Pivot Language Selection for Statistical Machine Translation, in: Proc. NAACL-HLT’09, 2009, pp. 221-224.

[3]

H. Wu and H. F. Wang, Revisiting Pivot Language Approach for Machine Translation, in: Proc. ACL’09, 2009, pp. 154-162.

[4]

B. X. Chen, G. Foster and R. Kuhn, Bilingual Sense Similarity for Statistical Machine Translation, in: Proc. ACL’10, 2010, pp. 834-843.

[5]

X. Y. Xiao, D. Y. Xiong, M. Zhang, Q. Liu and S. X. Lin, A Topic Similarity Model for Hierarchical Phrase-based Translation, in: Proc. ACL’12, 2012, pp. 750-758.

[6]

D. M. Blei, Latent Dirichlet Allocation, Journal of Machine Learning 3 (2003) 993-1022.

[7]

J. S. Su, H. Wu, H. F. Wang , Y. D. Chen, X. D. Shi, H. L. Dong and Q. Liu, Translation Model Adaptation for Statistical Machine Translation with Monolingual Topic Information, in: Proc. ACL’12, 2012, pp. 459-468.

[8]

P. Koehn, F. J. Och and D. Marcu, Statistical phrase-based translation, in: Proc. HLT-NAACL’03, 2003, pp. 127-133.

[9]

P. Koehn, Europarl: A Parallel Corpus for Statistical Machine Translation, in: Proc. MT Summit X’05, 2005, pp. 79-86.

[10] P. Koehn and C. Monz, Manual and Automatic Evaluation of Machine Translation between European Languages, in: Proc. the 2006 HLT-NAACL Workshop on Statistical Machine Translation’06, 2006, pp. 102-121. [11] J. Tiedemann, Parallel Data, Tools and Interfaces in OPUS, in: Proc. LREC’12, 2012, pp. 22142218. [12] S. Ralf, B. Pouliquen, A. Widiger, C. Ignat, T. Erjavec, D. Tufi¸s and D. Varga, The JRC-Acquis: A multilingual aligned parallel corpus with 20+ languages, in: Proc. LREC’06, 2006, pp. 24-26. [13] A. Stolcke, Srilm - An Extensible Language Modeling Toolkit, in: Proc. ICSLP’02, 2002, pp. 901-904. [14] P. Koehn, H. Hoang, A. Birch, C. Callison-Burch, M. Federico, N. Bertoldi, B. Cowan, W. Shen, C. Moran, R. Zens, C. Dyer, O. Bojar, A. Constantin and E. Herbst, Moses: Open source toolkit for statistical machine translation, in: Proc. ACL’07, 2007, pp. 177-180. [15] F. J. Och, Minimum Error Rate Training for Statistical Machine Translation, in: Proc. ACL’03, 2003, pp. 160-167. [16] H. S. Linag, M. Zhuang, T. J. Zhao, Forced decoding for minimum error rate training in statistical machine translation, Journal of Computational Information Systems 8 (2012) 861-868. [17] K. Papineni, S. Roukos, T. Ward and W. J. Zhu, BLEU: A Method for Automatic Evaluation of Machine Translation, in: Proc. ACL’02, 2002, pp. 311-318.

Pivot Probability Induction for Statistical Machine ... - Semantic Scholar

Oct 15, 2013 - paper, the latent topic structure of the document-level training data ... Keywords: Statistical Machine Translation; Topic Similarity; Pivot Phrase; Translation Model .... By analyzing the actual cases of the problem, we find out the ...

100KB Sizes 3 Downloads 245 Views

Recommend Documents

CALLUS INDUCTION, SOMATIC ... - Semantic Scholar
varieties viz., Coker 310 and Coker 312 were used to assess the somatic .... Observed data on callus ... callus induction, top six hormonal combinations viz.,.

CALLUS INDUCTION, SOMATIC ... - Semantic Scholar
The best medium for proliferation of callus was standardized as MS + 0.1 mgl-1 2,4-D + ... different combinations tried, MS + 1.9 gl-1 KNO3 was found suitable for ...

Efficient Search for Interactive Statistical Machine ... - Semantic Scholar
Efficient Search for Interactive Statistical Machine Translation. Franz Josef Och ... Current machine translation technology is not able ..... run on a small computer.

The RWTH Phrase-based Statistical Machine ... - Semantic Scholar
The RWTH Phrase-based Statistical Machine Translation System ... machine translation system that was used in the evaluation ..... OOVs (Running Words). 133.

Model Combination for Machine Translation - Semantic Scholar
ing component models, enabling us to com- bine systems with heterogenous structure. Un- like most system combination techniques, we reuse the search space ...

MACHINE LEARNING FOR DIALOG STATE ... - Semantic Scholar
output of this Dialog State Tracking (DST) component is then used ..... accuracy, but less meaningful confidence scores as measured by the .... course, 2015.

Belief Revision in Probability Theory - Semantic Scholar
Center for Research on Concepts and Cognition. Indiana University ... and PC(A2) > 0. To prevent confusion, I call A2 the ..... If we test a proposition n times, and the results are the same ... Pearl said our confidence in the assessment of BEL(E).

Approximating Discrete Probability Distributions ... - Semantic Scholar
Computer Engineering. University of Illinois ... Enterprise Systems Engineering. University of Illinois ... For Bayesian networks, graphical models are often used to represent ... This work will consider the specific setting where there are random ..

ACOUSTIC MODELING IN STATISTICAL ... - Semantic Scholar
The code to test HMM-based SPSS is available online [61]. 3. ALTERNATIVE ..... Further progress in visualization of neural networks will be helpful to debug ...

ACOUSTIC MODELING IN STATISTICAL ... - Semantic Scholar
a number of natural language processing (NLP) steps, such as word ..... then statistics and data associated with the leaf node needs to be checked. On the other ...

Query Rewriting using Monolingual Statistical ... - Semantic Scholar
expansion terms are extracted and added as alternative terms to the query, leaving the ranking function ... sources of the translation model and the language model to expand query terms in context. ..... dominion power va. - dominion - virginia.

N-gram based Statistical Grammar Checker for ... - Semantic Scholar
Center for Research on Bangla Language Processing, BRAC University, Dhaka, Bangladesh ... text is considered incorrect if the parsing does not suc- ceed.

N-gram based Statistical Grammar Checker for ... - Semantic Scholar
tion in our approach with possible solutions. Keywords: n-gram, grammar checker, ..... tional Conference on Artificial Intelligence (AAAI-. 94), Seattle, Washington ...

Backward Machine Transliteration by Learning ... - Semantic Scholar
Backward Machine Transliteration by Learning Phonetic Similarity1. Wei-Hao Lin. Language Technologies Institute. School of Computer Science. Carnegie ...

Support for Machine and Language Heterogeneity ... - Semantic Scholar
saging protocols or on higher-level remote procedure call systems such as RPC ... manually in the application or with the help of object caching middleware [20].

Support Vector Echo-State Machine for Chaotic ... - Semantic Scholar
Dalian University of Technology, Dalian ... SVESMs are especially efficient in dealing with real life nonlinear time series, and ... advantages of the SVMs and echo state mechanisms. ...... [15] H. Jaeger, and H. Haas, Harnessing nonlinearity: Predic

Wireless Sensor Network for Machine Condition ... - Semantic Scholar
We develop a wireless sensor network for ... important with accelerating advances in technology. ... acquisition systems to a new era of distributed wireless.

Support for Machine and Language Heterogeneity ... - Semantic Scholar
to) wide area networks (WANs), in order to take advantage of resources ... They have some important limitations, however, and their wide acceptance does not necessarily ... have spurred the development, over the past fifteen years, of a large number

Support Vector Echo-State Machine for Chaotic ... - Semantic Scholar
1. Support Vector Echo-State Machine for Chaotic Time. Series Prediction ...... The 1-year-ahead prediction and ... of SVESM does not deteriorate, and sometime it can improve to some degree. ... Lecture Notes in Computer Science, vol.

on the probability distribution of condition numbers ... - Semantic Scholar
Feb 5, 2007 - of the polynomial systems of m homogeneous polynomials h := [h1,...,hm] of ... We will concentrate our efforts in the study of homogeneous.

on the probability distribution of condition numbers ... - Semantic Scholar
Feb 5, 2007 - distribution of the condition numbers of smooth complete ..... of the polynomial systems of m homogeneous polynomials h := [h1,...,hm] of.