Online Vocabulary Adaptation Using Contextual Information and Information Retrieval Hagai Aronowitz IBM Haifa Research Labs, Haifa 31905, Israel [email protected]

Abstract This paper presents an algorithm for automatic online vocabulary adaptation based on contextual information and information retrieval. Experiments are presented on a transcription task of spoken annotations of business cards recorded by a hand-held device. Contextual information is used to trigger web search which is used to adapt the vocabulary for a given business card. Finally, the language model for the adapted vocabulary is modified by taking into account the relative value of each context information source. On the business card task, the proposed algorithm reduces 75% of the out-of-vocabulary rate and 16% of the word error rate. Index Terms: vocabulary adaptation, out-of-vocabulary reduction, language model adaptation, OOV

1. Introduction State-of-the-art automatic speech recognition (ASR) systems have been significantly improved during the past few years. For many tasks, acceptable accuracy can be achieved as long as large amounts of acoustic and language training data are used. However, when training data is sparse or does not match the test data, accuracy degrades significantly. This paper addresses the problem of high out-ofvocabulary (OOV) rate. In many real-life applications, a predefined vocabulary based on available corpora has a high OOV rate. This paper proposes a general scheme for dynamically coping with high OOV using available metadata. In order to cope with high OOV, a set of candidate words is selected and added to the vocabulary, and the language model is updated to include the new set of words. The proposed approach is demonstrated on a transcription task of annotated business cards, a task which is part of a larger project called PENSIEVE1 (Post Experiences, Navigate, Share Information, and rEliVE) [1]. The goal of PENSIEVE is to capture data from the physical world and bring them to the virtual world, where they can be transformed into usable knowledge. PENSIEVE takes a social tagging approach, where users intentionally tag objects, people, etc. using a state-of-the-art mobile phone. An example of an event that may be captured by PENSIEVE is meeting a new person. In this case, a business card handed may be scanned and have OCR applied to it; the location, date, and time saved; and a spoken annotation recorded. Another example is taking pictures of posters and flyers, “OCRing” the pictures, and recording a spoken annotation. In this paper we focus on the transcription of spoken annotations for the business card domain.

1 According to J. K. Rowling, the Pensieve is a magical stone basin that stores memories and thoughts.

This paper is organized as follows. In Section 2, related work for coping with OOV is reviewed. In Section 3, the business card task is introduced. In Section 4 the proposed method for coping with OOV is presented. Finally Section 5 concludes.

2. Related work For applications in which the list of OOV words is known at some stage of the application, the problem of high OOV rate can be handled as soon as the knowledge is available. This framework was used by [2] for Spoken Term Detection (STD) in audio archives that used a Large Vocabulary Continuous Speech Recognition (LVCSR) system with a fixed vocabulary to transcribe and index a speech archive. In-vocabulary query terms were retrieved using the transcription-based index, while OOV query terms were retrieved by searching on a phone-based index. Unfortunately, for many applications, explicit knowledge of the OOV words is unavailable. Several studies of exploiting available cues for potential OOV words have been reported in the literature. In [3-6] web crawling was used to retrieve textual data matched to the characteristics of the sparse training data. The retrieved texts were used to statically expand the vocabulary [4, 5] or to statically adapt the language model [3, 6]. Due to the limited power of static adaptation, [7-11] focused on dynamic adaptation of the vocabulary using the text obtained from the first ASR recognition pass as input for web queries. Text obtained from the first ASR recognition pass was used by [9-15] for dynamic language model adaptation. Text retrieval based on the first ASR pass is only one resource for vocabulary adaptation. In [16], metadata regarding the client was used to condition the language model for a customer-service-based task. For the task of lecture transcription, vocabulary and language model adaptation was done by [17] using the text of the projected slides. In [18], slide-based dynamic language model adaptation for lecture transcription was performed by identifying the projected slide and using it as a source for web search. In [19], dynamic language model adaptation for podcast transcription was achieved by exploiting available metadata such as the title and the description of the podcast. Nouns extracted from the metadata were used for web search and dynamic collection of adaptation data. In [20], the vocabulary used by a lecture transcription system was adapted by using available metadata (such as name, topic, and description of a lecture) for web search, and training a neural network to select words from the retrieved web pages for vocabulary expansion. The focus in [20] is the development of a word selection classifier that enables the filtering of large amounts of retrieved texts and selecting only a relatively small number of new words for vocabulary expansion. For a vocabulary of 56K words and an OOV rate of 5%, simply using the

retrieved texts resulted in a relative reduction of 70% in OOV with the cost of tripling the size of the vocabulary (adding 112K words). Instead, the authors of [20] selected only a subset of the retrieved words by using a neural network based on information-retrieval-based features. However, the neural network must be trained on task-dependent training data.

3. Indexing business cards The first application addressed by the PENSIEVE project was organizing personal contacts. A Sony-Ericsson w800i mobile phone was used as a tagging device for both photo and audio capturing. In-house OCR technology was used to extract the textual information from the scanned business cards. Figure 1 shows an image of a business card captured by the phone, followed by the output of the OCR and an automatic semantic annotation system. Figure 2 shows the manual transcription of the corresponding spoken annotation.

3.1. Data and metadata In order to evaluate the performance of the system, 44 business cards were collected, scanned, and OCRed. The business cards were received by researchers and engineers working in IBM Haifa. The recipients’ domains vary from speech processing to image processing, collaboration technologies, social network analysis, and document processing. Most of the business cards were received during international conferences. Correspondingly, spoken annotations were recorded by the persons who actually received the business cards. For each recipient of a business card (a user of PENSIEVE), we stored the full name (“Hagai Aronowitz”), the affiliation (“IBM”), and a short list of key terms (“speaker recognition”, “speaker diarization”, “speech recognition” and “language identification”). In addition, we annotated each business card with a short description of the event where the business card was received. A few examples are: “Interspeech 2007”; “IBM Haifa Research Lab”; and “After his presentation, in MIT”.

3.2. OCR and semantic annotation The OCR’s accuracy for scanned business cards is limited. On our dataset, only 66% of the names were correctly recognized by the OCR, and only 40% were correctly annotated as a person name. The affiliation is correctly recognized only in 29% of the business cards. These results emphasize the importance of incorporating spoken annotations into the PENSIEVE system and complicate the use of the recognized text for online vocabulary adaptation.

3.3. Spoken annotations transcription In our setup, audio was captured, sampled in 8 KHz, and compressed by the mobile phone. The audio was uploaded to a server and decoded using IBM’s state-of-the-art system for decoding English conversational telephony [21]. The system operates in three stages: speaker-independent decoding, adaptation, and speaker-adapted decoding. The speakerindependent decoding produces a 1-best transcript that is used by the adaptation stage to compute a VTLN warping factor, FMLLR transform, and several MLLR transforms (mean only). The speaker-adapted decoding then uses feature-space MPE (fMPE) and MPE trained acoustic models to produce a final 1-best transcript. Adaptation was done by pooling all the spoken annotations of a single speaker.

Name: (86%) Alex Sorin Email: (89%) [email protected] Phone-office: (92%) +97248296289 Phone-mobile: (92%) +972546424125 Figure 1: A scanned business card and the corresponding semantically annotated text after OCR. Estimated confidence levels are shown in parenthesis.

Alex Sorin. He is a developer of ETSI DSR standards. Worked on IBM’s embedded TTS. He is involved in a number of FP7 proposals. Figure 2: A spoken annotation recorded in correspondence to the business card shown in Figure 1. The transcription system uses two language models: a small language model (9M interpolated back-off 3-grams) for constructing static decoding graphs, and a large language model (103M interpolated back-off 4-grams) for in-memory lattice rescoring during the speaker-adapted transcription pass. The language models were trained on a collection of 335M words from the following data sources: 1996 CSR Hub4 Language Model data, EARS BN03 closed captions, GALE Phase 2 Distillation GNG Evaluation Supplemental Multilingual data, Hub4 acoustic model training transcripts, TDT4 closed captions, TDT4 newswire, GALE Broadcast Conversations, and GALE Broadcast News. The recognition vocabulary was chosen to maximize coverage of the training data, and consists of 84,170 words with 1.08 pronunciations in average per word. The system obtains a word error rate (WER) of 16.4% on the RT-04 English conversational telephony dataset. On a dataset collected internally (with a low OOV rate) using the mobile phone, the system obtained a WER of 16.8%. The spoken annotations dataset consists of the 44 annotations described in Subsection 3.1 and re-spoken by four English speakers. The total number of words in the dataset is 4576. The OOV rate of the dataset is 9.8%. The WER obtained by running the transcription system is 42.6% (with a 95% confidence interval of ±1.4%). For reference, we tested a similar 16 KHz-based system on the same audio recorded simultaneously by a close-talk microphone (without compression) and received a WER of 29%. Note also that the speakers have a slight foreign accent.

4. Online vocabulary adaptation In this section, we describe a method to reduce the WER by first identifying potential words that may be OOV and then incorporating these words in a transcription system. Subsection 4.1 describes the OOV reduction algorithm and Subsection 4.2 describes the method used to update the language model to account for the new words.

4.1. Identifying potential OOV words Given the metadata listed in Section 3 (namely OCR output, full name of PENSIEVE owner, affiliation of owner, short list of key terms for owner, and event description), we experimented with two methods of finding potential OOV words. The first method (named “NULL” in [20]) directly adds the OOV words from the metadata to the vocabulary. The second method (named “GREEDY” in [20]) uses the metadata for querying the web using the Yahoo! web search API [22] , and adds to the vocabulary all the OOV words in the top-N (we used N=100) retrieved documents. In any case, the metadata for one business card is used to adapt the vocabulary for that card only. Results for both methods are listed in Table 1. Note that using only the name of the PENSIEVE owner is not as effective as using both the name and the affiliation for a single web search. Using only the full name resulted in an OOV reduction of 2% for the “NULL” approach and 28% for the “GREEDY” approach (compared to 6% and 39%). Table 1. Comparative results for OOV reduction using context metadata information and OCR output. context name + affiliation Key terms Event description All context OCR All context + OCR

NULL OOV reduction (%) 6 2 11 18 16 32

GREEDY OOV reduction (%) 39 30 48 62 24 75

The advantage of using the “NULL” approach is that only a few words are added to the vocabulary (23 on average per business card). However, the reduction in OOV rate is limited (18% reduction using context only, 32% using context and OCR). The “GREEDY” approach is able to reduce the OOV rate much more significantly (62% using context only, and 75% using context and OCR) at the expense of adding many non-OOV words (9.3K words in average per business card). In [20], the “GREEDY” approach was improved by adding a trainable word selection stage in which 80% of the OOV reduction of the “GREEDY” approach was attained, while more then 80% of the non-OOV words were filtered out. Nevertheless, the remaining 20% of the non-OOV words (24.2K words in [20]) are not negligible. Furthermore, the approach taken in [20] requires training data for training the selection algorithm. Finally, it is not clear how to optimize the selection process. For some systems, losing 20% of the OOV words may not be worth the filtering of 80% of the non-OOV words. For other systems, more aggressive filtering may be more beneficial. The approach taken in this paper is to liberally add words to the vocabulary (as in the “GREEDY” approach) and to apply a soft word selection process implicitly in the language model component.

4.2. Updating the language model Given an n-gram w1,..,wn, we denote H=w1,..,wn-1 as the history of word wn and Pr(wn|H) as the n-gram probability. Our objective is to estimate the n-gram probability where the n-gram w1,..,wn may consist of either vocabulary words or new proposed words. The baseline language model we use was trained on the corpora listed in Subsection 3.3 with all OOV words assigned to a single word denoted by . Therefore it assigns n-gram probabilities even to n-grams that contain OOV words. Given an n-gram w1,..,wn, we first replace every new proposed word in the history w1,..,wn-1 with the word . This is motivated by the fact that the particular identity of a new proposed word is of secondary impact when it is in the history part of an n-gram, and furthermore it is hard to estimate it from the available data. Subsequently, if wn is not a new proposed word, the requested n-gram probability is available from the baseline language model. In case wn is indeed a new proposed word w, the n-gram probability is estimated using the model described in equation (1), where V represents the original vocabulary, and θ represents the metadata.

Pr(wn = w H ,θ ) = Pr(wn = w, wn ∉V H ,θ ) =

(1)

Pr(wn ∉V H ,θ )Pr(wn = w wn ∉V , H ,θ ) We assume that the probability of wn to be OOV (the first term in the RHS of equation (1)) is independent of the metadata θ. We further assume that the probability of wn to be a particular OOV word w given θ and wn being OOV (the second term in the RHS of equation (1)) is independent of the history H. The latter is assumed because we lack the data to estimate the dependency. Accordingly, equation (1) reduces to equation (2):

Pr(wn = w H ,θ ) ≅ Pr(wn ∉V H )Pr(wn = w wn ∉V ,θ ) (2)

(

)

The term Pr wn ∉ V H in the RHS of equation (2) is the probability of wn being OOV given the n-gram history. The motivation for being able to properly model this term is that most OOV words are nouns, names in particular; therefore, the history H is a direct indicative for wn being OOV. The term Pr wn ∉ V H is given by the baseline language model

(

)

Pr(< OOV > H ) . The second term in the RHS of equation (2),

Pr(wn = w wn ∉V ,θ ) , may be interpreted in general as soft OOV selection. A special case would be hard OOV selection when Pr wn = w wn ∉V ,θ is constant and equal to the

(

)

product of the reciprocal of the number of selected OOV words with the OOV detection probability of the word selection algorithm. In this work, we estimate Pr wn = w wn ∉V ,θ by generalizing the hard selection

(

)

framework. Assuming for instance that the new proposed word w was selected according to the “GREEDY” approach applied for “name+affiliation”, we estimate the term Pr wn = w wn ∉V ,θ as the product of two terms. The first

(

)

term is the reciprocal of the number of new words selected by the same method (“GREEDY”+”name+affiliation”). The

second term is the estimate of the proportion of OOV words that would be detected by the method. Although optimally the detection probability should be estimated from a development set, our experiments show that any reasonable estimate suffices.

7. References [1]

[2]

4.3. ASR results

[3]

The ASR results are listed in Table 2. The baseline WER for the ASR system and datasets described in Section 3 is 42.6%. The WER achieved using the proposed algorithm is 35.9%. To assess the degradation caused by the flaws of the proposed algorithm, namely failing to detect 25% of the OOV and having an OOV false alarm (FA) rate of 9.3K words per card, we conducted two cheating experiments. The first cheating experiment was to add the OOV words missed by the word selection algorithm to the vocabulary (and updating the language model correspondingly). This experiment resulted in a WER of 33.9%. The second cheating experiment was to both add all OOV words missed (as in cheating experiment #1) and to filter out all FAs (and updating the language model correspondingly). This experiment resulted in a WER of 32.8%. Table 2. Comparative ASR results for transcription of spoken annotations. System Baseline Proposed system 100% detection 100% detection 0 FAs

OOV rate (%) 9.8 2.4 0.0 0.0

FA (words/card) 0 9300 9300 0

WER (%) 42.6 35.9 33.9 32.8

[4]

[5] [6]

[7] [8]

[9]

[10]

[11]

[12]

5. Conclusions This paper has examined the task of online vocabulary adaptation using contextual information and information retrieval. The concept of soft word selection has been introduced by liberally adding words to the vocabulary while utilizing the expected value of each proposed word within the framework of the language model. In order to estimate the ngram probabilities of observing a new word, we combined pre-trained n-gram probabilities for observing an OOV with an online estimation of the unigram probability of the specific OOV word given the context. The proposed algorithm requires only few parameters to fix and does not require training data. The proposed algorithm was evaluated on spoken annotations of business cards. Contextual information was found more valuable than OCR output, but the combination of both was found to be best. The combined system achieved a 75% reduction in OOV and 16% reduction in WER. Future work will generalize the methods discussed in this paper for general language model adaptation.

6. Acknowledgments The authors would like to thank the following for their assistance: Eran Berlinsky, Brian Kingsbury, Hagen Soltau, Ron Hoory, Alexander Sorin, Jonathan Mamou, Yaakov Navon, Ella Barkan, and Boaz Ophir.

[13]

[14]

[15]

[16] [17]

[18]

[19]

[20]

[21]

[22]

Y. S. Maarek, N. Marmasse, Y. Navon, and V. Soroka, “Tagging the physical world,” in Proc. of Collaborative Web Tagging Workshop, 2006. J. Mamou, B. Ramabhadran, and O. Siohan, “Vocabulary independent spoken term detection,” in Proc. SIGIR, 2007. I. Bulyko, M. Ostendorf, and A. Stolcke, “Getting more mileage from web text sources for conversational speech language modeling using class-dependent mixtures,” in Proc. HLTNAACL, 2003. S. Schwarm, I. Bulyko, and M. Ostendorf, “Adaptive Language Modeling with Varied Sources to Cover New Vocabulary Items,” in IEEE Transactions on Speech and Audio Processing, 2004. A. Allauzen and J. Gauvain, “Diachronic vocabulary adaptation for broadcast news transcription,” in Proc. of Interspeech, 2005. I. Bulyko, M. Ostendorf, M.-H. Siu, T. Ng, A. Stolcke, and Ö. Çetin, “Web resources for language modeling in conversational speech recognition,” in ACM Trans. on Speech and Language Processing, 5(1): (2007). T. Kemp and A. Waibel, “Reducing the OOV rate in broadcast news speech recognition,” in Proc. ICSLP, 1998. H. Yu, T. Tomokiyo, Z. Wang, and A. Waibel, “New developments in automatic meeting transcription,” in Proc. ICSLP, 2000. B. Bigi, Y. Huang, and R. De Mori, “Vocabulary and language model adaptation using information retrieval,” in Proc. Interspeech, 2004. K. Othsuki, N. Hiroshima, M. Oku, and A. Imamura, “Unsupervised vocabulary expansion for automatic transcription of broadcast news,” in Proc. ICASSP 2005. G. Boulianne, J.-F. Beaumont, M. Boisvert, J. Brousseau, P. Cardinal,C. Chapdelaine, M. Comeau, P. Ouellet and F. Osterrath “Computer-assisted closed captioning of live TV broadcast in French,” in Proc. of Interspeech, 2006. A. Berger and R. Miller, “Just-in-time language modeling,” in Proc. ICASSP, 1998. M. Mahajan, D. Beeferman and X.D. Huang, “Improved topicdependent language modeling using information retrieval techniques,” in Proc. ICASSP, 1999. [L. Chen, J. L. Gauvain, L. Lamel and G. Adda, “Unsupervised language model adaptation for broadcast news,” in Proc. ICASSP, 2003. C. Martins, A. Tiexiera, and J. Neto, “Dynamic language modeling for a daily broadcast news transcription system,” in Proc. ASRU, 2007. M. Bacchiani and B. Roark, “Meta-data conditional language modeling,” in Proc. ICASSP, 2004. H. Yamazaki, K. Iwano, K. Shinoda, S. Furui, and H. Yokota, “Dynamic language model adaptation using presentation slides for lecture speech recognition,” in Proc. Interspeech, 2007. C. Munteanu, G. Penn, and R. Baecker, “Web based language modeling for automatic lecture transcription,” in Proc. Interspeech, 2007. J. Ogata, M. Goto, and K. Eto, “Automatic Transcription for a Web 2.0 Service to Search Podcasts,” in Proc. Interspeech 2007. C.E. Liu, K. Thambiranam, and F. Seide, “Online vocabulary adaptation using limited adaptation data,” in Proc. Interspeech, 2007. S. Chen, B. Kingsbury, L. Mangu, D. Povey, G. Saon, H. Soltau and G. Zweig, “Advances in Speech Transcription at IBM under the DARPA EARS Program,” in IEEE Transactions on Audio, Speech and Language Processing. Vol. 14, No. 5, 2006. Yahoo! web search API. Available online: http://developer.yahoo.co.jp/search

Online Vocabulary Adaptation Using Contextual ...

transcription task of spoken annotations of business cards ... vocabulary for a given business card. .... 335M words from the following data sources: 1996 CSR.

164KB Sizes 2 Downloads 232 Views

Recommend Documents

Online Vocabulary Adaptation Using Contextual ...
PENSIEVE takes a social ... Finally Section 5 ... expand the vocabulary [4, 5] or to statically adapt the ... technologies, social network analysis, and document.

Online Driver's Drowsiness Estimation Using Domain Adaptation with ...
This paper proposes a domain adaptation with model fusion (DAMF) online drowsiness estimation approach using EEG signals. By making use of EEG data ...

LANGUAGE MODEL ADAPTATION USING RANDOM ...
Broadcast News LM to MIT computer science lecture data. There is a ... If wi is the word we want to predict, then the general question takes the following form:.

SELF-ADAPTATION USING EIGENVOICES FOR ...
and Christian Wellekens. 2. 1. Panasonic Speech Technology Laboratory ... half an hour, while the complexity (the number of degrees of freedom) of the speech ...

Feature Adaptation Using Projection of Gaussian Posteriors
Section 4.2 describes the databases and the experimental ... presents our results on this database. ... We use the limited memory BFGS algorithm [7] with the.

Using the contextual redefinition strategy to improve ... - PUCV Inglés
The whole class sat the test and the score average was 34 (see Appendix E: Vocabulary Size Test. Scores), which ..... Retrieved from http://ejournal.upi.edu/index.php/L-E/article/view/583 ... http://181.112.224.103/bitstream/27000/3081/1/T-UTC-4018.p

IELTS Speaking- Vocabulary Brainstorming - Using English
IELTS Speaking Part One and Two Vocabulary Brainstorming and Speaking. Write as ... uni(versity), graduate school, (technical/ vocational) college ... year degree/ university degree/ undergraduate degree, junior college degree/ two year de-.

Patents Vocabulary and Discussion - Using English
3 types of patent (business method patents, chemical patents, patent protection for pharmaceutical inventions, patenting of genes, software patents) ... 8 one-year “grace” period for filing inventions (= inventors can use the application in pub-.

Academic Word List Vocabulary- Trends - Using English
correctly or not. academic interest in popular culture alternative energy ... exposure of companies to the world economy gender inequality importance of design.

Contextual term equivalent search using domain-driven ...
contextually appropriate term record is necessary to access the proper equivalent to be used in the target language text. The term bank TERMIUM Plus R , recently published as an open access repository, is chosen as a domain-rich resource for testing

Using the contextual redefinition strategy to improve ... - PUCV Inglés
students, and then the way in which the English subject is addressed in the school. .... observe those characteristics in the 10th graders I am teaching when I wanted to introduce a simple ..... Next, the students will have to be able to analyse the

Discovering Contexts and Contextual Outliers Using ...
outlier detection using the stationary distribution is a special case of our approach ... notion of a random walk graph, which is essentially a homogeneous Markov.

speaking rate adaptation using continuous frame rate ...
Most HMM-based ASR systems use a fixed frame rate and window ..... rd e rro r rate. (%. ) baseline normalized. Table 1. SI decoding results on GALE mandarin ...

Context-Based Adaptation of Mobile Phones Using ...
emerging technologies based on near-field communication. (NFC) and takes advantage of smart environments. NFC is a short-range wireless connectivity ...

Feature Adaptation Using Projection of Gaussian ...
We also see significant gains on top of .... best performance is obtained with about 64 Gaussians. ... In this paper we introduced a non-linear adaptation tech-.

Self-Adaptation Using Eigenvoices for Large ... - Semantic Scholar
However, while we are able to build models for, say, voice ... the system, then for all , we can write ... voice model on an arbitrary concatenation of speech seg-.

speaking rate adaptation using continuous frame ... - Semantic Scholar
This paper describes a speaking rate adaptation technique for auto- matic speech ... additional training, sometimes of multiple systems [5]-[7]. In con- trast, the ...

Fast Speaker Adaptation Using A Priori Knowledge
tion techniques could be applied to SD models to find a low-dimen- sional representation for speaker space, the .... 3.1. Conventional vs. Eigenvoice Techniques. We conducted mean adaptation experiments on the Isolet database. [l], which ..... “Pro

Contextual Advertising
Jun 18, 2011 - social networks alike embrace advertising as their preferred ..... 10Specifically, the intermediary determines the winner by ranking PPCi ∗ CTRi.

Extracting Contextual Evaluativity
(§4) strongly support a compositional approach to contextual evaluativity inference. A simple com- ... Note that for this application, we may simplify the compositional picture and treat functors as .... 2http://developer.yahoo.com/search/.

Contextual Contact Retrieval
relationship management software, the system may look-up contacts on your ... based on the user's context (e.g., the content of a displayed email message or ...