N-BEST PARALLEL MAXIMUM LIKELIHOOD BEAMFORMERS FOR ROBUST SPEECH RECOGNITION L. Brayda1 , C. Wellekens

M. Omologo

Institut Eurecom, 2229 Route des Cretes, 06904 Sophia Antipolis, France {brayda,wellekens}@eurecom.fr

ITC-irst, Via Sommarive 18, 38050 Povo (TN), Italy [email protected]

ABSTRACT This work aims at improving speech recognition in noisy environments using a microphone array. The proposed approach is based on a preliminary generation of N-best hypotheses. The use of an adaptive maximum likelihood beamformer (the Limabeam algorithm), applied in parallel to each hypothesis, leads to an updated set of transcriptions, among which the maximally likely to clean speech models is selected. Results show that this method improves recognition accuracy over both Delay and Sum Beamforming and Unsupervised Limabeam especially at low SNRs. Results also show that it can recover the recognition errors made in the first recognition step. 1. INTRODUCTION The use of microphone arrays in Automatic Speech Recognition (ASR) has shown significant improvements in the last years [1, 2]. The problem of coupling array processing with distant talking ASR has been recently addressed [3] due to the observation that improving speech intelligibility does not necessary increase recognition performance to the same extent. Advancements were obtained optimizing a Filter-and-Sum Beamformer through a Maximum Likelihood (ML) criterion: this setup showed improvements over both other adaptive filtering techniques [4] and the traditional Delay-and-Sum (D&S) beamformer [1]. Seltzer [5, 6, 3] proved that, by adapting a set of FIR filters on a set of clean speech models in an unsupervised manner, the input signal could be properly beamformed and the likelihood of test data increased. However, a careful analysis of this unsupervised maximum likelihood beamformer showed (see Section 5) that the improvements obtained over D&S were marginal with respect to the best results obtained with supervised methods. This suggested that when performing some kind of ASR feedback, one should rather consider the N-best hypotheses instead of the best one (in a maximum likelihood sense). In this work we propose and discuss an extension of the Limabeam algorithm. The proposed system applies Nbest parallel beamformers to multi-channel signals; it independently optimizes and recognizes each hypothesized utterance; then it re-scores the likelihoods to provide a new maximally likely transcription. The paper is organized as follows: Section 2 justifies the use of this approach with a microphone array. Section 3 outlines the Limabeam algorithm, Section 4 describes our N-best approach to the Unsupervised Limabeam, and Section 5 gives the experimental results. Finally, Section 6 presents discussion and future activity. 1 This

work was conducted while L. Brayda was at ITC-irst.

2. N-BEST RECOGNITION WITH MICROPHONE ARRAYS N-best recognition is known to be a useful approach when using progressive knowledge sources in multi-pass search strategies [7]. When performing ASR in noisy environments, it is well known that the correct transcription could be found in a very late position in an N-best list. Microphone array processing can be adopted to increase the SNR [8], to compensate for additive noise [4] and convolutional distortion or to maximize the likelihoods of the test set utterances with respect to given clean speech models [3]. Actually an array spans one dimension more (i.e. the spatial dimension) than the usual time-frequency domain spanned by its singlemicrophone counterpart. Thus, one can expect that the use of a microphone array can raise the position of the correct transcription in the N-best list. Let us assume to have a system that outputs the N-best hypotheses for known sentences and thus is able to select the correct one. Figure 1 depicts the performance of the system when one or eight microphones (in this case D&S is applied) are used in our experimental conditions (see Section 5). It is evident that the use of the array increases the amount of correct sentences among the N-best, i.e. the chances of picking up the correct transcription. This happens because all the candidates are more intelligible, less noisy, or better matching the models, depending on the array processing method adopted. At this point one would like the correct transcription be the first choice rather than the nth. One way to do that is to re-process each hypothesis until it better matches the models to which it will be compared. We chose the Limabeam algorithm (see Section 3) to optimize and recognize in parallel each hypothesis (see Section 4), then re-score them and produce a final transcription. 3. THE LIMABEAM ALGORITHM The Limabeam algorithm uses an adaptive filter-and-sum beamformer. Such a beamformer can be represented as follows: x[k] =

M

∑ hm [k] ∗ sm[k]

m=1

(1)

where sm [k] is the discrete time domain signal received at the m-th microphone, hm [k] is the FIR filter for the m-th channel, x[k] is the output of the beamformer, ∗ denotes convolution and k is the time index. The whole set of FIR coefficients of all the microphones can be represented by a super-vector h. For each frame, recognition features can be derived and expressed in function of h:  yL (h) = log10 W | FFT(x(h))|2 (2)

by Word Error Rate (WER), which should be the optimal criterion. Applying the Limabeam algorithm on each hypothesis, the ranking order of the N-best list changes. We show at experimental level that the new hypothesis chosen (the new maximally likely) in this new list has, on average, a lower WER than the first chosen in the old list. The system is described in the following. For each of the N-best hypotheses 2 we derive a set of FIR filters:

100

correct sentences (%)

90

80

70

hˆ n = argmax P(yL (h) |wn ) h

60

50

8 mics 1 mic 10

20

30

40

50

60

70

80

90

100

nbest

Figure 1: Percentage of correct sentences found by a system that recognizes the N-best hypotheses over transcribed sentences in a noisy environment. With more microphones the correct sentence is “pushed up” in the first alternatives. This result represents an upper-bound for system performance reported in the following. where x(h) is the observed vector, | FFT(x(h))|2 is the vector of individual power spectrum components, W is the Mel filter matrix and yL (h) is the vector of the Log Filter Bank Energies (LFBE). Cepstral coefficients are derived via a DCT rotation: yC (h) = DCT (yL (h)) . (3) Limabeam aims at deriving a set of M FIR filters, which maximize the likelihood of yL (h) given an estimated state sequence of a hypothesized transcription. This is expressed by: hˆ = argmax P(yL (h) |w ) (4) h

where w is the hypothesized transcription, P(y(h) |w ) is the likelihood of the observed features given the transcription considered and hˆ is the FIR parameter super-vector derived. The optimization is done via non-linear Conjugate Gradient. The state sequence can be estimated either using the array beamformed output (Unsupervised Limabeam) or, alternatively, assuming that the correct transcription is available (Oracle Limabeam). More details can be found in [5]. The Unsupervised Limabeam works well in noisy environments, even with a single channel. However, we found that preliminary experiments revealed two facts: first, the Oracle Limabeam performance on a single channel was close to the simple D&S on eight channels; second, there was still a margin of improvement between the Unsupervised and the Oracle Limabeam version applied to the multi-channel signals. 4. COMBINING N-BEST RECOGNITION AND ML BEAMFORMING The Limabeam algorithm increases the likelihood of the first hypothesized transcription after a first recognition step. We propose to apply N-best parallel such optimizations: this approach is based on the fact that the N-best list, prior to parallel optimization, is ordered by likelihood and not necessarily

(5)

where wn is the n-th hypothesized transcription at first recognition step, P(y(h) |wn ) is the likelihood of the observed features given the n-best transcription considered. Note that Equation (5) is equivalent to Unsupervised Limabeam when n is 1. After all the N-best FIR vectors are optimized in parallel, new features are calculated and recognition is performed. The transcription which gives the ML is then chosen: nˆ = argmax P(yC (hˆ n ) |wˆ n ) n

(6)

where wˆ n is the transcription generated at second step recognition and nˆ is the index of the most likely transcription, which is wˆ nˆ . Note that the optimization is done in the LFBE domain, while recognition is done in the Cepstral domain as in [3]. We re-score likelihoods in the Cepstral domain as well. The system we propose is depicted in Figure 2. The signal coming from a microphone array is processed via conventional D&S, then Feature Extraction (FE) and a first recognition step is performed (REC). The HMM recognizer generates N-best hypotheses. For each hypothesis and in parallel, the Limabeam algorithm is applied: first a Viterbi alignment is performed (switch to 1: ALIGN) and fixed, then FIR coefficients are adaptively optimized via Conjugate Gradient (switch to 2: OPT). After convergence, the N-best features are recognized (switch to 3: REC) and another set of new transcriptions is produced. Finally, the last block compares the new N-best Log-LikeliHoods (LLH-rescoring) choosing the highest and the recognized sentence is produced. 5. EXPERIMENTAL RESULTS Experiments were conducted using the HTK HMM-based recognizer [9] trained on the clean TI-digits corpus. Word models are represented by 18 state left-to-right HMMs. Output distributions are defined by 1 Gaussian pdf. The training set consists of 8440 utterances, pronounced by 110 speakers (55 men and 55 women). The test set consists of 104 phrases, pronounced by 52 men and 52 women. Both training and test data were up-sampled to 44.1 kHz via a polyphase filter prior to any processing. This sampling frequency was chosen in order to guarantee a high temporal resolution for the FIR filters and to ensure consistency with the real experimental framework being addressed [10]. In this work real noise coming from a computer fan was recorded in an office and added to the clean speech after being properly shifted channel by channel: this resulted in simulating a 8-microphone array with 7 cm inter-microphone distance and a source of fan noise in end-fire position. The FIR filters to be optimized 2 Note that here “N-best” results from a preliminary reduction to a list that does not include repetitions of the same word sequence, which could be caused by different number and location of silences/background noise units.

FE

2-best 1-best

N-best

REC 1

ALIGN

2

3

    

FE

FIR FIR FIR FIR

REC 1

ALIGN

2

OPT

FE

3   

2

REC 1

ALIGN OPT

new 2-best new 1-best

... ... ... ... ... ... ... ... ... N-best optimizer

 

...

OPT

FIR FIR FIR FIR

FIR FIR FIR FIR

3 

FE

FIR FIR FIR FIR

FIR FIR FIR FIR

REC

new N-best

1-best optimizer

FIR FIR FIR FIR

2-best optimizer

eplacements

Delay

    



  

 



            





 



                                           

...

LLH rescoring

We then applied the N-best Unsupervised Limabeam to the same test set. Figure 3 depicts its behavior at -5 dB in terms of digit accuracy against the number of N-best hypothesis considered. The solid line corresponds to D&S performance, while the upper flat line corresponds to the Oracle Limabeam. Except for the 2-best and the 3-best, where a major boosting in accuracy is observed, performance grow roughly linearly. When 19 hypotheses are considered in parallel, the N-best Unsupervised Limabeam is comparable to the Oracle Limabeam, showing a 34.4% relative improvement: in practice we achieved the same performance of a supervised algorithm in an unsupervised manner. The non-monotonic behavior is due to the mismatch between the maximum likelihood and the minimum WER criterion: especially if few hypotheses are considered (e.g 3 or 5 on this task, see Figure 3), after LLH rescoring some new maximally likely transcriptions can have a lower WER. The key is to consider as many hypotheses as possible, because this increases the probability of having a transcription which minimizes the WER as well. Further experiments with higher SNRs did not show significant improvements over D&S: Table 2 shows that the major gain of the N-best Unsupervised Limabeam is at low SNRs.

recognized sentence

82

Figure 2: Block diagram of the N-best Unsupervised Limabeam.

-5 dB Delay and Sum -5 dB Oracle Limabeam -5 dB N-best Uns. Limabeam

80

are 10 taps long. The feature extraction involves 12 Mel Frequency Cepstral Coefficients (MFCC) and the log-Energy together with their first and second derivatives, for a total of 39 coefficients. Features were calculated every 10 ms, using a 25 ms sliding Hamming window. The frequency range spanned by the Mel-Scale filterbank was limited to 100-7500 Hz to avoid frequency regions with no useful signal energy. Cepstral Mean Normalization is applied. While recognition is performed in the cepstral domain, the optimization process is done in the LFBE domain using 24 coefficients for the features and single-Gaussian output distributions [3] for the models, but without CMN. In our implementation of the Limabeam algorithm the Conjugate Gradient algorithm [11] is the same adopted in [3] and no modifications were applied to the original one. This was done to ensure compliance with Seltzer’s work. Preliminary results show the usefulness of a microphone array in a noisy environment. Table 1 reports the performance obtained with D&S, with Unsupervised Limabeam and with Oracle Limabeam (transcription is known). In this case the amount of fan and office noise caused a SNR of -5 dB. The Oracle on one channel performs only slightly better than the simple D&S on eight channels. Furthermore, note that Unsupervised Limabeam gives 16% relative improvement over D&S, while performance could be improved up to a 35%. -5 dB 1 ch 8 ch

D&S 63.79% 69.10%

Uns. Lim. 66.11%(6%) 74.09%(16%)

Oracle Lim. 70.43%(18%) 80.07%(35%)

Table 1: Usefulness of Limabeam with respect to a single microphone. Relative improvements over D&S are reported in parentheses. Results are in terms of digit accuracy.

digit accuracy (%)

78

76

74

72

70

68

0

2

4

6

8

10

12

14

16

18

20

nbest

Figure 3: Performance of the N-best Unsupervised Limabeam compared to D&S and to the supervised version of Limabeam at -5 dB: the N-best approach makes the unsupervised optimization comparable to Oracle Limabeam when N becomes high. For N=1 the system is equivalent to the original Unsupervised Limabeam.

15dB 5 dB -5 dB

D&S 98.34% 95.68% 69.10%

Uns.Lim. 98.34% 96.35% 74.09%

N-best Lim. 98.34% (1) 96.68% (9) 79.73% (19)

Oracle Lim. 98.34% 96.68% 80.07%

Table 2: Comparison among D&S, Limabeam and N-best Unsupervised Limabeam across different SNRs. At higher SNRs the use of a ML-based beamformer does not seem to provide significant improvement, while at low SNRs our Nbest approach performs almost as the Oracle Limabeam. The performance is the one relative to the n-best in parenthesis. Results are in terms of digit accuracy.

6. DISCUSSION AND FUTURE ACTIVITY We found that an N-best approach to Unsupervised Limabeam, which optimizes in parallel a set of FIR filters on the first N-best hypotheses in an unsupervised manner, is able to reach performance comparable to the Oracle Limabeam. This approach is an extension of the Limabeam algorithm, which has the advantage of modifying the FIR coefficients via a ML criterion. One could expect that the N-best hypotheses of the first recognition step should be optimized in the same way and, consequently, their ranks should not change significantly. This is generally not true. In fact an n-best hypothesis can have, after the first recognition step, a lower likelihood with respect to the 1-best, and a better WER at the same time. Multiple instances of the Limabeam algorithm generate different Viterbi alignments, objective functions to be minimized and FIR supervectors: it is our intuition that some of these filters significantly increase (with respect to the first step) the relative likelihoods of the minimal WER hypotheses. This may be due to the fact that the Viterbi alignment is closer to the correct transcription. Further work is planned to clarify this aspect. We also point out that in noisy environments it is not strictly necessary to have the correct transcription, among the Nbest, to generate a more likely transcription, as the literature states [7]. In conclusion, this work showed that performing N-best recognition is a way to improve the potential of Limabeam algorithm. The effectiveness of the proposed solution results from the synergy between a multi-channel signal processing and a multi-pass search strategy, which allows to recover errors introduced in early recognition steps. The robustness of the N-best Unsupervised Limabeam is still to be proved with a larger real database, recorded in reverberant environments, as meeting rooms, where such application would be desirable, but in which the overall performance is likely to decrease. Preliminary experiments on real-data indicated that the margin between D&S and the Oracle is smaller than in the here presented simulation. This is probably related to the regularities characterizing simulations versus variabilities in real-world experiments. As typical meeting room impulse responses are even more than 600ms long, the FIR filters length would be much higher. The temporal resolution of the filters should be guaranteed by the high sampling frequency (44.1 kHz) adopted, which complies with the microphone array we plan to use in our future studies [10]. To this regard, other issues to address in the next work will refer to the dimensionality of the problem (number of taps, number of microphones etc...) and to a possible reduction of the system complexity. 7. ACKNOWLEDGEMENTS The authors would like to thank Michael Seltzer for clarifications about the Limabeam algorithm. L. Brayda would also like to acknowledge the French MESR (Minist`ere de l’Einsegnement Sup´erieur et de la Recherche) and Istituto Trentino di Cultura for having supported this work. REFERENCES [1] M. Brandstein and D. Ward, Microphone arrays - signal processing techniques and applications, New York: Springer-Verlag, 2001.

[2] Acero A., Acoustical and environmental robustness in automatic spech recognition., Boston, MA:Kluwer Academic Publishers, 1993. [3] Seltzer M., Raj B., and Stern R. M., “Likelihoodmaximizing beamforming for robust hands-free speech recognition,” in IEEE Trans. on Speech and Audio Procesing, 2004, vol. 12, no, 5. [4] Bitzer J., Simmer U., and Kammeyer K.D., “Theoretical noise reduction limits of the generalized sidelobe canceler (gsc) for speech enhancement,” in Proceedings of ICASSP, Phoenix, AZ, USA, 1999. [5] Seltzer M., Microphone array processing for robust speech recognition, Ph.D. thesis, Carnegie Mellon University, Pittsburgh, PA, USA, 2003. [6] Seltzer M. and Raj B., “Speech recognizer-based filter optimization for microphone array processing,” in IEEE Signal Processing Letters, 2003, vol. 10, no, 3. [7] Huang X., Acero A., and Hon H., Spoken Language Processing, Carnegie Mellon University, 2001. [8] Flanagan J.L., Surendran A.C., and Jan E.E., “Spatially selective sound capture for speech and audio processing,” Speech Communication, 1993. [9] Young S. et al, The HTK Book Version 3.0., Cambridge University, 2000. [10] Brayda L., Bertotti C., Cristoforetti L., Omologo M., and Svaizer P., “Modifications on NIST MarkIII array to improve coherence properties among input signals,” in AES, 118th Audio Engineering Society Convention, Barcelona, Spain, 2005. [11] Press W. et al, Numerical recipes in C: The Art of Scientific Computing, Cambridge University Press, 1988.

n-best parallel maximum likelihood beamformers for ...

than the usual time-frequency domain spanned by its single- ... 100 correct sentences (%) nbest. 8 mics. 1 mic. Figure 1: Percentage of correct sentences found by a system ..... maximizing beamforming for robust hands-free speech recognition ...

114KB Sizes 1 Downloads 201 Views

Recommend Documents

Blind Maximum Likelihood CFO Estimation for OFDM ... - IEEE Xplore
The authors are with the Department of Electrical and Computer En- gineering, National University of .... Finally, we fix. , and compare the two algorithms by ...

Fast maximum likelihood algorithm for localization of ...
Feb 1, 2012 - 1Kellogg Honors College and Department of Mathematics and Statistics, .... through the degree of defocus. .... (Color online) Localization precision (standard devia- ... nia State University Program for Education and Research.

Properties of the Maximum q-Likelihood Estimator for ...
variables are discussed both by analytical methods and simulations. Keywords ..... It has been shown that the estimator proposed by Shioya is robust under data.

Maximum likelihood: Extracting unbiased information ...
Jul 28, 2008 - Maximum likelihood: Extracting unbiased information from complex ... method on World Trade Web data, where we recover the empirical gross ...

GAUSSIAN PSEUDO-MAXIMUM LIKELIHOOD ...
is the indicator function; α(L) and β(L) are real polynomials of degrees p1 and p2, which ..... Then defining γk = E (utut−k), and henceforth writing cj = cj (τ), (2.9).

Maximum likelihood training of subspaces for inverse ...
LLT [1] and SPAM [2] models give improvements by restricting ... inverse covariances that both has good accuracy and is computa- .... a line. In each function optimization a special implementation of f(x + tv) and its derivative is .... 89 phones.

5 Maximum Likelihood Methods for Detecting Adaptive ...
“control file.” The control file for codeml is called codeml.ctl and is read and modified by using a text editor. Options that do not apply to a particular analysis can be ..... The Ldh gene family is an important model system for molecular evolu

Asymptotic Theory of Maximum Likelihood Estimator for ... - PSU ECON
We repeat applying (A.8) and (A.9) for k − 1 times, then we obtain that. E∣. ∣MT (θ1) − MT (θ2)∣. ∣ d. ≤ n. T2pqd+d/2 n. ∑ i=1E( sup v∈[(i−1)∆,i∆] ∫ v.

MAXIMUM LIKELIHOOD ADAPTATION OF ...
Index Terms— robust speech recognition, histogram equaliza- tion, maximum likelihood .... positive definite and can be inverted. 2.4. ML adaptation with ...

Maximum Likelihood Eigenspace and MLLR for ... - Semantic Scholar
Speech Technology Laboratory, Santa Barbara, California, USA. Abstract– A technique ... prior information helps in deriving constraints that reduce the number of ... Building good ... times more degrees of freedom than training of the speaker-.

Reward Augmented Maximum Likelihood for ... - Research at Google
employ several tricks to get a better estimate of the gradient of LRL [30]. ..... we exploit is that a divergence between any two domain objects can always be ...

A maximum likelihood method for the incidental ...
This paper uses the invariance principle to solve the incidental parameter problem of [Econometrica 16 (1948) 1–32]. We seek group actions that pre- serve the structural parameter and yield a maximal invariant in the parameter space with fixed dime

Maximum Likelihood Detection for Differential Unitary ...
T. Cui is with the Department of Electrical Engineering, California Institute of Technology, Pasadena, CA 91125, USA (Email: [email protected]).

Asymptotic Theory of Maximum Likelihood Estimator for ... - PSU ECON
... 2010 International. Symposium on Financial Engineering and Risk Management, 2011 ISI World Statistics Congress, Yale,. Michigan State, Rochester, Michigan and Queens for helpful discussions and suggestions. Park gratefully acknowledges the financ

Blind Maximum Likelihood CFO Estimation for OFDM ...
vious at low SNR. Manuscript received .... inevitably cause performance degradation especially at lower. SNR. In fact, the ML .... 9, no. 4, pp. 123–126, Apr. 2002.

Maximum likelihood estimation of the multivariate normal mixture model
multivariate normal mixture model. ∗. Otilia Boldea. Jan R. Magnus. May 2008. Revision accepted May 15, 2009. Forthcoming in: Journal of the American ...

Maximum Likelihood Estimation of Random Coeffi cient Panel Data ...
in large parts due to the fact that classical estimation procedures are diffi cult to ... estimation of Swamy random coeffi cient panel data models feasible, but also ...

Maximum Likelihood Estimation of Discretely Sampled ...
significant development in continuous-time field during the last decade has been the innovations in econometric theory and estimation techniques for models in ...

Maximum likelihood estimation-based denoising of ...
Jul 26, 2011 - results based on the peak signal to noise ratio, structural similarity index matrix, ..... original FA map for the noisy and all denoising methods.

maximum likelihood sequence estimation based on ...
considered as Multi User Interference (MUI) free systems. ... e−j2π fmn. ϕ(n). kNs y. (0) m (k). Figure 1: LPTVMA system model. The input signal for the m-th user ...

Small Sample Bias Using Maximum Likelihood versus ...
Mar 12, 2004 - The search model is a good apparatus to analyze panel data .... wage should satisfy the following analytical closed form equation6 w* = b −.

Unifying Maximum Likelihood Approaches in Medical ...
priate to use information theoretical measures; from this group, mutual information (Maes ... Of course, to be meaningful, the transformation T needs to be defined ...

A Novel Sub-optimum Maximum-Likelihood Modulation ...
signal. Adaptive modulation is a method to increase the data capacity, throughput, and efficiency of wireless communication systems. In adaptive modulation, the ...

Agreement Rate Initialized Maximum Likelihood Estimator
classification in a brain-computer interface application show that. ARIMLE ..... variance matrix, and then uses them in (12) to compute the final estimates. There is ...