Reducing Computational Complexities of Exemplar-Based Sparse Representations With Applications to Large Vocabulary Speech Recognition Tara N. Sainath, Bhuvana Ramabhadran, David Nahamoo, Dimitri Kanevsky IBM T. J. Watson Research Center, Yorktown Heights, NY 10598, U.S.A {tsainath, bhuvana, nahamoo, kanevsky}@us.ibm.com

Recently, exemplar-based sparse representation phone identification features (Spif ) have shown promising results on large vocabulary speech recognition tasks. However, one problem with exemplar-based techniques is that they are computationally expensive. In this paper, we present two methods to speed up the creation of Spif features. First, we explore a technique to quickly select a subset of informative exemplars among millions of training examples. Secondly, we make approximations to the sparse representation computation such that a matrix-matrix multiplication is reduced to a matrix-vector product. We present results on four large vocabulary tasks, including Broadcast News where acoustic models are trained with 50 and 400 hours, and a Voice Search task, where models are trained with 160 and 1000 hours. Results on all tasks indicate improvements in speedup by a factor of four relative to the original Spif features, as well as improvements in word error rate (WER) in combination with a baseline HMM system.

1. Introduction Recently, various exemplar-based sparse representation (SR) techniques have successfully been applied to speech recognition tasks [3], [4], [7], [8]. Exemplar-based SR methods typically make use of individual training examples, which can be quite large (many millions) for large vocabulary systems. This makes training of SR systems quite time consuming, and thus most work has been limited to systems with less than 50 hours of training data (i.e., [3], [7]). In this paper, we look at making exemplar-based SR systems tractable to train at thousands of hours of data, while still preserving the WER improvements observed in past SR work. Specifically, we focus on improving the computational efficiency of the sparse representation phone identification features (Spif ), as first presented in [7]. In a typical SR formulation, we are given a test vector y and a set of exemplars hi from a training set belonging to different classes, which we put into a dictionary H = [h1 ; h2 . . . ; hn ]. y is represented as a linear combination of training examples by solving y = Hβ, subject to a spareness constraint on β. A Spif feature is then created by computing class-based posterior probabilities from the β elements belonging to different classes. The process of creating a Spif features at each frame involves two steps. First, we must construct an appropriate dictionary H at each frame. Secondly, β must be computed at each frame using a SR technique. Using the method in [7], the construction of H involves randomly selecting a small subset of training examples at each frame. The process of random selection without replacement from the entire training set has computational complexity which is linearly proportional to the number of training examples [5], which is millions of frames for large vocabulary tasks. To address this issue, we explore grouping training data aligning to

different HMM states into separate classes. Since frames aligning to a particular HMM state behave similarly, we hypothesize that the dictionary used for these frames can be the same. Thus, our first contribution is to randomly sample from the training data once for each HMM state, such that a fixed dictionary H is created for state. This computation of H per state can be done off-line, significantly minimizing the time to construct H. Second, in [7] the creation of Spif features by solving y = Hβ is done via the Approximate Bayesian Compressive Sensing (ABCS) SR method [1]. For a dictionary H of size m × n, the process of finding β involves a matrix-matrix multiplication, and is of order O(m2 n). Our second contribution is introducing an approximation to the ABCS algorithm such that the matrix-matrix multiplication is reduced to a matrix-vector multiplication, reducing computational complexity to O(mn). Our experiments are conducted on two large vocabulary corpora, namely a Broadcast News (BN) task [7] and an Voice Search (VS) task. Results are presented when Spif features are created and acoustic models are trained on 50 and 400 hours of BN, as well as 160 and 1,280 hours of VS data. Results on all four experiments show improvements in speedup by approximately a factor of four with no loss in accuracy compared to the original Spif features. In addition, the faster Spif features allow for an improvement in WER over a baseline HMM system for all tasks, demonstrating the benefit of exemplar-based techniques even with very large amounts of data. The rest of this paper is organized as follows. Section 2 reviews the creation of Spif features for speech recognition. Section 3 discusses the creation of a fixed dictionary, Section 4 highlights computational speedups of ABCS and Section 5 summarizes the algorithm to create both original and fast Spif features. Experiments and results are presented in Sections 6 and 7 respectively. Finally, Section 8 concludes the paper.

2. Overview of Spif Features In this section, we review the creation of Spif features [7]. 2.1. Creation of Spif Features Given a test frame y and a set of exemplars hi ∈ ℜm from training which we put into a dictionary H = [h1 , h2 , . . . hN ], the objective of any SR technique is to represent y as a linear combination of training examples by solving y = Hβ subject to a sparseness constraint on β. β = [β1 , β2 , . . . , βN ] is a vector where each element βi ∈ β corresponds to the weight given to column hi ∈ H. In addition, H is typically an over-complete dictionary (i.e., m << N ) where training examples hi belong to multiple classes (for example phoneme classes). The β vector obtained by solving y = Hβ is often sparse and only non-zero for the elements in H which belong to the

same class as y. Using this β vector, a classification score for different classes in H can be calculated using a variety of different rules ([6], [8]). As described in [6] one approach is to define a selector δi (β) ∈ ℜN as a vector whose entries are zero except for entries in β corresponding to class i. We then compute the l2 norm for β for class i as ∥ δi (β) ∥2 . The best class for y will be the class in β with the largest l2 norm. Building on the SR classification work in [6], [7] uses these SR classification scores to construct Spif features. Specifically, i the Spif score for a specific class i is defined to be the classification score for that class, namely ∥ δi (β) ∥2 . After a score is computed for each class in H, the scores are normalized across Si i = ∑ pifi . A Spif feature is all classes, in other words S¯pif i

Spif

then created by concatenating the individual class scores into a C 1 0 ], where C , . . . , S¯pif , S¯pif vector, in other words Spif = [S¯pif is the total number of classes. 2.2. Using Spif Features for Recognition Now that Spif features have been defined, in this section we review from [7] how Spif features are used in recognition, including creating a dictionary H and solving for β. 2.2.1. Constructing Dictionary In speech recognition, each training example can be labeled by the HMM state that aligns to it. Pooling together all training data from all states into H will make the columns of H large, and will make solving for β intractable. Therefore, to reduce computational costs, the data is decoded with an HMM to find the best aligned state at each frame. For each state, the closest 24 states to this state are found as described in [8]. H is then seeded with the training data aligning to these 25 states. Since using all training data aligning to the top 25 states typically amounts to thousands of training samples in H, we must sample this further. Specifically, for each state N training examples are randomly sampled from all training frames that aligned to this state. In total, [7] uses 700 training examples to construct H, distributed across the 25 states. 2.2.2. Solving for β In [7], the ABCS SR method [1] was used to solve the problem y = Hβ, subject to a semi-Gaussian sparsity constraint on β. y is assumed to satisfy a linear model as: y = Hβ + ζ where ζ ∼ N (0, R). This allows us to represent p(y|β) ∼ N (Hβ, R) as a Gaussian distribution. Assuming β is a random variable with prior p(β), the maximum a posteriori (MAP) estimate for β can be obtained as follows: β ∗ = arg maxβ p(β|y) = maxβ p(y|β)p(β). In the ABCS formulation, p(β) is the product of two prior constraints, namely a Gaussian constraint pG (β) and a semi-Gaussian constrain pSG (β) which enforces sparsity. pG (β) is assumed to have a Gaussian distribution, i.e., pG (β) = N (β|β0 , P0 ). Here β0 and P0 are initialized statistical moments and will be described in more detail below. We have observed that if an informative Gaussian prior is used which gives more weight to examples in H belonging to the same class as y, then the semiGaussian has no additional benefit. Thus, the optimal β solves the following problem: β ∗ = arg maxβ p(y|β)pG (β). [1] shows that the ABCS solution to this problem has a closed-form expression given by Equation 1. ( ) β ∗ = I − P0 H T (HP0 H T + R)−1 H β0 +

(1) P0 H T (HP0 H T + R)−1 y At each frame, P0 and β0 must be initialized. β0 = 0 since it is assumed to be sparse [6]. The diagonal elements of P0 have dimension equal to the number of examples in H, and thus each column in P0 has a corresponding class label. The class label is defined as follows. Each column of H is associated with a particular HMM state (i.e. “aa-b-0”) which can be mapped to a monophone class label (“a”), where multiple states (i.e. “aa-b0” and “aa-m-2”) can map to the same monophone class. Diagonal elements of P0 are initialized such that the entries corresponding to a particular class are proportional to the HMM state posteriors which map to that class. Specifically, given the 25 unique HMM states {λ1 , λ2 , . . . , λ25 } in H, we can calculate a posterior p(λi |y) ≈ p(y|λi )p(λi ) for each state i. Here p(y|λi ) is computed by a GMM that is used to model state i and p(λi ) is a prior on state i. Then, the posterior for a “class” is calculated by summing the state posteriors which map to that class. Each diagonal element of P0 is set to be proportional to the posterior of the class i it belongs to. The intuition behind setting P0 in this fashion is that the larger the initial P0 , the more weight is given to examples in H belonging to this class. Therefore, the class posteriors picks out the most likely supports, and ABCS provides an addition step by using the actual training data to refine these supports. This approach to setting P0 , as opposed to setting it to a constant value, was found to provide more than a 2% absolute improvement in WER on a phonetic classification task [6].

3. Use of Fixed Dictionary H In this section, we propose an approach to reduce the time to create dictionary H per frame. 3.1. Motivation The approach described in Section 2.2.1 to create H has computational drawbacks. Particularly, for a given HMM state, P training points must be randomly selected without replacement from the N training points which aligned to that state. This has a computational complexity of O(N ) [5]. A 50-hour BN task contains roughly N = 6, 000 frames per state. Since we need to select training points from 25 different states, the overall computation time to seed H is O(25 × 6, 000). 3.2. Creation of Fixed Dictionary H In [8], two different methodologies for selecting training examples belonging to each HMM state were compared. The first method was to randomly sample N training points from a state, while the second method was to select samples from the state which well reflected the overall GMM distribution. It was found creating Spif features with the two methods did not provide a difference in WER. Instead, it mattered more what classes were use to seed H, rather than the actual training points in H. Using this intuition, we propose the following approach to seed H. Recall that for two frames y1 and y2 which align to the same state, the top 25 states and classes in H are identical. Therefore, we suspect that the Spif features for these two frames would not be different if H was chosen for both frames through random selection, or the same H was used for both. The creation of a fixed H for each HMM state is done as follows. Given an HMM state and knowledge of the 24 closest HMMs to this state, we randomly sample training data from each of the 25 states once, and use this to create a fixed H per HMM state.

0.7 0.6

Mean/Std Per Class

We test this hypothesis of using a fixed H by creating a f ixed var set of Spif features where H is chosen per frame, and Spif where H is chosen to be fixed per HMM state. Figure 1 shows a f ixed var histogram of the difference in Spif − Spif for over 500 Spif features. Notice that with over 90% frequency the difference between the two Spif values is less than 0.1 absolute, supporting our claim to use a fixed H. The fixed dictionary approach is computationally much less expensive. H is created per state ahead of time, and thus at each frame it is not necessary to randomly sample the training data.

0.5 0.4 0.3 0.2 0.1 0 0

0.8 0.7

2

4

6 8 Classes in H

10

12

14

Figure 2: Mean/Std of Class Posteriors

Frequency

0.6 0.5

multiplication given in Equation 1 can be computed off-line. We denote this multiplication per state s as M s in Equation 2.

0.4 0.3

¯ s )T (H ¯ s P¯0s (H ¯ s )T + R)−1 M s = P¯0s (H

0.2 0.1 0 −0.8

−0.6

−0.4

−0.2 0 S Difference

0.2

0.4

0.6

pif

f ixed var Figure 1: Spif − Spif Distribution

4. Speeding up SR Computation In this section, we propose a method to speed up ABCS. 4.1. Motivation To better understand how to speed up ABCS, we first calculate the computational complexity of the original ABCS algorithm. The calculation of β is given by Equation 1. Using the fact that β0 = 0, H ∈ ℜm×n , P0 ∈ ℜn×n and β ∈ ℜn×1 , the most computationally expensive step involves the product HP0 H T . More specifically, assuming P0 is a diagonal matrix, then the multiplication HP0 is O(mn), and produces a matrix HP0 which is of size m × n. Multiplying HP0 by H T , which is an n × m matrix, has a complexity of O(m2 n) [9]. 4.2. Speeding up ABCS Using the ABCS methodology given in Equation 1, even if H is computed off-line, P0 is computed for every frame y, and thus a matrix-matrix multiplication HP0 H T is necessary. For all frames y which align to the same state, the classes in H and are the same. We hypothesize that for these frames, the class posteriors used to initialize P0 will behave similarly across different frames, and the standard deviation of class posteriors will be small. To test this hypothesis, Figure 2 plots the mean and standard deviation of different class posteriors for all frames which aligned to a particular HMM state. The figure shows that a class probability differs from its mean at most with a standard deviation of 0.1, indicating class probabilities of frames belonging to the same HMM state behave similarly. Based on the above conclusion, we explore using a fixed P0 per HMM state. Specifically, for all frames in training which align to the same HMM state s, we calculate a P0t,s value by computing the class posteriors at each frame t. We then average all P0 values for all frames aligned to a specific state s, ∑ whicht,s as denoted by P¯0s = N1 t∈N P0 . Given a dictionary fixed ¯ s and a fixed P¯0s per state, this implies that the matrix-matrix H

(2)

Given a frame y and corresponding M s factor, β can be found in one step by solving β = M s × y. The computational complexity now involves multiplying M s , an n × m matrix, with an m-dimensional vector y. The overall computational complexity in solving for β is reduced to O(mn) [9].

5. Summary of Creating Spif Features Now that we have described the process of creating Spif features, in this section we summarize both techniques. Algorithm 1 shows the steps involved in creating the original Spif features, while Algorithm 2, referred to as fast Spif , outlines the steps using the proposed changes discussed in Sections 3 and 4. The main differences between the two approaches include: • Creation of H is done per frame in Step 5 of Algorithm 1, while it is created off-line in Step 3 of Algorithm 2. • Solving of β with ABCS. A matrix-matrix multiplication per frame is required in Step 7 of Algorithm 1. In Algorithm 2, we compute P¯0s off-line. Therefore, only a matrix-vector multiplication per frame is required. Algorithm 1 Original Method of Creating Spif Features [7] 1: Given an existing set of HMM models, perform a forced path alignment on the training set to obtain the state-level alignments at each frame. This is done off-line. 2: For each HMM state sj , save out all training frames which align to this state. This is done off-line. 3: For each HMM state sj , calculate the 24 closest HMM states to this HMM. This is done off-line. 4: With an existing set of HMM models, decode the test data to get a state-level alignment S = {s1 , s2 , . . . , sT } for all frames Y = {y1 , y2 , . . . , yT }. 5: For each frame yt ∈ Y and corresponding st ∈ S, get the top 25 HMM states from Step 3. Randomly sample training examples aligning to these 25 HMMs in order to seed H with 700 total examples. 6: Calculate P0 by scoring the current frame yt against each of the 25 HMM states. 7: Use ABCS to solve yt = Ht βt . The solution for βt is given by Equation 1. 8: Use βt create Spif t features for each class i.

Algorithm 2 Fast Method for Creation of Spif Features 1: Given an existing set of HMM models, perform a forced path alignment on the training set to obtain the state-level alignments at each frame. This is done off-line. 2: For each HMM state sj , save out all training frames which align to this state. This is done off-line. 3: For each HMM state sj , we randomly sample training data from ¯ sj per each of the 25 classes once, and use this to create a fixed H HMM state. This is done off-line. s 4: Calculate an average P¯0 j per state by averaging all P0 values from all frames in training aligning to the same state sj . This is done off-line. 5: With an existing set of HMM models, decode the test data to get a state-level alignment S = {s1 , s2 , . . . , sT } for all frames Y = {y1 , y2 , . . . , yT }. 6: For each frame yt ∈ Y and corresponding st ∈ S, use the appro¯ st and fixed P¯ st . priate fixed dictionary H 0 st s ¯ ¯ t 7: Given yt , H and P0 , solve for βt using Equation 2. 8: Use βt create Spif t features for each class i.

6. Experiments 6.1. Broadcast News The first set of experiments are conducted on an English broadcast news transcription task [7]. Two different acoustic models are explored, trained on 50 hours and 400 hours of data from the 1996 and 1997 English Broadcast News Speech Corpora. Results are reported on 3 hours of the EARS Dev-04f set. The initial acoustic features are 19-dimensional PLP features. Speaker adaptation (SA) and discriminative training (DT) are applied to both features and models create the best baseline HMM system. 6.2. Voice Search The second set of experiments are conducted on a large vocabulary Voice Search task. Unlike the English Broadcast News speech corpora, there is no publicly available standardized training and test sets exist in the community for this real-world, large vocabulary task. However, there have been few publications in the literature [2] that have reported on similar tasks with baselines WERs ranging widely from 16.0% to 28%. Two baseline research system were trained on 160 hours and 1280 hours of in-house voice search data yielding WERs of 24.8% and 23.0% respectively. Similar to the broadcast news task, these systems were trained using PLP features and include SA and DT.

7. Results 7.1. Computation Time Table 1 shows the average time per frame to seed H and compute β for the original and fast Spif methods. The overall time is also provided. The fast Spif is approximately four times faster than the original Spif method with a majority of speedup coming from using a fixed dictionary H per state. Notice that there is a time to seed H in the fast Spif method as we must load an H per state into memory. The time to create Spif features scales linearly with the amount of training data. SR Step Seeding H Solving for β, ABCS Overall Per Frame

Original Spif [7] 0.22 sec 0.05 sec 0.27 sec

Fast Spif 0.066 sec 0.00076 sec 0.071 sec

Table 1: Comparison of Computation Times

In addition, the results in Table 2 indicate that there is no degradation in performance when using the fast Spif features. Corpus BN-50 VS-160

Original Spif [7] 19.5 25.5

Fast Spif 19.4 25.6

Table 2: Comparison of WER 7.2. Accuracy Finally, we explore the WER using Spif features across different tasks. As in [7], gains with Spif features are observed when used in model combination with an HMM baseline system. We see that across all four tasks, Spif features allow for improvements in WER. Naturally as the number of hours increases, the gains are less substantial than with smaller amounts of data. However, the table still indicates the fast Spif method allows for the creation of Spif features on large amounts of data. These features are still able to capture the benefit of exemplarbased methods and provide gains in combination with an HMM. System Baseline Fast Spif Model Comb.

BN-50 18.7 19.4 18.1

BN-400 17.3 17.2 17.1

VS-160 24.8 25.6 23.8

VS-1280 23.0 23.1 22.7

Table 3: WER Results, BN and VS

8. Conclusions In this paper, we explored speeding up the creation of Spif features to make them more tractable to create for large vocabulary tasks. Specifically, we investigated the use of a fixed dictionary H per state created off-line. Second, we computed P0 in the ABCS algorithm off-line, allowing for β to be computed using a matrix-vector product. We presented results on four large vocabulary tasks. Results on all tasks indicated improvements in speedup by a factor of four relative to the original Spif features, allows for creating of Spif features on large amounts of data. In addition, we observed that the fast Spif features are still able to capture the benefit of exemplar-based methods and provide gains in combination with a baseline HMM system.

9. References [1] A. Carmi, P. Gurfil, D. Kanevsky, and B. Ramabhadran, “ABCS: Approximate Bayesian Compressed Sensing,” Tech. Rep., 2009. [2] C. Chelba and et al., “Query Language Modeling for Voice Search,” in Proc. IEEE Workshop on Spoken Language Technology, 2010. [3] J. F. Gemmeke, B. Cranen, and U. Remes, “Sparse Imputation for Large Vocabulary Noise Robust ASR,,” Computer Speech and Language, vol. 25, no. 2, pp. 462–479, 2011. [4] J. F. Gemmeke and T. Virtanen, “Noise robust exemplar-based connected digit recognition,” in Proc. ICASSP, 2010. [5] S. Goodman and S. Hedetniemi, Introduction To The Design And Analysis Of Algorithm. McGraw Hill, 1977. [6] T. N. Sainath, A. Carmi, D. Kanevsky, and B. Ramabhadran, “Bayesian Compressive Sensing for Phonetic Classification,” in Proc. ICASSP, 2010. [7] T. N. Sainath, D. Nahamoo, B. Ramabhadran, D. Kanevsky, V. Goel, and P. M. Shah, “Exemplar-Based Sparse Representation Phone Identification Features,” in Proc. ICASSP, 2011. [8] T. N. Sainath, B. Ramabhadran, D. Nahamoo, D. Kanenvsky, and A. Sethy, “Sparse Representation Features for Speech Recognition,” in Proc. Interspeech, 2010. [9] G. Strang, Introduction to Linear Algebra. Wellesley-Cambridge Press, 2003.

Reducing Computational Complexities of Exemplar ...

Recently, exemplar-based sparse representation phone identification features ... quite large (many millions) for large vocabulary systems. This makes training of ...

89KB Sizes 0 Downloads 258 Views

Recommend Documents

Reducing the Computational Complexity of ... - Research at Google
form output of each spatial filter is passed to a longer-duration ... The signal is passed through a bank of P spatial filters which convolve .... from 0 to 20 dB. Reverberation is simulated using the image model [15] – room dimensions and micropho

Approaches for Reducing the Computational Cost of ...
otherwise, set y = y′ and go to Step 2. and go to Step 2. III. ENHANCEMENTS TO THE KM ALGORITHMS. Five enhancements to the KM algorithms are introduced in this section. Their computational costs are also compared with the original KM algorithms. Al

A Method for Reducing the Computational Complexity ...
E. N. Arcoverde Neto, A. L. O. Cavalcanti Jr., W. T. A. Lopes, M. S. Alencar and F. Madeiro ... words, the encoder uses the encoding rule C(x) = bI if d(x, wI ) < d(x, ...

New Literacies, New Complexities: Professional Development with ...
Development with Chinese Teachers. Hiller A. ... Complexity Science and Educational Research 2010 Conference ... of literacy, evolving technologies, and.

Determinantal complexities and field extensions
2 Institute of Computing Technology, Chinese Academy of Sciences [email protected] ... Laboratory for Information Science and Technology, Department of Computer. Science and .... an algebraic extension of degree d to the base field increases the

Reducing the impact of interference during programming
Nov 4, 2011 - PCT/US2008/074621, ?led Aug. 28, 2008. (Continued). Primary Examiner * Connie Yoha. (74) Attorney, Agent, or Firm *Vierra Magen Marcus ...

INVESTIGATIONS ON EXEMPLAR-BASED FEATURES FOR SPEECH ...
from mobile and video sharing speech applications. How- ever, most ... average distance between the hypothesis X and the k-nearest .... the hypothesis in the lattice with the lowest edit cost. ... scribed in Section 2 for development and tuning.

Exemplar-Based Sparse Representation Phone ...
1IBM T. J. Watson Research Center, Yorktown Heights, NY 10598. 2MIT Laboratory for ... These phones are the basic units of speech to be recognized. Moti- vated by this ..... to seed H on TIMIT, we will call the feature Sknn pif . We have also.

Exemplar-Based Sparse Representation Features ...
in LVCSR systems and applying them on TIMIT to establish a new baseline. We then .... making it difficult to compare probabilities across frames. Thus, to date SVMs ...... His conversational biometrics based security patent was recognized by.

Unsuppressible Repetition Suppression and exemplar ...
different process from predicting a particular face with specific features (e.g. John vs. ... Our design involved 3 types of blocks: “Prediction Blocks”, “Repetition Blocks” .... and defined our ROIs as 4-mm–radius spheres centered on local

Don't let the reporting complexities of the Affordable Care ... - ADP.com
*ACA Study, ADP Research Institute, 2014. Mid-sized employers defined as 50-999 employees. Are they able to gather it all? In order to properly complete ACA ...

Concealments of Problems: An Incentive of Reducing ...
Jun 25, 2017 - ... Database, http://www.sozogaku.com/fkd/en/cfen/CA1000622.html. .... justified by assuming that players take care of social norms. ...... φ1,r (s0) < 0 for all r, subordinate 0's best response is to conceal: r0 = 0. ..... )st + pSD(

Human health consequences of reducing emissions of ... - Woodsmoke
Oct 1, 2014 - Association (UNEP/WMO) to reduce short-lived greenhouse pollutants (methane (CH4), black carbon (BC) and ozone precursors) are .... that our returns from the standard types of investment will buy us much less in reducing ...... Proceedi

Reducing Systemic Cybersecurity Risk - OECD
Jan 14, 2011 - views of the OECD or of the governments of its member countries. ...... seeking to punish downloaders of copyright material, against the .... to focus more on the process of analysing risk rather than simply having a long list ... abou

Computational Vision
Why not just minimizing the training error? • Never select a classifier using the test set! - e.g., don't report the accuracy of the classifier that does best on your test ...

Human health consequences of reducing emissions of ... - Woodsmoke
Oct 1, 2014 - climate benefits, the health co-benefits are more local and can be achieved more quickly and directly, making them more tangible and .... Strong Pacific trade winds are reported to be increasing subsurface ocean ... The IPCC's AR4 lists

Reducing Systemic Cybersecurity Risk - OECD.org
Jan 14, 2011 - patches to operating systems and applications; the deployment of anti-malware, firewall and intrusion detection products and services; the use ...

Active Dual Supervision: Reducing the Cost of ...
we deal only with text classification in this paper, all .... and 1000 negative reviews from the Internet Movie .... next best feature to label is one that will result in.

Reducing Effect of Outliers in Landmark-based Spatial ...
This paper proposes a framework to improve accuracy of the existing landmark-based localization .... finally selects the top-ranked hypothesis as the result. The landmark ... ment and the size of environment is 10 × 18, then s is 2 and ν is 180.