APPLICATION SPECIFIC LOSS MINIMIZATION USING GRADIENT BOOSTING Bin Zhang1∗ , Abhinav Sethy2 , Tara N. Sainath2 , Bhuvana Ramabhadran2 1

University of Washington, Department of Electrical Engineering, Seattle, WA 98125 2 IBM T. J. Waston Research Center, Yorktown Heights, NY 10598 [email protected], {asethy,tsainath,bhuvana}@us.ibm.com ABSTRACT Gradient boosting is a flexible machine learning technique that produces accurate predictions by combining many weak learners. In this work, we investigate its use in two applications, where we show the advantage of loss functions that are designed specifically for optimizing application objectives. We also extend the original gradient boosting algorithm with Newton-Raphson method to speed up learning. In the experiments, we demonstrate that the use of gradient boosting and application specific loss functions results in a relative improvement of 0.8% over a 82.6% baseline on the CoNLL 2003 named entity recognition task. We also show that this novel framework is useful in identifying regions of high word error rate (WER) and can provide up to 20% relative improvement depending on the chosen operating point. Index Terms— Gradient boosting, named entity recognition, confidence estimation 1. INTRODUCTION Boosting algorithms have received a lot of attention in the field of machine learning, and they have gained popularity in various natural language processing (NLP) applications. The strength of boosting is that it is able to generate a very accurate model based on the combination of many simple and rough models, which are usually called base or weak learners. A well-known boosting algorithm, AdaBoost [1], has demonstrated its power and effectiveness in practice. It has a nice modular framework to re-weight training samples based on the performance of base learners in each boosting iteration, which can be shown to be equivalent to greedily minimizing an exponential classification loss [2]. However, it is designed for classification, and the loss function cannot be arbitrarily changed. This limits its application in other tasks such as regression and occasions where the loss function needs to be specifically defined. Gradient boosting [3] is a boosting algorithm that is able to minimize a variety of loss functions. Similar to AdaBoost, it generates an additive model that is a weighted sum of many ∗ This

work was performed when the first author was an intern at IBM.

base learners. Unlike AdaBoost, however, it isolates the base learners from the loss function, by only fitting base learners to the negative functional gradient of the loss function in least square sense. This enables the application of gradient boosting to a multitude of differentiable loss functions, with little requirement to the base learners. Recently, gradient boosting has been applied in robust training of hidden Markov models for automatic speech recognition (ASR) [4]. In this paper, we study the gradient boosting algorithm in the context of classification, regression and ranking. First, we apply gradient boosting as a classifier in CoNLL 2003 English named entity recognition. An F-measure loss function is introduced to directly optimize for the task evaluation metric. Second, we use gradient boosting to identify utterances which are transcribed poorly by an ASR system. Compared to the traditional regression view of this problem, the proposed ranking loss fits the evaluation metric more precisely and yields better performance. We also show that replacing the gradient descent component with Newton-Raphson updates speeds up learning and improves the performance furthermore.

2. BACKGROUND The gradient boosting algorithm can be found in [3]. The very original algorithm is designed for regression, where we want to fit a function F (·) of input feature vector x to the target value y. It can also be applied to a J-class classification problem by fitting J regression functions Fj (x), and the posterior probability of each class can be computed using logistic regression. Regularizations in gradient boosting is accomplished by controlling the degree of freedom of the base learner, and by using a small enough step size. The most popular base learner is regression trees. Stumps, which are one-level binary regression trees, have the smallest degree of freedom, and they are widely used in many other boosting algorithms as well. The degree of freedom increases as the tree level increases, and it can be usually tuned together with the step size on development data. Moreover, a benefit of using trees more than one-level is that feature conjunctions can be learned.

A modification to gradient boosting is stochastic gradient boosting [5]. It sub-samples the training data in each iteration at random. Empirical study shows that it can improve overall performance. The sub-sampling fraction is another hyper-parameter that can be tuned.

Therefore, an F1 loss function may be defined correspondingly (various formulations exist, they will, however, lead to similar analyses afterwards),

3. APPLICATION SPECIFIC LOSSES

where n1 is the number of samples from class 1. The nij ’s are non-differentiable functions of F (x), therefore differentiable approximations are needed. For example, one can compute expected counts by using the posterior probabilities from logistic regression [6]. However, it does not work well with gradient boosting, because of small misclassifications penalty. To utilize the heavy misclassification penalty from exponential loss, and, in the mean time, promote high F1 , we propose weighted exponential loss. From Eq. (1), we observe that n10 plays a more important role in changing LF 1 , as it is in both numerator and denominator. We can thereby use a weighted misclassification loss where n10 is weighted more,

With great flexibility in the choice of loss functions, users can tailor gradient boosting to different applications. There are several standard loss functions that we use in our experiments for classification and regression [3] (Table 1). There are also other regression loss functions such as absolute deviation and Huber loss that penalize outliers less heavily. Due to specific evaluation metrics, in practice, we may still need to design our own loss functions. We will describe our approach to loss function design in the following two examples. Loss function Exponential loss (AdaBoost) Negative log likelihood Square loss

Definition P exp(−yj Fj (x)) Pj − j:yj =1 log(Fj (x)) (y − F (x))2

Table 1. Standard loss functions. They are represented as individual sample loss. The first two are classification loss functions, where yj is 1 if the sample has label j, and 0 otherwise. Square loss is a regression loss function. 3.1. F-measure Loss F-measure is commonly used to evaluate an object detection system, especially when the distribution of classes is skewed. The most frequently used F-measure is the F1 score, 2 · Precision · Recall . Precision + Recall It is used in the evaluation of our named entity recognition task, where the names are the objects to be detected. Usually the loss functions for minimizing error rate can yield good results. In this section, however, we show that we can achieve even better results by using the following F-measure loss. In a two-class scenario, there is a class to be detected and a reject class. The F-measure can be computed based on the contingency table (Table 2), where the nij ’s are the counts of samples with corresponding reference and hypothesis.

LF 1 =

n10 + n01 n10 + n01 1 −2= = , F1 n11 n1 − n10

(1)

Lw = λn10 + n01 , λ > 1. +n01 If we choose λ = nn11 −n , then for any changes in n10 and 10 n01 , denoted by ∆n10 and ∆n01 , LF 1 decreases with Lw , because ∆Lw = n1 − n10 − ∆n10 ≥ 0. ∆LF 1

This indicates that we can use Lw as a proxy for LF 1 . Furthermore, since exponential loss performs well in reducing misclassification, we can intuitively replace the misclassification counts with individual exponential losses, yielding the weighted exponential loss X X Lwexp = λ exp(−F (xi )) + exp(F (xi )), i:yi =1

i:yi =0

F1 =

Hypothesis Reference

1 0

1 n11 n01

0 n10 n00

Table 2. Two class contingency table Assuming that class 1 is the object to be detected and class 0 is the reject class, the F-measure can be computed as F1 =

2n11 . 2n11 + n01 + n10

where i is the sample index, and λ is updated in each gradient boosting iteration according to the error counts n10 and n01 . Note that λ approaches infinity when n1 = n10 , which may happen in the first iteration. This can be prevented by safeguarding λ with a positive upper limit. We use 10 in our experiments. In a multi-class (J-class) scenario, the object to be detected come from multiple classes. For example, there are four types of names to be detected in the named entity recognition task. Not only is missing a name an error, but also classifying a name into a wrong type. Denote n11e as the number of detected objects but with wrong classes, we have 2n11e + n10 + n01 , n1 − n11e − n10 Lw = λ1 n11e + λ2 n10 + n01 ,

LF 1 =

+n11e +n01 where the choice of λ2 = nn11 −n and λ1 = λ2 + 1 11e −n10 similarly guarantees that LF 1 and Lw change in the same direction. Hence the weighted exponential loss for multi-class

can be defined as Lwexp =

X

J X

exp(Fj (xi ))

i:yi =0 j=1

 +

X

 X

 i:yi 6=0

λ1 exp(−Fj (xi )) + λ2 exp(−F0 (xi )) ,

j:j6=0

where we assume classes labeled other than 0 are the objects to be detected.

Algorithm 1: Gradient boosting with Newton-Raphson Input: training data D = {(xi , yi )}, i = 1, . . . , N , and base learner type L Output: Learner F (x) Initialize F0 (x); for m = 1 to M do (x)) gm (xi ) = ∂L(D,F , i = 1 to N ; ∂F (x) F (x)=Fm−1 (xi ) 2 (D,F (x)) hm (xi ) = ∂L∂F , i = 1 to N ; (x)2 fm (x) =

3.2. Ranking Loss Our motivation for studying the ranking loss comes from the problem of identify utterances with high error rate in the ASR output. We are interested in identifying a set of utterances whose mean error rate would pass some quality threshold specified as WER. It is possible to view this task as a regression problem where the learning algorithm is trained to predict the WER of an utterance using features such as ASR confidence, number of consensus bins, etc. We can use the regression losses described in Table 1 in the gradient boosting framework to train such a WER prediction model. However, we believe that the task is better formulated as a ranking problem, due to the use of cumulative WER of sorted hypotheses in its evaluation. Therefore, we propose to use the following ranking loss function X Lrank = exp(−(yi − yj )(F (xi ) − F (xj )). i
Here yi is the WER of utterance i, and xi is the utterancelevel features. Similar to [7], it is a sum of pairwise exponential loss terms that promote correct ordering of pair (F (xi ), F (xj )). Moreover, it weights the pairwise losses by the difference between the target WER pair (yi , yj ). This is motivated by the fact that it is usually more crucial to rank the pairs with larger target difference in the right order, because any wrong ordering of these pairs leads to more adversarial cumulative WER degradations. Our results in Section 5.2 validate our motivation for the ranking loss. 4. SPEEDING UP GRADIENT BOOSTING Gradient boosting has a slower learning speed than AdaBoost, as fitting learners to negative gradient direction is suboptimal to direct loss minimization. Nevertheless, one may use optimization methods other than gradient descent, e.g., conjugate gradient [8] or Newton-Raphson [2], to achieve faster learning speed. In particular, we found in our experiments that the Newton-Raphson method (Algorithm 1) introduces significant speedup (Section 5.1). Note that, for the base learners, the least square fitting is changed to weighted least square fitting with the weights from second-order gradients.

F (x)=Fm−1 (xi ) ˛ “ ”˛ PN gm (xi ) ˛2 ˛ i=1 hm (xi )˛f (xi )− − hm (xi ) ˛ PN ; argminf ∈L i=1 hm (xi )

αm = argminρ L(D, Fm−1 (x) + ρfm (x)); Fm (x) = Fm−1 (x) + αm fm (x); end PM F (x) = FM (x) = m=1 αm fm (x); 5. EXPERIMENTS 5.1. Named Entity Recognition We use data sets from the CoNLL 2003 shared task [9]. The target is to detect four types of names from English news wire text, including person names, location names, organization names, and miscellaneous names. There are 219k words in training data, 55k words in development data, and 50k words in evaluation data. The features are extracted using the Stanford NER system [10]. There are 850k binary word-level features. Gradient boosting results with different optimization methods are shown in Table 3. Using 10000 boosting iterations with stumps, Newton-Raphson update (NR) has significant improvements over gradient descent (GD) or conjugate gradient (CG) for both loss functions. Loss Exponential loss Neg. log likelihood

Optimization GD CG NR GD CG NR

Dev F1 (%) 85.52 85.70 88.84 85.87 85.76 88.35

Eval F1 (%) 78.86 78.25 82.31 78.25 78.18 82.27

Table 3. Comparison of gradient boosting with different optimization methods In Table 4, we compare gradient boosting with other machine learning methods1 . Gradient boosting with NewtonRaphson update outperforms MaxEnt (by Mallet [11]) and AdaBoost (by icsiboost [12]), especially with 2-level regression trees as base learners. Meanwhile, F-measure loss improves the F1 score furthermore. 1 Results are obtained using 10000 iterations. Other parameters are tuned on dev. Note that there are other methods yielding better performance on this task [9], but with expanded feature sets and additional training data.

Dev F1 (%) 88.48 88.23 88.84 88.89 89.41 89.27

Eval F1 (%) 82.57 81.87 82.31 82.66 83.15 83.28

35

25 20 15 10

Table 4. Comparison of gradient boosting and other methods

5

5.2. Identifying High Error Rate Decodes

00

Predicting WER of ASR hypotheses is useful for controlling the error rate. For instance, we may wish to identify the portion of hypotheses with high error rate which we can further process with a more accurate but slower and more complex ASR system. As discussed in Section 3.2, the WER prediction can be cast as a regression problem and a ranking problem. We believe that since the ordering of the confidence estimator output is more critical than its absolute value, gradient boosting with ranking loss should perform better than regression loss. The data set in this experiment is the broadcast news ASR output from the IBM speech recognition system. We use the DEV04 set (888 utterances) for training and the RT04 set (1761 utterances) for test. For each utterance, we have an 11-dimensional continuous feature vector, with features including statistics of the consensus, cross-system (speakerindependent vs. speaker-dependent) error rates, and lattice density. Figure 1 compares the cumulative WER when the hypotheses are sorted by the confidence estimates in ascending order. The results are obtained using 1000 iterations of gradient descent with 3-level trees and 0.1 step size. Hypothesislevel WER values are used in training as targets for regression (square loss) or ranking (ranking loss). A better confidence estimator produces a cumulative WER curve closer to the oracle curve by WER. Compared to the baseline estimator, average probability of the top consensus hypothesis (avgtoptokprob), ranking loss performs significant better in most of the regions, and it also consistently outperforms square loss, suggesting that the ranking formulation is a better fit to the application. At the hypothesis percentages of our interest (30% and 50%), square loss performs worse than the baseline, while ranking loss improves cumulative WER: 4.7% vs. 6.0% (@30%) and 6.2% vs. 7.6% (@50%). 6. CONCLUSIONS We have demonstrated the flexibility of gradient boosting framework with novelly designed loss functions to match our applications. The proposed F-measure loss optimizes the named entity recognition F-measure. Together with the speedups from Newton-Raphson updates, we have achieved 0.7% absolute improvement over the MaxEnt baseline. We have also introduced the ranking loss for identifying high

WER (oracle) avgtoptokprob Square loss Ranking loss

30 Cumulative WER (%)

Approach MaxEnt (Mallet) AdaBoost (icsiboost) Stump NR exp loss Stump NR F loss 2-level tree NR exp loss 2-level tree NR F loss

20

40 60 Percentage of hypotheses (%)

80

100

Fig. 1. Comparison of cumulative WER for different loss functions error rate ASR outputs. It not only outperforms the conventional square regression loss, but also offers up to 20% relative improvements over the baseline confidence estimator. 7. ACKNOWLEDGMENTS We greatly thank Raimo Bakis for all the helpful discussions. 8. REFERENCES [1] Robert Schapire and Yoram Singer, “Improved boosting algorithms using confidence-rated predictions,” Machine Learning, vol. 37, no. 3, pp. 297–336, December 1999. [2] Jerome Friedman, Trevor Hastie, and Robert Tibshirani, “Additive logistic regression: A statistical view of boosting,” The Annals of Statistics, vol. 28, pp. 337–407, 2000. [3] Jerome Friedman, “Greedy function approximation: A gradient boosting machine,” The Annals of Statistics, vol. 29, no. 5, pp. 1189–1232, 2001. [4] Hao Tang, Mark Hasegawa-Johnson, and Thomas S. Huang, “Toward robust learning of the gaussian mixture state emission densities for hidden Markov models,” in Proceedings of ICASSP 2010, 2010, pp. 5242– 5245. [5] Jerome Friedman, “Stochastic gradient boosting,” Computational Statistics & Data Analysis, vol. 38, no. 4, pp. 367–378, Feburary 2002. [6] Martin Jansche, “Maximum expected F-measure training of logistic regression models,” in Proceedings of HLT 2005, 2005, pp. 692–699. [7] Yoav Freund, Raj Iyer, Robert E. Schapire, and Yoram Singer, “An efficient boosting algorithm for combining preferences,” in Journal of Machine Learning Research, 2003, vol. 4, pp. 170–178. [8] Kristian Kersting and Bernd Gutmann, “Unbiased conjugate direction boosting for conditional random fields,” in Proceedings of MLG 2006, 2006, pp. 157–164. [9] Erik F. Tjong Kim Sang and Fien De Meulder, “Introduction to the CoNLL-2003 shared task: Language-independent named entity recognition,” in Proceedings of CoNLL-2003, 2003, pp. 142–147. [10] Jenny Finkel, Trond Grenager, and Christopher Manning, “Incorporating non-local information into information extraction systems by gibbs sampling,” in Proceedings of ACL 2005, 2005, pp. 363–370. [11] Andrew McCallum, “Mallet: A machine learning for language toolkit,” http://mallet.cs.umass.edu, 2002. [12] Benoit Favre, Dilek Hakkani-T¨ur, and Sebastien Cuendet, “Icsiboost,” http://code.google.come/p/icsiboost.

Application Specific Loss Minimization Using Gradient ...

controlling the degree of freedom of the base learner, and by ... degree of freedom increases as the tree level increases, and ..... Statistics & Data Analysis, vol.

174KB Sizes 3 Downloads 262 Views

Recommend Documents

Method for intercepting specific system calls in a specific application ...
Sep 30, 2004 - (12) Statutory Invention Registration (10) Reg. No.: Tester. States .... a security application executing on a host computer system in accordance ...

Method for intercepting specific system calls in a specific application ...
Jul 3, 2007 - NETWORK 126. APPLICATION. 106. OPERATING. SYSTEM. 104. MEMORY114 ..... recipient, such as a system administrator. From optional .... physically located in a location different from processor 108. Processor 108 ...

Application-Specific Integrated Circuits - Addison Wesley Michael ...
Application-Specific Integrated Circuits - Addison Wesley Michael John Sebastian Smith.pdf. Application-Specific Integrated Circuits - Addison Wesley Michael ...

Application-Specific Memory Management in ... - Semantic Scholar
The Stata Center, 32 Vassar Street, Cambridge, Massachusetts 02139. Computer Science and ... ware mechanism, which we call column caching. Column caching ..... in part by the Advanced Research Projects Agency of the. Department of ...

Application Specific Parallel Architectures
an evaluation system that can help the designer in the ... A system for ASA development has to be ...... two neighbors and the host, the 2N memory loads and.

Using Quadtrees for Energy Minimization Via Graph Cuts
For illustration, in figure 5 we reconstruct .... for representing 3D arrays such as those proposed in [18]. 6.3 Application .... In Tutorial at ACM. SIGGRAPH 2005 ...

two-level boolean logic minimization using microbial ...
In conventional GA, a generational system is used. From one generation a .... the system prematurely converged to a solution with a fitness of 19,000 which is.

Using Quadtrees for Energy Minimization Via Graph Cuts
How to efficiently... (ordered in increase order of difficult). – compute them? – compute their areas? – compute the perimeters of common borders? – compute their neighborhood system? Page 8. 14:12. 8. Quadtrees. Our proposal: Page 9. 14:12.

Using Low Level Gradient Channels for ...
Gradient has been traditionally used in the computer vision field [3]. A lot of well ... filtered by 10 Gabor filters to form the image channel. The Haar features are ...

Bayesian Sampling Using Stochastic Gradient ... - Research at Google
dom walk effect, and the Metropolis step serves as a correction of the .... first look at a 1-D illustration of the SGNHT sampling in the presence of unknown noise. ..... is run for a total number of 50k iterations with burn-in of the first 10k itera

GRADIENT IN SUBALPINE VVETLANDS
º Present address: College of Forest Resources, University of Washington, Seattle .... perimental arenas with alternative prey and sufficient habitat complexity ...... energy gain and predator avoidance under time constraints. American Naturalist ..

Coupling-Driven Bus Design for Low-Power Application-Specific ...
wire-to-wire spacing is shrinking for higher densities and the as- ... ing wire widths. For example of metal 3 layer in typical 0.35 µm. CMOS process, the lateral component of capacitance reaches 5 times the sum of fringing and vertical components w

Site-specific synthesis and application of Deuterium-Labeled ... - Arkivoc
Dec 27, 2016 - 1974, 22, 2656-2661. 617 http://dx.doi.org/10.1248/cpb.22.2656. 618. 29. Seo, S.; Yoshimura, Y.; Satoh, T.; Uomori, A.; Takeda, K. i., J. Chem.

Site-specific synthesis and application of Deuterium-Labeled ... - Arkivoc
Dec 27, 2016 - Chemistry, Fried, J.; Edwards, J. A., Eds. Van Nostrand Reinhold Company: New York, 1972; Vol. 1, pp. 572. 145–221. 573. 8. Shinkyo, R.; Xu ...

System and method for obtaining and using location specific information
(73) Assignee: Apple Inc., Cupertino, CA (US). (21) App1.No.: 12/874,155. (22) Filed: Sep. 1, 2010. Related US. Patent Documents. Reissue of: (64) Patent No.:.

Specific Binding of Cancer Cells Using a ...
Jan 16, 2010 - X. Zheng1, S.L. Cheung1, L. Wang2, J. Schroeder3, R. Heimark4,. J.C. Baygents2, R.Guzman2 & Y. Zohar1. November 19, 2009. 1Aerospace & Mechanical Engineering, 2Chemical & Environmental Engineering. 3Molecular and Cellular Biology, 4D

Fair Simulation Minimization - Springer Link
Any savings obtained on the automaton are therefore amplified by the size of the ... tions [10] that account for the acceptance conditions of the automata. ...... open issue of extending our approach to generalized Büchi automata, that is, to.

Customized Cognitive State Recognition Using Minimal User-Specific ...
minimum number of user-specific training examples. METHODS .... Pz, O1, and O2 according to the international 10-. 20 system for EEG ... successive R waves) in the epoched window. We .... of classification accuracy, TL+ACS may require a.

Influence of using date-specific values when extracting ... - CiteSeerX
The TIMESAT [1] software program is the most advanced, ..... [1] P. Jonsson and L. Eklundh, “TIMESAT - a program for analyzing time- series of satellite sensor ...

Mining Advertiser-specific User Behavior Using Adfactors
mining the long-term patterns in the advertising data. Categories and Subject Descriptors. H.4.m [Information Systems]: Miscellaneous; J.4 [Social.

System and method for obtaining and using location specific information
Sep 1, 2010 - supports the coordinate entry or linked to an existing Web ..... positions to any GPS receiver that is within the communica tion path and is tuned ...

Online Determination of Track Loss Using Template Inverse Matching
Sep 29, 2008 - match between the real target and one of the virtual targets. Under this .... The experiments are implemented on a standard PC. (Pentium IV at ...

Application Communities: Using Monoculture for ...
curity risks of software monoculture. ... both analytical and experimental results that show ACs are ... Software monoculture has been identified as a major.