A Multiple Operator-valued Kernel Learning Approach to Functional Regression with Functional Responses

Hachem Kadri SequeL Team, INRIA Lille - Nord Europe, Villeneuve d’Ascq, France

[email protected]

Alain Rakotomamonjy [email protected] LITIS, UFR de Sciences, Universit´e de Rouen, St Etienne du Rouvray, France Francis Bach Sierra Team, INRIA, Ecole Normale Sup´erieure, Paris, France Philippe Preux LIFL/CNRS/INRIA, Universit´e de Lille, Villeneuve d’Ascq, France

Abstract This paper addresses the problem of learning a finite linear combination of operatorvalued kernels. We study this problem in the case of kernel ridge regression for functional responses with a `r -norm constraint on the combination coefficients (r ≥ 1). We propose a multiple operator-valued kernel learning algorithm based on solving a system of linear operator equations by using a block coordinate descent procedure. We experimentally validate our approach on a functional regression task in the context of finger movement prediction in Brain-Computer Interface (BCI).

1. Introduction During the past decades, a large number of algorithms have been proposed to deal with learning problems in the case of single-valued functions. Recently, there has been considerable interest in estimating vector-valued functions (Micchelli & Pontil, 2005b). Much of this interest has arisen from the need to learn tasks where the target is a complex entity, not a scalar variable as usual. Typical learning situations include multitask learning (Evgeniou et al., 2005) and functional regression (Kadri et al., 2010). “Object, functional and structured data” Workshop. ICML Workshop series, Edinburgh, Scotland, UK, 2012. Copyright 2012 by the author(s)/owner(s).

[email protected] [email protected]

In this paper, we are interested in the problem of functional regression with functional responses in the context of Brain-Computer Interface (BCI) design. More precisely, we are interested in finger movement prediction from electrocorticographic signals (Miller & Schalk, 2008). Indeed, from a set of signals measuring brain surface electrical activity on d channels during a given period of time, we want to predict, for any instant of that period whether a finger is moving or not and the amplitude of the finger flexion. Formally, the problem consists in learning a functional dependency between a set of d signals and a vector of labels and between the same set of signals and vector of real values (the amplitude). This problem can benefit from multiple operator-valued kernel learning framework. Indeed, one of the difficulties arises from the unknown latency between the signal set-on related to the finger movement and the actual movement. Hence, instead of fixing in advances some value of this latency in the regression model, our framework allows to learn it from the data by means of several operator-valued kernels.

2. Problem Setting We first consider the problem of estimating a function f such that f (xi ) = yi when observed data (xi , yi )i=1,...,n are assumed to be elements of infinite dimensional separable Hilbert spaces. In the following we denote by Gx and Gy the domains of xi and yi respectively. X = {x1 , . . . , xn } denotes the training set with corresponding targets Y = {y1 , . . . , yn }. Since Gx and Gy are spaces of functions, the problem can be

Multiple Operator-valued Kernel Learning

thought of as an operator estimation problem, where the desired operator maps a Hilbert space of factors to a Hilbert space of targets. We can define the regularized operator estimate of f ∈ F fλ , arg min f ∈F

n X

kyi − f (xi )k2Gy + λkf k2F

(1)

i=1

3. Solving the MovKL Problem After having presented the framework, we now devise an algorithm for solving this multiple operator-valued kernel learning. 3.1. Block-coordinate descent algorithm

In this work, we are looking for a solution to this minimization problem in a reproducing kernel Hilbert space F of function-valued functions on some functional input space Gx . Let F be a linear space of operators on Gx with values in Gy . We assume that F is a Hilbert space with inner product h·, ·iF . Let L(G y ) be the set of bounded linear operators from Gy to Gy .

Since the optimization problem has the same structure as a multiple scalar kernel learning problem, we can build our algorithm upon the MKL literature. Hence, we propose to borrow from Kloft et al. (2011), and consider a block-coordinate descent method. Our algorithm iteratively solves the problem with respects to α with d being fixed and then with respects to d with α being fixed. This boils down to the following steps :

Definition 1 (function-valued RKHS) A Hilbert space F of functions from Gx to Gy is called a RKHS if there is a nonnegative L(G y )-valued kernel KF (w, z) on Gx × Gx such that:

1. with {dk } fixed , the resulting optimization problem with respects to α has a simple form which solution is given by: (K + λI)α = 2y

i. KF (w, ·)g belongs to F, ∀z ∈ Gx , w ∈ Gx , g ∈ Gy , ii. ∀f ∈ F, hf, KF (w, ·)giF = hf (w), giGy

where K =

Definition 2 (operator-valued kernel) An L(G y )-valued kernel KF (w, z) on Gx is a function KF (·, ·) : Gx × Gx −→ L(G y ); furthermore: i. KF is Hermitian if KF (w, z) = KF (z, w)∗ , where KF (z, w)∗ is the adjoint operator of KF (z, w) ii. KF is nonnegative on Gx if it is Hermitian and ∀r ∈ N∗ and ∀{(wi , ui )ri=1 } ∈ Gx × Gy , KF (wi , wj )uj , ui iGy ≥ 0. Let us now consider that the function f (·) is sum of M functions {fk (·)}M k=1 where each fk belongs to an Gy -valued RKHS of kernel Kk (·, ·). Similarly to scalarvalued multiple kernel learning, we can cast the problem of learning these functions fk as min

d∈D

min

kfk k2F k k=1 dk

PM

fk ∈Fk

with ξi = yi −

+

PM

k=1

1 λ

P i

kξi k2Gy

(2)

fk (xi )

with P rd = [d1 , · · · , dM ], D = {d : ∀k, dk ≥ 0 and k dk ≤ 1} and 1 ≤ r ≤ ∞. A partial dualization of this problem leads to the following equivalent one min max −λkαk2Gyn − hKα, αiGyn + 4hα, yiGyn

d∈D α∈Gyn

where K =

M P

(3)

dk Kk and Kk is the block operator

k=1

kernel matrix associated to the operator-valued kernel Kk . The KKT conditions also state that at optimality dk X Kk (xi , ·)αi . we have fk (·) = 2 i

M P

(4)

dk Kk . While the form of solution

k=1

is rather simple, solving this linear system is very challenging and we propose an algorithm for its resolution in the sequel. 2. with {fk } fixed, according to problem (2), we can rewrite the problem as min

d∈D

M X kfk k2F

k

k=1

dk

(5)

which has a closed-form solution and for which optimality occurs at (Micchelli & Pontil, 2005a): 2

kfk k r+1 dk = P 2r ( k kfk k r+1 )1/r

(6)

In this algorithm, we have to solve a linear system involving a block-operator kernel matrix which is a combination of basic kernel matrices associated to M operator-valued kernels. We present an algorithm for solving it in the next paragraph. 3.2. Solving a linear system with multiple block operator-valued kernels One common way to construct operator valued kernels is to build scalar-valued ones which are carried over to the vector-valued (resp. function-valued) setting by a positive definite matrix (resp. operator). In this setting an operator-valued kernel has the form K(w, z) = G(w, z)T , where G is a scalar-valued kernel

Multiple Operator-valued Kernel Learning

where

Algorithm 1 `r norm MovKL Input Kk for k = 1, . . . , M 1 d1k ←− for k = 1, . . . , M M α ←− 0 for t = 1, 2, . . . do α0 ←− α P K ←− k dtk Kk

si = 2yi −

(t−1)

K(xi , xj )αj

,

j=i+1

Now, by means of a variable splitting approach, we are able to decouple the role of the different kernels. Indeed, the above problem is equivalent to the following one : min (t) αi

1 ˆ (t) (t) (t) hK(xi , xi )αi , αi iGyM − hsi , αi iGyM 2 (t)

(t)

with αi,1 = αi,k

for k = 1, . . . , M

and T is an operator in L(G y ). Now, we show how to solve linear operator equations (4) in the general case where multiple operator-valued kernels are combined as follows M X



n X

Kk = dk Gk Tk , ∀k ∈ {1, . . . , M }, and KM +1 = λI.

2 r+1

K(w, z) =

(t) K(xi , xj )αj

j=1

α ←− solution of (K + λI)α = 2y if kα − α0 k <  then break end if kfk k dt+1 ←− P 2r k ( k kfk k r+1 )1/r end for

i−1 X

dk Gk (w, z)Tk

(7)

k=1

Inverting the associated block operator kernel matrix K is not feasible in this case, that is why we propose a Gauss-Seidel iterative procedure to solve the system of linear operator equations (4). Starting from an initial vector of functions α(0) , the idea is to iteratively compute, until a convergence condition is satisfied, the functions αi according to the following expression

for k = 2, . . . , M + 1

ˆ i , xi ) is the M + 1 × M + 1 diagonal matrix where K(x (t) (t) (t) +1 [Kk (xi , xi )]M is the vector (αi,1 , . . . , αi,M +1 ) k=1 . αi and the M + 1-dimensional vector si = (si , 0, . . . , 0). We now have to deal with a quadratic optimization problem with equality constraints. Writing down the Lagrangian of this optimization problem and then deriving its first-order optimality conditions leads us to the following set of linear equations  P  K1 (xi , xi )αi,1 − si + M = 0 k=1 γk (9) Kk (xi , xi )αi,k − γk = 0  αi,1 − αi,k = 0 where k = 2, · · · , M + 1 and {γk } are the Lagrangian multipliers related to the M equality constraints. Finally, in these set of equations, the operator-valued kernels have been decoupled and thus, if their inversion can be easily computed (which is the case in our experiments), one can solve the problem (9) with respects to {αi,k } and γk by means of another Gauss-Seidel algorithm after simple reorganisation of the linear systems.

4. Experiments (t)

[K(xi , xi ) + λI]αi

= 2yi −

i−1 P j=1



n P j=i+1

(t)

K(xi , xj )αj (t−1)

(8)

K(xi , xj )αj

where t is the iteration index. This problem is still challenging because the kernel K(·, ·) still involves a positive combination of operator-valued kernels. Our algorithm is based on the idea that instead of inverting the finite combination of operator-valued kernels [K(xi , xi ) + λI], we can consider the variational formulation of this system

In order to highlight the benefit of our multiple operator-valued kernel learning approach, we focus on the problem of estimating the finger movement amplitude from the ECoG signals. To this aim, the fourth dataset from the BCI Competition IV (Miller & Schalk, 2008) was used. For our finger movement prediction task, we have kept 5 channels that have been manually selected and split the AM feature signals in portions of 200 samples. For each of these time segments, we have the real movement amplitudes. An example of input-output signals are depicted in Figure 1. In a nutshell, the problem boils down to be a supervised functional regression task.

M +1

min (t) αi

1 X (t) (t) (t) h Kk (xi , xi )αi , αi iGy − hsi , αi iGy 2 k=1

To evaluate the performance of the multiple operatorvalued kernel learning approach, we use the residual

Multiple Operator-valued Kernel Learning Ch. 1

20

6

0 −20

5 0

20

40

60

80

100

120

140

160

180

Table 1. Residual Sum of Squares Error (RSSE) results for finger movement prediction.

200

Ch. 2

10

4

−10

0

20

40

60

80

100

120

140

160

180

200

0

20

40

60

80

100

120

140

160

180

200

0

20

40

60

80

100

120

140

160

180

200

Ch. 3

5 0 −5

Ch. 4

5

Finger Movement

0

3 2 1

Algorithm

RSSE

KRR - scalar-valued KRR - functional response MovKL - `∞ norm MovKL - `1 norm MovKL - `2 norm -

88.21 79.86 76.52 78.24 75.15

0 −5

0

Ch. 5

5

−1

0 −5

0

20

40

60

80

100

120

140

160

180

200

Time samples

0

50

100 Time samples

150

200

Figure 1. Example of a couple of input-output signals in our BCI task. (left) Amplitude modulation features extracted from ECoG signals over 5 pre-defined channels. (right) Real amplitude movement of the finger.

sum of squares error (RSSE) as evaluation criterion for curve prediction Z X RSSE = {yi (t) − ybi (t)}2 dt (10) i

where ybi (t) is the prediction of the function yi (t) corresponding to real finger movement or the finger movement state. For the multiple operator-valued kernels having the form (7), we have used a Gaussian kernel with 5 different bandwidths and a polynomial kernel of degree 1 to 3 combined with three operators T : identity T y(t) = y(t), multiplication operator associated with the func2 2 tion e−t defined by T y(t) = e−t y(t), and the integral Hilbert-Schmidt operator with the kernel e−|t−s| proR −|t−s| posed in (Kadri et al., 2011) T y(t) = e y(s)ds. Empirical results on the BCI dataset are summarized in Table 1. The dataset was randomly partitioned into 65% training and 35% test sets. We compare our approach in the case of `1 and `2 -norm constraint on the combination coefficients with: (1) the baseline scalar-valued kernel ridge regression algorithm by considering each output independently of the others, (2) functional response ridge regression using an operator-valued kernel constructed from the integral operator (Kadri et al., 2011), (3) kernel ridge regression with evenly-weighted sum of operator-valued kernels, which we denote by `∞ -norm MovKL. As in the scalar case, using multiple operator-valued kernels leads to better results. By directly combining kernels constructed from identity, multiplication and integral operators we could reduce the residual sum of squares error. Best results are obtained using the MovKL algorithm with `2 -norm constraint on the combination coefficients. RSSE of the baseline kernel ridge regression are significantly outperformed by the operator-valued kernel based functional response regression. These results confirm that by taking into account the relationship between outputs we can improve

performance. This is due to the fact that an operatorvalued kernel induces a similarity measure between two pairs of input/output.

5. Conclusion In this paper we have presented a new method for learning simultaneously an operator and a finite linear combination of operator-valued kernels. We have extended the MKL framework to deal with functional response kernel ridge regression and we have proposed a block coordinate descent algorithm to solve the resulting optimization problem. The method is applied on a BCI dataset to predict finger movement in a functional regression setting. Experimental results show that our algorithm achieves good performance outperforming existing methods.

References Evgeniou, T., Micchelli, C. A., and Pontil, M. Learning multiple tasks with kernel methods. JMLR, 2005. Kadri, H., Duflos, E., Preux, P., Canu, S., and Davy, M. Nonlinear functional regression: a functional rkhs approach. In AISTATS, 2010. Kadri, H., Rabaoui, A., Preux, P., Duflos, E., and Rakotomamonjy, A. Functional regularized least squares classification with operator-valued kernels. In ICML, 2011. Kloft, M., Brefeld, U., Sonnenburg, S., and Zien, A. `p norm multiple kernel learning. Journal of Machine Learning Research, 12:953–997, 2011. Micchelli, C. and Pontil, M. Learning the kernel function via regularization. Journal of Machine Learning Research, 6:1099–1125, 2005a. Micchelli, C. A. and Pontil, M. On learning vector-valued functions. Neural Computation, 17:177–204, 2005b. Miller, K. J. and Schalk, G. Prediction of finger flexion: 4th brain-computer interface data competition. BCI Competition IV, 2008.

A Multiple Operator-valued Kernel Learning Approach ...

not and the amplitude of the finger flexion. Formally, the problem consists in learning a functional depen- dency between a set of d signals and a vector of la- bels and between the same set of signals and vector of real values (the amplitude). This problem can benefit from multiple operator-valued kernel learning frame-.

327KB Sizes 0 Downloads 199 Views

Recommend Documents

Unsupervised multiple kernel learning for ... -
The comparison of these two approaches demonstrates the benefit of our integration ... ity of biological systems, heterogeneous types (continuous data, counts ... profiles to interaction networks by adding network-regularized con- straints with .....

Multiple Kernel Learning Captures a Systems-Level Functional ... - PLOS
Dec 31, 2013 - This is an open-access article distributed under the terms of the ... out in the UK. ..... Multi-center, international initiatives for data sharing and.

A Conscience On-line Learning Approach for Kernel ...
2School of Mathematics and Computational Science. Sun Yat-sen University, P. R. China. The 10th IEEE International Conference on Data Mining. Wang et al.

A Conscience On-line Learning Approach for Kernel ...
Iterative procedure such as k-means can be used to seek one of the local minima. .... prototypes is characterized by the convergence criterion. = ∑. =1. ∥ − ˆ ∥2 ..... obtain a meaningful comparison, on each dataset, the same value is used f

Multiple Kernel Clustering
leviate this problem. Examples include semi-definite programming (SDP) [26, 27, 28], alternating optimiza-. ∗Department of Automation, Tsinghua University, ...

Hyperparameter Learning for Kernel Embedding ...
We verify our learning algorithm on standard UCI datasets, ... We then employ Rademacher complexity as a data dependent model complexity .... probability of the class label Y being c when the example X is x. ..... Mining, pages 298–307.

A parallel multiple reference point approach for multi ...
Available online 11 January 2010. Keywords: ... problems: even the most powerful computer of any generation can be easily .... tion levels and the resulting objective vector is called a reference point and can ...... Computer Science, vol. 2632.

Amalgams: A Formal Approach for Combining Multiple Case Solutions
solutions. In the paper we define amalgam as a formal operation over terms in a generalization space, ...... In In 11th International Conference on Industrial and ...

Amalgams: A Formal Approach for Combining Multiple ...
Enric Plaza (2010), Amalgams: A Formal Approach for Combining Multiple ... are multiple open problems such as what knowledge is required for adaptation.

A multiple-population evolutionary approach to gate ...
Jan 15, 2004 - On the hardware side, computer networks, multi- processor computers ..... dure that executes the job swap move 10.g times in each individual ...

A multiple-objective approach for the vulnerability ...
of infrastructure networks. M. Claudio & S. ... The infrastructure enterprise is key to sustaining the economy ... neering and risk analysis have provided tools and.

An experiment on learning in a multiple games ...
Available online at www.sciencedirect.com ... Friedl Schoeller Research Center for Business and Society, and the Spanish Ministry of Education and Science (grant .... this does not prove yet that learning spillovers do occur since behavior may be ...

Learning Relationships between Multiple Modalities and Words
*This work was partially supported by JST, CREST. 1Akira Taniguchi and Tadahiro Taniguchi are with Ritsumeikan Univer- sity, 1-1-1 Noji Higashi, Kusatsu, Shiga 525-8577, Japan {a.taniguchi, taniguchi} @em.ci.ritsumei.ac.jp. 2Angelo Cangelosi is with

Multiple Categorization by iCub: Learning ...
For example, if infant grasps a red ball at hand, the parent may describe an action of infant and an object using a sentence. “look at red apple” apple ? red ?

Learning Relationships between Multiple Modalities and Words
that can learn multiple categorizations and words related to any of four modalities (action, object, position, and color). This paper focuses on a cross-situational learning using the co-occurrence of sentences and situations. We conducted a learning