GRAPH REGULARIZED LOW-RANK MATRIX RECOVERY FOR ROBUST PERSON RE-IDENTIFICATION Ming-Chia Tsai, Chia-Po Wei, Yu-Chiang Frank Wang Research Center for Information Technology Innovation, Academia Sinica, Taipei, Taiwan ABSTRACT Robust person re-identification (PRID) refers to the problem of matching individuals across non-overlapping camera views, while the images captured by either camera might be occluded or even missing. To address this challenging task, we propose a low-rank matrix recovery (LR) based approach in this paper. In addition to observing the global structure of cross-camera images via LR, we further exploit their local geometrical information via graph regularization, which preserves the recovered images with recognition guarantees. Our experiments verify the effectiveness and robustness of our approach, which is shown to perform favorably against state-of-the-art PRID methods. Index Terms— Person Re-Identification, Low-Rank Matrix Recovery, Graph Regularization 1. INTRODUCTION Determining the identities of images across non-overlapping camera views is known as person re-identification (PRID). It attracts the attention of researchers in the fields of computer vision and image processing, and benefits applications like video surveillance and computation forensics. Due to significant visual appearance variations like lighting and view point changes, including differences in camera parameters, recognizing images across different views is a difficult problem. Existing PRID approaches can be divided into two categories. Matching-based methods aim at searching for representative features, so that the distances between cross-view images can be properly measured. For example, Gary and Tao [1] exploited localized visual features for relating crosscamera images. Bak et al. [2] extracted Haar and DCD based feature descriptors with spatial covariance information for improving the matching robustness. Recently, while saliencybased features were considered in [3], Li et al. [4] advanced deep neural networks to learn the features for PRID. In contrast to matching-based approaches, researchers also advance learning techniques to cope with the visual appearance differences between cameras. For example, metric learning has been applied to derive a proper distance metric, aiming at projecting cross-camera images into a feature space for PRID purposes. Recently developed approaches include Large Margin Nearest Neighbor (LMNN) [6], Information

978-1-4799-8339-1/15/$31.00 ©2015 IEEE

4654

Fig. 1. Illustration of robust PRID. Each column depicts crosscamera images of the same subject with possible missing and occluded data at either view. Note that the images are from VIPeR [5].

Theoretic Metric Learning (ITML) [7], and Logistic Discriminant Metric Learning (LDML) [8]. On the other hand, B¨auml et al. [9] chose to minimize the matching and reconstruction loss by utilizing multinomial logistic regression, and Xiong et al. [10] applied kernel-based metric learning with discriminant features for improved recognition. In addition, Prosser et al. [11] and Avraham et al. [12] regarded PRID as ranking and domain adaptation problems, respectively. In practice, one cannot always expect that images captured by different cameras are always available for matching or learning purposes. As illustrated in Figure 1, in addition to corrupted image regions due to occlusion or severe lighting and view point changes, one might have missing data at either camera view during the training process. Thus, robust PRID refers to the use of such corrupted and incomplete data. In this paper, we propose a low-rank matrix recovery (LR) based approach for robust PRID. We note that, both Liu et al. [13] and Fu et al. [14] applied LR as preprocessing/refinement techniques for PRID. However, the former did not consider any missing data, while the latter only derived an approximated solution for LR. Our proposed algorithm not only advances LR for retrieving the global structural information from cross-camera images (including missing or corrupted ones), we further exploit local geometrical information between observed images by enforcing graph regularization [15]. As confirmed by our experiments, our method would perform favorably against state-of-the-art PRID approaches and achieve improved recognition performance.

ICIP 2015

2. OUR PROPOSED METHOD 2.1. Graph Regularized Low-Rank Matrix Recovery For robust PRID, we advocate the extraction of global structures from cross-camera images, while their local geometrical information is exploited for handling such data. In our work, we advance the technique of low-rank matrix recovery for solving this task. Let X = [x1 , x2 , . . . , xn ] ∈ Rm×n and Y = [y1 , y2 , . . . , yn ] ∈ Rm×n be the image sets, in which xi and yi are the m-dimensional feature vectors of the ith instance of cameras 1 and 2, respectively. We utilize the feature extraction method suggested in [13] to generate a m = 1, 984 dimensional feature for describing (as discussed in the experiments). Now, we define the cross-view image matrix Z = [X; Y] ∈ R2m×n , i.e., each column of Z contains a cross-camera image pair observed by the cameras. We note that missing data could occur at either camera view, i.e., either xi or yi might not be available for instance i. Since the cross-camera images are expected to be linearly related to each other in terms of their color-based features [13], Fu et al. [14] considered the following optimization problem for recovering the low-rank structure from the cross-view image matrix Z: min rank(A) + λ kEk0 s.t. Z = A + AΩ + E,

A,AΩ ,E

(1)

where A ∈ R2m×n is the recovered low-rank version of Z, and AΩ ∈ R2m×n and E ∈ R2m×n are the matrices for representing missing data and outliers, respectively. The upper part of Figure 2 illustrates the structures of Z, A, AΩ , and E, respectively. The underlying observation of [14] is that, if there exist missing instances at either view in Z, the corresponding and complementary vectors will be recovered in A and AΩ , respectively. Although promising results have been reported in [14] for the task of PRID, we observe two issues which would limit its performance. First, the formulation in (1) does not impose constraints on AΩ for enforcing the sparsity in column (i.e., the recovered complementary parts of the missing data). As a result, the derived solution of (1) might not possess the desirable structure for AΩ as shown in Figure 2. Second, Fu et al. [14] did not consider the locality information presented in Z. More precisely, let zk and ak be the kth columns of Z and A, respectively. If zi is close to zj , there is no guarantee in [14] that ai and aj would be similar. This is crucial especially when dealing with corrupted image data. Nevertheless, the low-rank property proposed in [14] is not sufficient for robust PRID. To address the above two issues, we propose to solve the following optimization problem: min rank(A) + λ kEk0 + RG (A) + RM (AΩ )

A,AΩ ,E

(2)

s.t. Z = A + AΩ + E.

4655

Fig. 2. Illustration of our proposed method. Note that Z is the input cross-view image data, A is the derived low-rank matrix, while the associated matrices AΩ and EΩ aim at handling missing instances and sparsely corrupted image regions, respectively.

In (2), the regularizer RG (A) preserves the local geometric information by enforcing the similarity between ai and aj based on the observed xi and xj . It is defined as follows: RG (A) =

n 1 X kai − aj k22 Wij , 2 i,j=1

(3)

n×n where Wij is the (i, j)-entry of matrix and is   W ∈ R 2 −kxi −xj k2 defined as Wij = exp with σ as a width paσ rameter. Note that W is viewed as a weight matrix of a graph, which constructed by the instances observed from camera 1. This is the reason why we have RG (A) as a graph regularizer [16, 17]. We can further rewrite RG (A) as: n n n X X 1 X kai − aj k22 Wij = aTi ai Dii − aTi aj Wij 2 i,j=1 (4) i=1 i,j=1

= tr(ADAT ) − tr(AWAT ) = tr(ALAT ),

where tr(·) stands for the trace of the input Pn matrix, D ∈ Rn×n is a diagonal matrix with Dii = j=1 Wij , and L := D − W is called the Laplacian matrix [15]. On the other hand, RM (AΩ ) in (2) is to compensate the recovered missing data in A, and thus has the following form: RM (AΩ ) = tr(ATΩ M) + α kAΩ k2F ,

(5)

where M ∈ R2m×n is a binary matrix (i.e., its entries are either 0 or 1) with the same dimensions as AΩ . It can be seen that, if Mij corresponds to an entry of missing instances in AΩ , Mij is set as 1; otherwise Mij is set as 0. As a result, the term tr(ATΩ M) enforces AΩ to exhibit the desirable structure as shown in Figure 2 (i.e., observed sparsity in columns). However, to prevent the magnitude of entries of AΩ to be very large which favors the minimization (2), we

2

need to add an additional regularization term kAΩ kF in (5) to limit the magnitude of the entries of AΩ . With (3), (4), and (5), the proposed optimization problem can be expressed as min rank(A) + λ kEk0

A,AΩ ,E

+ tr(ALAT ) + tr(ATΩ M) + α kAΩ k2F

(6)

s.t. Z = A + AΩ + E.

To make (6) tractable, we solve its convex relaxation version: min

A,AΩ ,E,B

kAk∗ + λ kEk1

+ tr(BLBT ) + tr(ATΩ M) + α kAΩ k2F

(7)

s.t. Z = A + AΩ + E, A = B,

Algorithm 1 Graph Regularized Low-Rank Matrix Recovery Input: Cross-view data Z, step size ρ, parameter µ, missing data indicator matrix M, and Laplacian matrix L Initialization: A0 = Z, B0 = Z, AΩ,0 = 0, E0 = 0 while Not Converge do Ak+1 = arg minA F (A, AΩ,k , Ek , Bk ) Ek+1 = arg minE F (Ak+1 , AΩ,k , E, Bk ) 1 (Φ1,k + µk (Z − Ak+1 − Ek+1 ) − M) AΩ,k+1 = (µk +2α) Bk+1 = (µk Ak+1 − Φ2,k )(2L + µk I)−1 Φ1,k+1 = Φ1,k + µk (Z − Ak+1 − AΩ,k+1 − Ek+1 ) Φ2,k+1 = Φ2,k + µk (Bk+1 − Ak+1 ) µk+1 = ρµk end while Output: Recovered cross-view data A

by solving:

where kAk∗ denotes the nuclear norm of A, and kEk1 sums up the absolute values of each entry in E. We note that, an auxiliary variable B is introduced in (7) for facilitating the optimization process (as discussed in the next subsection). 2.2. Optimization To solve (7), we form the Lagrange function of (7) as follows: min

A,AΩ ,E,B

F (A, AΩ , E, B) = T

min

A,AΩ ,E,B

kAk∗ + λ kEk1

tr(ATΩ M)

arg min λ kEk1 + < Φ1 , Z − A − AΩ − E > E

µ kZ − A − AΩ − Ek2F 2 λ 1 = arg min kEk1 + kE − CE k2F , E µ 2 +

(10)

where CE = Z − A − AΩ + Φ1 /µ. From [18], the closedform solution of (10) is obtained as E = T [CE ], where T [·] is the soft thresholding operator with  = λ/µ. Updating AΩ Similarly, to update AΩ , we solve:

α kAΩ k2F

+tr(BLB ) + + (8) µ + < Φ1 , Z − A − AΩ − E > + kZ − A − AΩ − Ek2F 2 µ + < Φ2 , B − A > + kB − Ak2F , 2

in which Φ1 and Φ2 ∈ R2m×n are the Lagrange multipliers, and the equality constraints are regularized by parameter µ. We apply the inexact ALM method [18] for iteratively deriving the optimal solution A, AΩ , E, and B (see Algorithm 1 for the pseudo code). We now detail how to update the above variables in each iteration during optimization. Updating A With fixed AΩ , E, and B in (8), we update A by solving the following optimization problem: µ arg min kAk∗ + < Φ2 , B − A > + kB − Ak2F A 2 µ + < Φ1 , Z − A − AΩ − E > + kZ − A − AΩ − Ek2F 2 1 1 2 = arg min kAk∗ + kA − CA kF , A µ 2

(9) where CA = 0.5(Z−E+B+Φ1 /µ+Φ2 /µ). From [18], the closed-form solution of (9) is obtained as A = UT [S]VT , 1 where T [·] is the soft thresholding operator with  = 2µ , and USVT is the singular value decomposition of CA . Updating E With the same techniques above, we update E

4656

µ kZ − A − AΩ − Ek2F 2 + < Φ 1 , Z − A − AΩ − E > .

min tr(ATΩ M) + α kAΩ k2F + E

(11)

Setting the derivative of (11) with respect to AΩ to zero, we 1 obtain the optimal AΩ as (µ+2α) (µ(Z−A−E+Φ1 /µ)−M). Updating B Finally, we fix A, AΩ , and E in (8) to update B. It can be solved by: µ 2 min tr(BLBT )+ < Φ2 , B − A > + kB − AkF . (12) B 2 With the derivative of the above function with respect to B to zero, we obtain the optimal B as (µA − Φ2 )(2L + µI)−1 . After obtaining A, E, AΩ , and B, the Lagrange multipliers Φ1 and Φ2 are updated by the corresponding formulas in Algorithm 1. The convergence of these variables indicates the termination of our algorithm (typically 20 iterations in our experiments). 3. EXPERIMENTS We consider the VIPeR dataset [5] for evaluating the performance of our method. The VIPeR dataset contains images of 632 subjects, and each subject has a pair of images captured by two cameras 1 and 2 with non-overlapping views. We randomly and equally split this dataset into training and test sets. We report the average performance of 10 random trials.

Table 1. Recognition rates of different methods on VIPeR. Methods NN ELF [1] PRSVM [11] SDALF [19] PRDC [20] MC [13] LR [14] Ours

Rank=1 0.3 12 13 20 16 12.7 16.4 16.8

10 2.1 43 50 53 53.8 56.1 57.5 58.2

20 4.9 60 67 67 70 72 73.4 76.9

50 15.6 81 85 84 87 88 88.4 93.7

For the test set data, we consider images of camera 1 as the probe set, while those of camera 2 as the gallery set to be matched (as did in [13, 14]). Based on the setting of [14], we divide each image into 10 overlapping horizontal stripes. For each stripe, the color histograms of RGB, HSV, and Cb , Cr are calculated (with 8 bins for each each channel). As a result, each image will be represented by a (16 + 15) × 8 × 8 = 1984 dimensional feature vector. When constructing the matrix Z, we include all cross-camera image pairs from the training set and 316 probe images observed from camera 1 only. Once the LR process is complete, we will recover the instances at camera 2, and matching between such instances and the gallery ones can be performed accordingly. 3.1. Re-Identification with Corrupted Data We first consider the PRID problem with corrupted data, i.e., both training and test images might be corrupted due to occlusion or extreme view point changes. As a result, all 316 crossview image pairs from the training set are available when constructing Z. Once the optimization of our proposed algorithm is complete, we apply the metric of the Bhattacharya distance [19] for performing matching. To quantitatively evaluate the recognition performance, we cosnider Cumulative Matching Characteristic (CMC) in our experiments, which reflect the accuracy of the matching of top n candidates. To compare our approach with recent PRID ones, we have state-of-the-art approaches of ensemble of localized features (ELF) [1], person re-identification SVM (PRSVM) [11], symmetry-driven accumulation of local features (SDALF) [19], probabilistic relative distance comparison (PRDC) [20], matrix completion (MC) [13], and low-rank recovery (LR) [14]. We also have nearest neighbor classifiers (NN) as a baseline approach, which performs directly cross-camera image matching for PRID. Table 1 lists and compares the approaches. From this table, it can be seen that our approach was able to achieved improved recognition accuracy. We note that, methods like ELF and SDALF extracted more sophisticated visual features, while we only considered color-based ones for representation. 3.2. Re-Identification with Corrupted and Missing Data To address the more challenging and practical PRID problem, we consider not only corrupted images but also the missing

4657

Fig. 3. CMC comparisons of different approaches on VIPeR. Note that, in addition to corrupted image regions, 20% of training instances at either camera view are randomly disregarded (and thus viewed as missing data).

ones when constructing Z. In this part of the experiment, besides having the training data corrupted by background or cluttered regions, we further randomly disregard 20% of the instances at either camera view from the training set. With this setting, 20% of the subjects in the training set (i.e., 64 people) have only one image available. As for the test set, all instances at camera 1 are still available in the probe set (while those at camera 2 as the gallery ones). Figure 3 compares the CMC performances of NN, LR [14] and our proposed method. From this figure, we see that our improvements over LR and NN were more significant than those reported in Section 3.1 (i.e., performance with corrupted data only). More importantly, we can see that our approach outperformed all other approaches over different rank numbers. Based on the above experiments, the superiority of our proposed algorithm for robust PRID can be successfully verified.

4. CONCLUSION In this paper, we presented a novel graph regularized low-rank matrix recovery algorithm for robust person re-identification. The proposed algorithm aims at recovering the global structure of images observed across cameras. More importantly, it is able to exploit their local geometrical information by advancing graph regularization. This is the reason why our proposed algorithm allows us to handle corrupted and missing data in PRID problems. Our experiments not only confirmed the use of our method for such challenging PRID problems, we also verified that our method would perform favorably against state-of-the-art PRID methods. Acknowledgement This work is supported in part by the Ministry of Science and Technology of Taiwan via MOST103-2221-E-001021-MY2.

5. REFERENCES [1] D. Gray and H. Tao, “View Point Invariant Pedestrian Recognition with an Ensemble of Localized Features,” in ECCV, 2008. [2] S. Bak, E. Corvee, F. Br´emond, and M. Thonnat, “Person Re-identification Using Spatial Covariance Regions of Human Body Parts,” in IEEE AVSS, 2010. [3] Y. Yang, J. Yang, J. Yan, S. Liao, D. Yi, , and S. Z. Li, “Salient Color Names for Person Re-Identification,” in ECCV, 2014. [4] W. Li, R. Zhao, T. Xiao, and X. Wang, “DeepReID: Deep filter pairing neural network for person reidentification,” in IEEE CVPR, 2014.

[15] M. Belkin and P. Niyogi, “Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering,” in NIPS, 2001. [16] M. Zheng, J. Bu, C. Chen, C. Wang, L. Zhang, G. Qiu, and D. Cai, “Graph regularized sparse coding for image representation,” IEEE TIP, 2011. [17] X. Lu, Y. Wang, and Y. Yuan, “Graph-regularized lowrank representation for destriping of hyperspectral images,” IEEE Trans. Geoscience and Remote Sensing, 2013. [18] Z. Lin, M. Chen, L. Wu, and Y. Ma, “The augmented Lagrange multiplier method for exact recovery of corrupted low-rank matrices,” UIUC Tech. Rep. UILUENG-09-2215, 2009.

[5] D. Gray, S. Brennan, and H. Tao, “Evaluating Appearance Models for Recognition, Reacquisition, and Tracking,” in IEEE PETS Workshop, 2007.

[19] M. Farenzena, L. Bazzani, A. Perina, V. Murino, and M. Cristani, “Person Re-Identification by SymmetryDriven Accumulation of Local Features,” in IEEE CVPR, 2010.

[6] K. Q. Weinberger and L. K. Sau, “Fast solvers and efficient implementations for distance metric learning,” in In Proc. Intl Conf. on Machine Learning, 2008.

[20] W.-S. Zheng, S. Gong, and T. Xiang, “Person Reidentification by Probabilistic Relative Distance Comparison,” in IEEE CVPR, 2011.

[7] J. V. Davis, B. Kulis, P. Jain, S. Sra, and I. S. Dhillon, “ Information-Theoretic Metric Learning,” in Intl Conf. on Machine Learning, 2007. [8] M. Guillaumin, J. Verbeek, and C. Schmid, “Is that you? metric learning approaches for face identification,” in ICCV, 2009. [9] M. B¨auml, M. Tapaswi, and R. Stiefelhagen, “Semisupervised Learning with Constraints for Person Identification in Multimedia Data,” in IEEE CVPR, 2013. [10] F. Xiong, M. Gou, O. Camps, and M. Sznaier, “Person Re-Identification Using Kernel-Based Metric Learning Methods,” in ECCV, 2014. [11] B. Prosser, W.-S. Zheng, S. Gong, T. Xiang, and Q. Mary, “Person Re-Identification by Support Vector Ranking,” in BMVC, 2010. [12] T. Avraham, I. Gurvich, M. Lindenbaum, and S. Markovitch, “Learning Implicit Transfer for Person Re-identification,” in ECCV, 2012. [13] K. Liu, X. Guo, Z. Zhao, and A. Cai, “Person ReIdentification Using Matrix Completion,” in IEEE ICIP, 2013. [14] M.-H. Fu, Y.-C. F. Wang, and C.-S. Chen, “Exploiting Low-Rank Structures from Cross-Camera Images for Robust Person Re-Identification,” in IEEE ICIP, 2014.

4658

GRAPH REGULARIZED LOW-RANK MATRIX ...

also advance learning techniques to cope with the visual ap- ... Illustration of robust PRID. .... ric information by enforcing the similarity between ai and aj.

2MB Sizes 5 Downloads 251 Views

Recommend Documents

Programming Exercise 5: Regularized Linear Regression ... - GitHub
where λ is a regularization parameter which controls the degree of regu- larization (thus ... Finally, the ex5.m script should also plot the best fit line, resulting in an image similar to ... When you are computing the training set error, make sure

Alternative Regularized Neural Network Architectures ...
collaboration, and also several colleagues and friends for their support during ...... 365–370. [47] D. Imseng, M. Doss, and H. Bourlard, “Hierarchical multilayer ... identity,” in IEEE 11th International Conference on Computer Vision, 2007., 2

Feature Selection via Regularized Trees
Email: [email protected]. Abstract—We ... ACE selects a set of relevant features using a random forest [2], then eliminates redundant features using the surrogate concept [15]. Also multiple iterations are used to uncover features of secondary

Regularized Latent Semantic Indexing
optimization problems which can be optimized in parallel, for ex- ample via .... edge discovery, relevance ranking in search, and document classifi- cation [23, 35] ..... web search engine, containing about 1.6 million documents and 10 thousand.

LOCALITY REGULARIZED SPARSE SUBSPACE ...
Kim, Jung Hee Lee, Sung Tae Kim, Sang Won Seo,. Robert W. Cox, Duk L. Na, Sun I. Kim, and Ziad S. Saad, “Defining functional {SMA} and pre-sma subre-.

Feature Selection via Regularized Trees
selecting a new feature for splitting the data in a tree node when that feature ... one time. Since tree models are popularly used for data mining, the tree ... The conditional mutual information, that is, the mutual information between two features

Efficient Computation of Regularized Boolean ...
enabled the development of simple and robust algorithms for performing the most usual and ... Some practical applications of the nD-EVM are also commented. .... Definition 2.3: We will call Extreme Vertices of an nD-OPP p to the ending ...

Curvelet-regularized seismic deconvolution
where y is the observed data, A is the convolution operator (Toeplitz matrix), .... TR-2007-3: Non-parametric seismic data recovery with curvelet frames, Geophys.

Knowledge Graph Identification
The web is a vast repository of knowledge, but automatically extracting that ... Early work on the problem of jointly identifying a best latent KB from a collec- ... limitations, and we build on and improve the model of Jiang et al. by including ....

Graph Theory.PDF
(c) Find the chromatic number of the Graph 4. given below. If the chromatic number is k, ... MMTE-001 3. Page 3 of 5. Main menu. Displaying Graph Theory.PDF.

Graph Algorithms
DOWNLOAD in @PDF Algorithms in C++ Part 5: Graph Algorithms: Graph Algorithms Pt.5 By #A#. Download EBOOK EPUB KINDLE. Book detail.

Robust Graph Mode Seeking by Graph Shift
ample, in World Wide Web, dense subgraphs might be communities or link spam; in telephone call graph, dense subgraphs might be groups of friends or families. In these situations, the graphs are usually very sparse in global, but have many dense subgr

Graph Coloring.pdf
Sign in. Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying.

VECTOR & MATRIX
Lists - arbitrary collections of objects of any type, e.g. list of vectors, list of ... ”R”contains different libraries of packages. Packages .... Numeric vectors. • Character ...

Graph Theory Notes - CiteSeerX
To prove the minimality of the set MFIS(X), we will show that for any set N ..... Prove that for a non-empty regular bipartite graph the number of vertices in both.

Graph Theory Notes - CiteSeerX
Let us define A = {v1,...,vm} and B = V (G)−A. We split the sum m. ∑ i=1 di into two parts m. ∑ i=1 di = C + D, where C is the contribution of the edges with both ...

Graph Theory.pdf
Page. 1. /. 2. Loading… Page 1 of 2. Page 1 of 2. Page 2 of 2. Page 2 of 2. Main menu. Displaying Graph Theory.pdf. Page 1 of 2.

Graph representations
Models for small world? ▫ Erdos-Renyi model. ▫ n nodes, each node has a probability p of ... Barabasi-Albert model. ▫ Graph not static, but grows with time.

graph-contraction.pdf
... loading more pages. Retrying... graph-contraction.pdf. graph-contraction.pdf. Open. Extract. Open with. Sign In. Main menu. Displaying graph-contraction.pdf.

Field Matrix - Kean University
Educational Leadership Department ... improvement of school programs and culture. 2.2 ... C. Explore the in-service and continuing educational opportunities.