2010 International Conference on Pattern Recognition

Boosting Clusters of Samples for Sequence Matching in Camera Networks Valtteri Takala, Yinghao Cai, Matti Pietik¨ainen Machine Vision Group University of Oulu Oulu, Finland {vallu, yinghao, mkp}@ee.oulu.fi

A large number of different object features (shape, color, texture, etc.) are extracted from background subtracted tracking sequences to assemble a comprehensive template description of the object. These are then clustered using k-means clustering to get a few descriptive samples representing the final object model for view-to-view re-recognition. The mean clusters are utilized to address the temporal dimension and problems caused by momentary occlusion and sudden lighting changes. The features themselves are also Gaussian filtered or moving averages, thus improving the model’s resistance against rapid changes. The proposed learning algorithm combines Gentle AdaBoost [4] and classification and regression trees (CART) [1] with the suggested object model to create a classifier for each individual object. After learning phase, each cluster extracted from a new object sample sequence (provided by an object tracker, for example) is classified separately, and the final outcome for the sequence is determined by voting. The sequence matching approach is tested on a demanding dataset of 16 different human objects appearing multiple times in an IP camera network of five cameras installed in an office environment.

Abstract—This study introduces a novel classification algorithm for learning and matching sequences in view independent object tracking. The proposed learning method uses adaptive boosting and classification trees on a wide collection (shape, pose, color, texture, etc.) of image features that constitute a model for tracked objects. The temporal dimension is taken into account by using k-mean clusters of sequence samples. Most of the utilized object descriptors have a temporal quality also. We argue that with a proper boosting approach and decent number of reasonably descriptive image features it is feasible to do view-independent sequence matching in sparse camera networks. The experiments on real-life surveillance data support this statement. Keywords-boosting; recognition; sequence matching; camera networks

I. I NTRODUCTION Object tracking [15] is one of the basic challenges in computer vision. In camera networks, object tracking extends from single viewpoint tracking to following the target across multiple camera views. The challenge is great as the views, illumination, shadows and many other aspects of image are different. Many of these obstacles can be tackled by calibration and overlapping cameras, which are, though, very laborious to maintain. In uncalibrated camera networks, the challenge of object tracking becomes significantly harder. The track must be maintained without the extra information and constraints provided by the calibration procedures. One of the first studies related to view-independent tracking in camera networks was done by Huang et al. [7] who presented a probabilistic approach for object re-identification with two cameras. Bayesian formalization and linear programming were employed by Kettnaker and Zabih [8] to reconstruct the paths of moving objects in multi-camera environment. Porikli and Divakaran [10] used a color calibration model to estimate the optimal alignment function between the object appearance histograms in different views and a Bayesian belief network for reasoning the correspondences. More recently, Chen et al. [2] proposed an adaptive learning method that uses spatio-temporal and appearance cues for tracking objects across multiple cameras with disjoint views. In this paper, we introduce a machine learning approach for object re-recognition in an uncalibrated camera network. 1051-4651/10 $26.00 © 2010 IEEE DOI 10.1109/ICPR.2010.106

II. O BJECT M ODEL Visual cues define objects often better than their raw images. Our object model consists of a versatile collection of different features: Color, as being one of the most intuitive cues for human perception, is described by the standard color histogram [11], color correlogram [6] and two different opponent color space based histogram descriptor variants [13]. The chosen four color histograms define object’s color characteristics in two different color spaces and are also invariant to translation and rotation. Opponent colors have the following relation to RGB values:   O1  O2  =   O3 

R−G √ 2 R+G−2B √ 6 R+G+B √ 3

  ,

(1)

in which the intensity information is in channel O3 and the color is stored in O1 and O2 . These two color models are shift-invariant with respect to light intensity. While O1 and O2 still contain some intensity information, we can divide 404 400

Feature Contour area Width-height ratio: latest and avg. Object ellipse axis angle Temporal difference: min., latest and avg. Bounding box: x, y (upper left corner), width, height Ellipse bounding box: x, y (center), width, height Color histogram 1 & 2 (for the upper and lower half of the bounding box and foreground) Color correlogram 1 & 2 (upper and lower) LBP 1 & 2 (upper and lower) Opponent color 1 & 2 (upper and lower) W invariant 1 & 2 (upper and lower)

one-vs-others or one-vs-one approach. We selected the onevs-others as our classifier scheme because it offers certain benefits. First of all, we need to train only one classifier per class. The cascade structure of the whole classification algorithm is also smaller and faster as there are less strong classifiers to go through while evaluating a sample. These advantages, of course, come at the expense of discrimination power if compared to one-vs-one approach. For matching the tracking sequences in dynamic camera networks, we propose a classification and learning algorithm based on Gentle AdaBoost [4] and CART decision trees [1] in one-vs-others manner. The algorithm uses kmeans clustering of sequential samples and voting between the clusters for final decision making. The detailed structure of a classifier is shown in Algorithm 1. First we cluster sequential samples xji (i = 1, 2, ..., N, j = 1, 2, ..., S) using k-means clustering and train a boosted classifier on the obtained clusters cki (i = 1, 2, ..., N, k = 1, 2, ..., K), where N is the number of training sequences, S the samples in a sequence, and K the corresponding clusters. After training, the classification of new clusters of samples xj is carried out using the learned classifier C and voting. The boosted CART classifier outputs for all mean clusters of the tracking sequence are summed up and the final decision is the sign [−1, 1] of the sum.

Size 1 2 1 3 4 4 432 256 512 1024 128

Table I O BJECT FEATURES .

them by O3 to get a descriptor called W invariant [5] which has more invariance to intensity changes. Another important visual cue is texture. In our experiments, we describe the textural properties of the image using the popular local binary pattern (LBP) method [9] which is invariant to monotonic changes in gray-scale and inexpensive to compute. Each histogram descriptor consists of two separate histograms extracted from the upper and lower half of the object foreground area. While subtracting an object from the background and tracking, many other, more straightforward measures are available. The temporal difference feature describes the motion inside the object as a normalized scalar sum of absolute differences on all color channels on all pixels inside the object contour. The differences are calculated between the last two frames. In addition to internal motion detection and description, this single valued feature can be understood as a sort of temporal texture descriptor as its value is heavily affected by the texture content inside the moving object contour. The shape and pose are taken into account by utilizing the contour area, bounding box width-height ratio and the angle between the length (major) axis of the object ellipse and the horizontal axis. Object contour and ellipse bounding boxes are also covered as features, even though they do not produce much useful information on objects which have large variation in size while being tracked inside a view and moving across the views. The main object descriptors exploited in this study are described in Table I. Any of the features alone does not offer enough discrimination power for object recognition across camera views but together they provide a wide coverage on many aspects of the object’s properties.

Algorithm 1 Boosted classification 1: Initialization: 2: - N training sequences (each consisting of S samples) xji with labels yi , i = 1, 2, ..., N, j = 1, 2, ..., S 3: - Number of k-means clusters K 4: - Number of boosting iterations I 5: 1. Learning: j 6: - Cluster training sample sequences xi to K mean clusters cki . 7: - Teach a boosted classifier C using clusters ck i and labels yi (correct labels are set to +1, others to -1), and iterate I times. 8: 2. Classification: 9: - Cluster a new sample sequence xj to K mean clusters ck , k = 1, 2, ..., K. 10: - Classify clusters ck using the learned classifier C. 11: - Sum classification results C(ck ) of all clusters ck . 12: - Get the final classification decision F (ck ) = PK sign( 1 C(ck )). IV. E XPERIMENTS The object sequence matching experiments were carried out on a dataset obtained using an object tracker [12] on a collection of surveillance videos. The following three subsections describe the testing environment, tracker parameters, and the performance evaluation dataset. The last one presents the classification tasks and results for Algorithm 1.

III. B OOSTED L EARNING AND VOTING Multi-class classification requires either a genuine multiclass capable classifier [3] or a proper combination of twoclass classifiers. In the latter, one can adopt, for example, a

401 405

A. Environment The software environment was composed of MS Visual Studio and Matlab. A 2.66 GHz Intel Core 2 Duo PC with 2 GB of memory was used for processing, data capturing and object tracking. A camera network of five Axis 210A IP network cameras was employed to produce the raw data for object tracking. The cameras were set in an indoor office environment where illumination conditions were relatively stable though not controlled. The views were effectively non-overlapping, meaning that only small image regions were covering the same areas and could be filtered out by the object tracker. The video frame size was a common 352 × 288 for all IP cameras and the frame rate range was 10-20 fps. The proposed classification algorithms were implemented using Matlab. In our experiments, we used the Gentle AdaBoost and CART implementations of GML AdaBoost Matlab Toolbox [14]. The decision trees were initialized with a maximum branching factor of two, which was kept constant in all the classification experiments. B. Feature Extraction Figure 1.

A one-view object tracker [12] was used for tracking and feature extraction. The tracker was driven with the same parameters as in the original study. The foreground/background ratio for object detection was set to 0.02 as was suggested by the authors for the detection of sizable objects. All tracking data was saved in XML format.

Example images of all image classes.

dominant cluster out of three available gives the outcome. The dominance is determined by a percentage of samples from the temporal center of the original tracking sequence. The last classifier configurations (Sample Position SP 5, 10 and 20) take advantage of single samples selected from the indicated positions from the original sequences. As can be seen from the figure, the voting and best cluster selection approaches are clearly superior with less than 6 % error rates that stabilize after about 50 boosting iterations. Other classifiers stay above 10 % error and are not able improve while iterations are increased. The voting algorithm reaches its maximum performance at less than 40 iterations, which indicates that only a few of the available over 2000 features are needed for building a single classifier. The frequency histogram of all features selected by all classifiers of the voting approach is represented in Figure 3. Many features of object descriptors have been selected, especially the LBP (bins 715-1226) and W invariant opponent color histograms (2251-2379). According to the figure, the color histogram (bins 27-458), color correlogram (459-714) and the basic opponent color histogram (1227-2250) provide relatively small amount of usable features, even though some have been frequently selected by the boosted learning trees. The isolated spike (bin 7) in the beginning of the histogram corresponds to object contour area. All temporal difference values (11-13) are also included in the collection of active features.

C. Description of Test Data The raw test data consist of 53 sample video sequences from five camera views. In these videos, 16 different objects (persons) appear in 1-5 views for at least once per view. The data was gathered during normal office working days. The video sequences were processed by the object tracker to obtain 53 XML sequence descriptions which contain the features of tracked objects for all frames. The histogram features were Gaussian filtered by the object tracker to remove noise. The sequence descriptions of individual objects were labeled with corresponding classes. Figure 1 displays example images of all 16 different tracked objects. D. Results The developed boosting algorithm (Algorithm 1) was evaluated on the whole dataset of 16 classes. Each class was tested separately against the other classes. The validation was based on a common 10-fold cross-validation procedure. Figure 2 shows the performance of the algorithm with and without k-means clustering. The selected classification schemes include the proposed cluster voting algorithm for one and three clusters (VotingCluster 1 and 3), a variant that makes the final classification decision according to the highest cluster output of three clusters (BestCluster 3), and a scheme (DC 25%, 50% and 75%) in which the

402 406

R EFERENCES

Classification Error 0.13 SP 5 SP 10 SP 20 DC 25 DC 50 DC 75 VotingCluster 1 VotingClusters 3 BestCluster 3

0.12

0.11

Error

0.1

[1] L. Breiman. Random forests. Machine Learning, 45:5–32, 2001. [2] K.-W. Chen, C.-C. Lai, Y.-P. Hung, and C.-S. Chen. An adaptive learning method for target tracking across multiple cameras. In CVPR, pages 1–8, 2008.

0.09

[3] Y. Freund and R. E. Schapire. A decision-theoretic generalization of on-line learning and an application to boosting. JCSS, 55:119–139, 1997.

0.08

0.07

[4] J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: A statistical view of boosting. The Annals of Statistics, 28:337–407, 2000.

0.06

0.05

0

10

20

Figure 2.

30

40

50 Iterations

60

70

80

90

100

[5] J.-M. Geusebroek, R. van den Boomgaard, A. Smeulders, and H. Geerts. Color invariance. IEEE TPAMI, 23:1338–1350, 2001.

Different sampling and voting schemes.

[6] J. Huang, S. R. Kumar, M. Mitra, W.-J. Zhu, and R. Zabih. Image indexing using color correlograms. In CVPR, pages 762–768, 1997.

Feature Selection Histogram 0.07 0.06

[7] T. Huang and S. Russell. Object identification in a bayesian context. In IJCAI, pages 1276–1283, 1997.

Frequency

0.05 0.04

[8] V. Kettnaker and R. Zabih. Bayesian multi-camera surveillance. In CVPR, pages 253–259, 1999.

0.03 0.02

[9] T. Ojala, M. Pietik¨ainen, and T. M¨aenp¨aa¨ . Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE TPAMI, 24:971–987, 2002.

0.01 0

27

459

715

Figure 3.

1227 Feature

2251

[10] F. Porikli and A. Divakaran. Multi-camera calibration, object tracking and query generation. In ICME, pages 653–656, 2003.

Boosted feature selection.

[11] M. Swain and D. Ballard. Color indexing. In ICCV, pages 11–32, 1990.

V. C ONCLUSIONS

[12] V. Takala and M. Pietik¨ainen. Multi-object tracking using color, texture and motion. In CVPR, pages 1–7, 2007.

This paper presents a novel approach for sequence matching in a camera network. The classifier algorithm uses k-means clustering, Gentle AdaBoost and CART decision trees to build classifiers for feature-based classification of object tracking sequences. The results show that there is a significant boost in performance while using multiple mean clusters of samples and voting instead of single isolated samples or clusters. They also indicate a positive effect of temporally oriented image features. For the future studies, we are interested in evaluating the proposed algorithm in a real-time tracking system. We would also like to compare our algorithm against another sequence matching algorithm if one was readily available.

[13] K. van de Sande, T. Gevers, and C. Snoek. Evaluation of color descriptors for object scene recognition. In CVPR, pages 1–8, 2008. [14] A. Vezhnevets. GML AdaBoost Toolbox. retrieved January 25, 2010, http://graphics.cs.msu.ru/science/research/.

Matlab from

[15] A. Yilmaz, O. Javed, and M. Shah. Object tracking: A survey. ACM Computing Surveys, 38:1–45, 2006.

ACKNOWLEDGMENT This study was supported by the Finnish Funding Agency for Technology and Innovation (TEKES) and the European Regional Development Fund (ERDF).

403 407

Boosting Clusters of Samples for Sequence Matching in ...

Abstract—This study introduces a novel classification algo- rithm for learning and matching sequences in view independent object tracking. The proposed learning method uses adaptive boosting and classification trees on a wide collection (shape, pose, color, texture, etc.) of image features that constitute a model for ...

639KB Sizes 0 Downloads 257 Views

Recommend Documents

Time Warping-Based Sequence Matching for ... - Semantic Scholar
The proliferation of digital video urges the need of video copy detection for content and rights management. An efficient video copy detection technique should be able to deal with spatiotemporal variations (e.g., changes in brightness or frame rates

Time Warping-Based Sequence Matching for ... - Semantic Scholar
The proliferation of digital video urges the need of ... proposed an ordinal signature generated by ... and temporal signature matching, which obtains better.

PARTIAL SEQUENCE MATCHING USING AN ...
where the end point of the alignment maybe be unknown. How- ever, it needs to know where the two matching sequences start. Very recently [6] proposed an ...

samples of business plans for small businesses pdf
samples of business plans for small businesses pdf. samples of business plans for small businesses pdf. Open. Extract. Open with. Sign In. Main menu.

Samples-of-phishing-mails.pdf
Loading… Page 1. Whoops! There was a problem loading more pages. Samples-of-phishing-mails.pdf. Samples-of-phishing-mails.pdf. Open. Extract. Open with.

Profile of impurities in polycrystalline silicon samples ...
Solar Energy Materials & Solar Cells 79 (2003) 347–355. Profile of impurities in polycrystalline silicon samples purified in an electron beam melting furnace.

Samples
http://digital-photography-school.com/99-remarkable-photographers-portfolios http://www.rleggat.com/photohistory/ http://www.artcyclopedia.com/. Cari Ann Wayman. Man Ray. Richard Misrach. Robert Frank. Robert Park Harrison. Sophie Calle. Scott Mutter

Exploiting Multi-core Architectures in Clusters for ...
Performance of the Parallel Bootstrap Simulation Algorithm. The 12th IEEE ... Hybrid programming model with MPI and OpenMP: ▻ fine grain (Hybrid-I).

Exploiting multi-core architectures in clusters for ...
connected via a Gigabit Ethernet network. ... ASP (Alternate Service Pattern) model describes an ... the queues and P the number of service patterns. FAS (First ...

samples of business plans pdf
Sign in. Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying.

Authentication of forensic DNA samples - Semantic Scholar
by an automatic search of the database (e.g. CODIS). ..... samples are routinely searched against these databases (e.g. by. Fig. 5. .... Int. 160 (2006) 90–101.

Instrumentness in Clusters of Artefacts -- a First ... - Semantic Scholar
Software is not an instrument in the same sense as a violin, but several instrument ... reference to Wartofsky's (1973) analysis of how human perception develops ...

Instrumentness in Clusters of Artefacts -- a First ... - Semantic Scholar
Collaboration, instrumentness, computer mediated creativity ... specifics of computer-mediated creativity. .... revisit ten collaboration-oriented lessons on complex.

SAMPLES OF GOALS AND OBJECTIVES FOR ...
from one telephone interview and one print or on-line resource to establish ... daily assignment in 4 classes by the second 6-weeks report card period. Example ...

Parallel generation of samples for simulation techniques ... - CiteSeerX
The current SAN solver, PEPS software tool [4], works with less than 65 million ..... multiprocessor technology machine and perform the three simulation approaches for a ..... ular Analytical Performance Models for Ad Hoc Wireless. Networks.

Parallel generation of samples for simulation ...
Ricardo M. Czekster, Paulo Fernandes, Afonso Sales, Dione Taschetto and Thais Webber. ∗. Pontifıcia ... The current SAN solver, PEPS software tool [4], works.

PRIZM Clusters
When relocating due to a job transfer, people usually find a neighborhood that is very ... Claritas uses factor analysis of census data to uncover the demographic and lifestyle variables that explain the ..... Their diligence has enabled them to.

order matters: sequence to sequence for sets - Research at Google
We also show the importance of ordering in a variety of tasks in which seq2seq has ..... In Advances in Neural Information Processing Systems, 2010. 10 ...

Cross-lingual Word Clusters for Direct Transfer of Linguistic ... - SODA
guistic structure prediction (Smith, 2011). Regret- ... Smith, 2009; Gillenwater et al., 2010; Naseem et al., ...... Scott Miller, Jethran Guinness, and Alex Zamanian.

Stability mechanism of cuboctahedral clusters in UO2+x
May 20, 2008 - Color online a Cuboctahedral cluster COT-v incor- porated .... x = 2 VU − IU. + IO + ... FIG. 2. Color online Defect concentrations of point oxygen.

Parallel generation of samples for simulation ... - Semantic Scholar
Analytical modeling of complex systems is crucial to de- tect error conditions or ... The current SAN solver, PEPS software tool [4], works with less than 65 million ...

Parallel generation of samples for simulation ... - Semantic Scholar
This advantage justifies its usage in several contexts where .... The main advantage of. SAN is due to ..... ular Analytical Performance Models for Ad Hoc Wireless.

Online Management of Jobs in Clusters using Virtual ...
the allocated resources at all times. On the opposite, dynamic scheduling algorithms manipulate in live the state of the jobs using preemption and the location of their tasks using process migration. These algorithms provide a finer use of resources