POSE-ROBUST REPRESENTATION FOR FACE VERIFICATION IN UNCONSTRAINED VIDEOS Hoang Anh B. Nguyen, Wen Li School of Computer Engineering, Nanyang Technological University, Singapore ABSTRACT Face recognition in unconstrained videos has been actively studied with the increasing popularity of surveillance and personal cameras. Compared with the traditional task of face recognition in images, recognizing faces in unconstrained videos is much more challenging due to the large variations in poses, expressions and lighting conditions. To handle varying poses in the videos, in this paper, we propose a two-level representation approach for face verification in unconstrained videos. Specifically, we first recover the full-pose face representation for each video which contains all the pose categories ranging from the most left to the most right profile faces. The missing poses are synthesized using keyframes of known ones. Then, we further propose a cross-pose video pair representation for face verification task, which consists of the similarity scores of all frame-level pairs across two videos. Extensive experiments on the benchmark YouTube Faces (YTF) dataset clearly demonstrate the effectiveness of our proposed method. Index Terms— Face recognition, video representation. 1. INTRODUCTION With the rapid adoption of surveillance systems, digital cameras and handphones, the face recognition problem has been widely studied over the last decades. Generally, face recognition applications fall into two categories: face identification which aims to identify the subject of each face, and face verification which aims to decide whether a pair of faces are from the same subject. While many face recognition systems have shown promising results under restricted environments with cooperative users, face recognition under unconstrained environments remains challenging due to large variations in poses, lighting conditions, expressions, and so on. Many works have studied the image-based face identification [1, 2, 3, 4, 5, 6] and face verification [7, 8, 9, 10] problems under unconstrained environments. However, for the videobased face recognition, previous works mainly focus on the face identification problem [11, 12, 13, 14, 15, 16], only few works [17] have been proposed for the face verification under unconstrained environments. Under unconstrained settings, face recognition in videos is much more challenging than that in images because videos

978-1-4799-2341-0/13/$31.00 ©2013 IEEE

usually contain more intra-class variations in terms of poses, expressions and lighting conditions. Among all these difficulties, pose variation is considered to be most challenging as pointed out in previous works [18, 19]. Different videos may contain different poses and the number of keyframes for each may be different as well. As a result, the two videos of different persons with a overlapped set of poses may appear more similar, when compared with the videos of captured under completely different viewpoints. Hence, using traditional methods to directly compare the videos may not achieve promising results. In this paper, we present a new method for face verification in unconstrained videos. To explicitly handle the large pose variance in different videos, we first recover full-pose representation for each video which covers all the poses from the most left to the most right profile faces. Specifically, the keyframes are first divided into k predefined pose categories according to the detected head poses. For pose categories with no keyframes, we synthesize face frames using either a flip operation or a generative model with keyframes of known poses. After that, the keyframes in each pose category are averaged to obtain one representative frame, which both simplifies the problem and reduces the possible variances in expressions and lighting conditions. The full-pose representation helps to capture the pose structure of each video. However, we are more interested in the specific face verification task in which face matching is performed between each pair of videos. In order to better utilize the information of different poses during the matching stage, we further propose a pair-level cross-pose representation for each pair of videos. In particular, we use k × k classifiers targeting at k × k different pose combinations between any two videos. For each pair of input videos, we populate the k × k pairs of keyframes between them and use the corresponding classifier to obtain the output similarity. Then, the k × k outputs from these classifiers are concatenated into a single feature representation for pose-robust face verification. Finally, an SVM classifier is trained on the this feature representation to perform the face verification task. Figure 1 shows the pipeline of our proposed method for a pair of input videos. Our method significantly outperforms the state-of-the-art methods on the YouTube Faces (YTF) dataset. Our proposed

3715

ICIP 2013

Fig. 1. An overview of our face verification scheme. Given the video pair, we first extract the full-pose representation for each video, using either flip operation (frame with yellow cross) or transfer operation (white cross) to fill in all missing pose categories. Using multiple cross-pose classifiers, we obtain the similarity scores between every frame-level pairs between the two representations and concatenate them together to form the high-level feature vector. Finally, an SVM classifier is used to learn the threshold to decide whether the two videos are from the same subject. scheme achieves 79.12%, 79.58% and 80.92% in terms of recognition rate by using the intensity, LBP and Gabor features, respectively. It can further achieves 82.24% by fusing different types of features, which outperforms the published state-of-the-art result of 76.4% [17]1 .

patches in addition to the pixel surrounding. In [7], Kumar et al. proposed to use the face attributes and the similarity of faces to specific reference people as features. Taigman et al. [23] also utilized the similarity scores of one face with another training on different negative sets as a high-level representation.

2. RELATED WORKS Video-based face recognition can be seen as a problem of measuring the similarities between two video sequences. Previous methods can be generally divided into two main directions considering the use of temporal information of the video. The first one extensively makes use of the spatiotemporal information and dynamic structure of the video when modelling the video sequence. The temporal and face motion have been encoded in the learning of Hidden Markov Models [20] or joint probabilistic model [21]. In the second direction, without the temporal information each video is transformed into an unordered set of images. Researches in this direction modelled the image sets as parametric distributions [15], non-parametric subspaces [14] or manifolds [13, 16]. There are two major components in a verification framework: face representation and face matching. Correspondingly, most previous works can be divided into two main approaches, descriptor-based methods [22, 7, 23] and similaritybased methods [8, 9, 24, 25]. The former one mainly focuses on finding effective features for face representation while the latter one aiming at a novel similarity metric for comparing facial descriptors. Currently, descriptor-based methods have showed promising result for the face verification task. In [22], Wolf et al. proposed two simple extensions of the Local Binary Patterns feature so as to capture the relationship of neighbourhood 1 The result was from the official website of YTF at the time of submitting this paper. We also note there are better numbers when this paper is under review, however, our method still outperforms these methods.

3. OUR APPROACH In this section, we present the details of our proposed approach for video-based face verification task. The first step is to recover the missing poses for each video such that each video contains all face poses, which leads to our full-pose video representation. Thereafter, we extract the cross-pose video pair representation by concatenating the similarities of all frame-level pairs between the two input videos. 3.1. Full-pose video representation In unconstrained videos, different videos may contain persons in different poses and the number of keyframes may be different for each pose as well. If two videos of different persons have similar or overlapped set of poses, they may appear more similar than videos of the same person captured under completely different viewpoints. As a result, directly comparing the videos using traditional methods may not achieve satisfying results. To reduce such pose variation in different videos, we therefore propose to represent each video with the fullpose face representation, which covers all the poses from the most left to the most right profile. Formally, let us denote a video as V = {fi |m i=1 } where fi is a keyframe and m is the total number of keyframes. We first define k pose categories {Pj |kj=1 } which cover all head poses from the most left to the most right profile. Then, we estimate the head pose of each frame fi in the original video and allocate this frame to the corresponding pose category accordingly. However, a real-world video usually does not cover all of the k poses but rather only several of them; for example,

3716

in the news or interview videos, the person (narrator) is often captured in frontal view with only slight head movements. As a result, some categories may contain no keyframes, which means the pose variance between different videos still exists. To fill these categories, we therefore propose two operations to synthesize face frames for them. Flip Operation: Since the face is approximately symmetric, our first solution is to horizontally flip keyframes in the category on one side (e.g., the most left profile) to fill its corresponding category on the other side (e.g., the most right profile) if that is empty. In other words, we flip and allocate the keyframes in category Pj to the category Pk−j+1 if Pk−j+1 does not have any keyframes. With this simple operation, we can reduce the number of missing poses by more than 30%. Transfer Operation: To fill the remaining empty categories, we propose to use keyframes in the non-empty categories to synthesize virtual frames for the empty ones. Particularly, we employ the deformable model in [26]. Although their initial intention is for face alignment, the model can be applied to transfer the face appearance from one viewpoint to another. Similar to [26], we firstly obtain the response maps for 65 fiducial landmarks using the feature point detector in [27] and jointly optimize the locations using the subspace constrained mean-shifts. Then we define the transformation to rotate face frames in the non-empty pose categories to each empty pose category. Thus, we obtain a set of synthesized keyframes to fill each empty pose category. By using the above two operations, we have represented the video with a set of pose categories, each of which contains some keyframes with similar poses. To further reduce expression and lighting variations within each category, we propose to average all the keyframes in one pose category into a single face frame. Thus, we finally represent each video as a set of keyframes, each of which represents a distinct pose and all of which cover the poses from the most left to the most right profile, i.e., V = {rj |kj=1 } frames where rj is the average frame for the j-th pose category.

input videos. Since each video contains exactly k average keyframes corresponding to k pose categories, there are k × k frame-level pairs between the two videos. The similarity of a frame-level pair is the output of the classifier learnt from available training data. Recall that in one video, each keyframe represents a distinct pose, so the k × k frame-level pairs actually represent k × k cross-pose pairs of the faces from the input videos. As one universal classifier may not well model the similarity measurement for all cross-pose pairs, we propose to learn k × k cross-pose classifiers, each corresponds to one cross-pose combination. Formally, let us denote the classifier trained on a cross-pose pair (Pi , Pj ) as Cij , where Pi is the i-th pose in the first video and Pj is the j-th pose in the second video with i, j = 1, . . . , k. To train each individual classifier Cij , we employ the Side-Information Linear Discriminant analysis (SILD) [24] approach, which extends the traditional Linear Discriminant Analysis by defining the between-class and within-class scatter matrices using only the side information. In the training process, we divide frame-level pairs from the n training video pairs into k × k non-overlapped subsets {Gij }i,j=1,...,k . Specifically, for the video pair V1 = {r1i |ki=1 } and V2 = {r2j |ki=1 }, we populate all cross-pose frame-level pairs between the two videos as {(r1i , r2j )|ki,j=1 }. The pair of keyframes (r1i , r2j ) is allocated to the set Gij for training the classifier Cij . Therefore, each subset Gij contains n frame-level pairs, in which the first frame in each pair belongs to pose Pi while the second frame belongs to pose Pj . Following [24], the within-class scatter ij matrix Sij W and between-class scatter matrix SB for the pose combination (i, j) are respectively defined as: X Sij (r1i − r2j )(r1i − r2j )T , (1) W = (r1i ,r2j )∈S

Sij B

=

X

(r1i − r2j )(r1i − r2j )T .

(2)

(r1i ,r2j )∈D

Similar to LDA, the projection matrix for pose combination (i, j) is calculated by solving this optimization problem:

3.2. Cross-pose video pair representation In the face verification setting, one is given a set of video pairs divided into two categories, same-person and different-person pairs. Specifically, a same-person video pair (V1 , V2 ) ∈ S means that the two videos are from the same subject, while (V1 , V2 ) ∈ D means that the videos are from two different subjects. The task of face verification is to predict whether two unseen test videos are from the same person or not. The face verification task can be treated as a binary classification problem, in which each training sample is a pair of videos. As most traditional classification methods were designed for training samples represented as feature vectors, we propose to represent each video pair into one single high-level feature vector by using a cross-pose video pair representation. In this high-level feature vector, each entry is the similarity of a frame-level pair, i.e., a pair of keyframes between the two

wij = arg max u

uT Sij Bu uT Sij Wu

(3)

After projecting the original frame-level pair (r1i , r2j ) to a lower dimension using wij , the classification score sij can be obtained by calculating their cosine similarity. Finally, we concatenate the scores from all k×k frame-level pairs to form a feature vector of length k 2 for each pair of videos. Finally, a binary SVM classifier with RBF kernel [28] is trained on the cross-pose features obtained from the training video pairs. For any given pair of test videos V1 and V2 , we first represent them as a single feature vector and then use the SVM classifier to predict whether they are from the same person or not.

3717

4. EXPERIMENTAL RESULTS We conduct the experiments on the benchmark YouTube Faces (YTF) dataset [17] to evaluate our proposed method for face verification in unconstrained environment. The YTF dataset contains 3425 videos of over 1595 subjects. We follow the standard benchmark setting on the YTF dataset, which consists of 5000 video pairs equally divided into 10 independent splits, with 250 same-class pairs and 250 differentclass pairs in each split. In all experiments, each keyframe is cropped to the size of 150 × 80 pixels and allocated to 1 of 7 pose categories according to the detected head pose provided in YTF dataset. To obtain the full-pose video representation, we employ the pre-trained model from FaceTracker library2 to synthesize missing pose categories. For the cross-pose video pair representation, we adopt several well-known features in face recognition including intensity, Local Binary Patterns [29] and Gabor wavelet feature [30]. Due to memory limitation, we apply down-sampling and PCA with 98% energy preserved, which leads to about 500 dimensions for these feature vector. For the cross-pose classifiers projection, 30% of the columns are kept as suggested in [24]. Following the standard setting on YTF dataset, results are reported in terms of recognition rate and standard error by using 10-fold cross validation. There are two main components in our framework as (a) the full-pose video-level representation and (b) the crosspose pair-level representation for video verification. In order to evaluate the individual contribution of each part towards our final result, we adopt three experiments for comparison: • Original: This is the baseline without using both representations. Instead, one SILD classifier is trained on all frame-level pairs between two input videos and the video similarity is calculated by averaging classification scores of all pairs. • Video level representation: We employ the full-pose video representation in which each video is represented by 7 average keyframes across all pose categories. Similar to the previous baseline, one SILD classifier is used to make the prediction. • Pair level representation: Our unified representation consists of full-pose video descriptor and then learning the similarities between every cross-pose frame-level pairs to form the high-level feature. original video level pair level

Intensity 77.12±1.51 77.60±1.43 79.12±1.81

LBP 76.68±2.06 77.98±2.16 79.58±2.01

the results by around 1% from the baseline for all the feature types, which clearly demonstrates that our full-pose video representation better reduces the intra-person variance and helps to normalize all the pose differences between videos. Moreover, we also observe that using the pair level representation helps to boost the result by more than 1.5%. Since each element in the new feature vector is the similarity score extract from the corresponding cross-pose classifier, the resultant similarity is thus more robust to pose changes. The result reveals that the cross-pose feature is highly suitable for the verification task as it captures the similarities of two faces in every aspects. Finally, our proposed pose-robust representation improves the recognition rate over the original feature by 2.00%, 2.90% and 2.68% for the intensity, LBP and Gabor features, respectively. Finally, we combine all three feature types with an additional layer of SVM. Figure 2 shows the performance comparison of our proposed method to the state-of-the-art results, in which MBGS [17] is currently the best published result on YTF dataset obtained by training a discriminative classifier for each video sequence. We also include the ROC curves from two recent papers, APEM-FUSION [25] and STFRD+PMML [10], which are accepted for publication when this paper is under review. Our fusion method achieves the best recognition rate of 82.24 ± 1.14, which significantly outperforms the published state-of-the-art of 76.4 ± 1.8 reported in [17] by 5.8% as well as all latest results from this year [31, 25, 10] by almost 3%. 5. CONCLUSION

Gabor 78.24±1.48 79.26±1.55 80.92±1.10

Table 1. Recognition rate of our representations and the baseline with different feature types. The results of the above three experiments are summarized in Table 1. The full-pose video representation helps to improve 2 https://github.com/kylemcdonald/FaceTracker

Fig. 2. Performance comparison on the YTF dataset.

We have introduced a novel two-level representation for video-based face recognition by leveraging the pose information. Our two main contributions include a video-level descriptor that helps to reduce intra-person pose variance and a video pair-level representation for face verification task. We achieve better performance than the original feature representation on the YouTube Faces dataset using different features. We also have shown that our proposed method significantly outperforms the current state-of-the-art results.

3718

6. REFERENCES [1] M. Turk and A. Pentland, “Eigenfaces for recognition,” Journal of cognitive neuroscience, vol. 3, no. 1, pp. 71– 86, 1991. [2] P.N. Belhumeur, J.P. Hespanha, and D.J. Kriegman, “Eigenfaces vs. fisherfaces: Recognition using class specific linear projection,” T-PAMI, vol. 19, no. 7, pp. 711–720, 1997. [3] J. Wright, A.Y. Yang, A. Ganesh, S.S. Sastry, and Y. Ma, “Robust face recognition via sparse representation,” TPAMI, vol. 31, no. 2, pp. 210–227, 2009. [4] Y. Huang, D. Xu, and T. J. Cham, “Face and human gait recognition using image-to-class distance,” T-CSVT, vol. 20, no. 3, pp. 431–438, 2010. [5] S. Yan, D. Xu, Q. Yang, L. Zhang, X. Tang, and H. Zhang, “Multilinear discriminant analysis for face recognition,” T-IP, vol. 16, no. 1, pp. 212–220, 2007. [6] Y. Huang, D. Xu, and F. Nie, “Patch distribution compatible semi-supervised dimension reduction for face and human gait recognition,” T-CSVT, vol. 22, no. 3, pp. 479–488, 2012.

[17] L. Wolf, T. Hassner, and I. Maoz, “Face recognition in unconstrained videos with matched background similarity,” in CVPR, 2011. [18] X. Zhang and Y. Gao, “Face recognition across pose: A review,” Pattern Recognition, vol. 42, no. 11, pp. 2876– 2896, 2009. [19] X. Liu and T. Chen, “Pose-robust face recognition using geometry assisted probabilistic modeling,” in CVPR, 2005. [20] X. Liu and T. Chen, “Video-based face recognition using adaptive hidden markov models,” in CVPR, 2003. [21] S. Zhou, V. Krueger, and R. Chellappa, “Probabilistic recognition of human faces from video,” CVIU, vol. 91, no. 1, pp. 214–245, 2003. [22] L. Wolf, T. Hassner, Y. Taigman, et al., “Descriptor based methods in the wild,” in Workshop on Faces in’Real-Life’Images: Detection, Alignment, and Recognition, 2008. [23] Y. Taigman, L. Wolf, and T. Hassner, “Multiple oneshots for utilizing class label information,” in BMVC, 2009.

[7] N. Kumar, A.C. Berg, P.N. Belhumeur, and S.K. Nayar, “Attribute and simile classifiers for face verification,” in ICCV, 2009.

[24] M. Kan, S. Shan, D. Xu, and X. Chen, “Sideinformation based linear discriminant analysis for face recognition,” in BMVC, 2011.

[8] Q. Yin, X. Tang, and J. Sun, “An associate-predict model for face recognition,” in CVPR, 2011.

[25] H. Li, G. Hua, Z. Lin, J. Brandt, and J. Yang, “Probabilistic elastic matching for pose variant face verification,” in CVPR, 2013.

[9] T. Berg and P.N. Belhumeur, “Tom-vs-pete classifiers and identity-preserving alignment for face verification,” in BMVC, 2012. [10] Z. Cui, W. Li, D. Xu, S. Shan, and X. Chen, “Fusing robust face region descriptors via multiple metric learning for face recognition in the wild,” in CVPR, 2013. [11] R. Chellappa, M. Du, P. Turaga, and S.K. Zhou, “Face tracking and recognition in video,” Handbook of Face Recognition, pp. 323–351, 2011. [12] U. Park, A.K. Jain, and A. Ross, “Face recognition in video: Adaptive fusion of multiple matchers,” in CVPR, 2007. [13] R. Wang, S. Shan, X. Chen, and W. Gao, “Manifoldmanifold distance with application to face recognition based on image set,” in CVPR, 2008. [14] G. Shakhnarovich, J. Fisher, and T. Darrell, “Face recognition from long-term observations,” ECCV, pp. 851–865, 2002. [15] O. Yamaguchi, K. Fukui, and K. Maeda, “Face recognition using temporal image sequence,” in FG, 1998.

[26] J.M. Saragih, S. Lucey, and J.F. Cohn, “Face alignment through subspace constrained mean-shifts,” in CVPR, 2009. [27] Y. Wang, S. Lucey, and J. F. Cohn, “Enforcing convexity for improved alignment with constrained local models,” in CVPR, 2008. [28] C-C. Chang and C-J. Lin, “LIBSVM: A library for support vector machines,” ACM Transactions on Intelligent Systems and Technology, vol. 2, pp. 27:1–27:27, 2011. [29] T. Ahonen, A. Hadid, and M. Pietikainen, “Face description with local binary patterns: Application to face recognition,” T-PAMI, vol. 28, no. 12, pp. 2037–2041, 2006. [30] C. Liu and H. Wechsler, “Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition,” T-IP, vol. 11, no. 4, pp. 467–476, 2002. [31] L. Wolf and N. Levy, “The svm-minus similarity score for video face recognition,” in CVPR, 2013.

[16] Z. Cui, S. Shan, H. Zhang, S. Lao, and X. Chen, “Image sets alignment for video-based face recognition,” in CVPR, 2012.

3719

pose-robust representation for face verification in ...

School of Computer Engineering, Nanyang Technological University, ... two-level representation approach for face verification in un- .... available training data.

424KB Sizes 0 Downloads 238 Views

Recommend Documents

Learning Prototype Hyperplanes for Face Verification in the Wild
(referred to as the prototype hyperplane) of one Support Vector. Machine (SVM) model, in which a sparse set of support vectors is selected from the unlabeled ...

Total Variability Modelling for Face Verification
Sep 29, 2012 - In this work, we extend the application of i-vectors beyond the domain of ..... Recent progress in speaker verification has seen the development of .... handheld mobile devices, this could conceivably have been caused by ...

Face Verification using Gabor Filtering and Adapted ...
2École Polytechnique Fédérale de Lausanne, Switzerland. 1. Abstract. ▷ Face authentication is difficult because of the high-variability of face images.

Face Verification using Gabor Filtering and Adapted ...
This is a Gaussian whose center is at a distance f from the origin with an angle θ from the horizontal axis. .... instead fed with Gabor images as inputs. ... of the protocols, enrollment and test data is selected in matching scenarios, that is from

Exploring Instructional Practices in Face-to-Face and Hybrid Courses
Education Commission of the States, and The Johnson Foundation provides evidence of its application to a broad range of undergraduate curricula and learning ..... When instructors required students in hybrid classes to read an article, however, they

Multithread Face Recognition in Cloud
Oct 12, 2016 - biometrics systems have proven to be an essential security tools, ... showcase cloud computing-enabled face recognition, which utilizes ...

Multithread Face Recognition in Cloud
Oct 12, 2016 - this responsibility to third-party vendors who can maintain cloud ... showcase cloud computing-enabled face recognition, which utilizes ...

Extraction Of Head And Face Boundaries For Face Detection ieee.pdf
Loading… Whoops! There was a problem loading more pages. Whoops! There was a problem previewing this document. Retrying... Download. Connect more apps... Extraction Of ... ction ieee.pdf. Extraction Of H ... ection ieee.pdf. Open. Extract. Open wit

An Approach For Integrity Verification In Multi Cloud Storage ... - IJRIT
using virtual infrastructure management (VIM) , a multi-cloud allows clients to easily ... These tools help cloud providers construct a distributed cloud storage ...

Extraction Of Head And Face Boundaries For Face Detection.pdf ...
Extraction Of Head And Face Boundaries For Face Detection.pdf. Extraction Of Head And Face Boundaries For Face Detection.pdf. Open. Extract. Open with.

An Approach For Integrity Verification In Multi Cloud Storage ... - IJRIT
IJRIT International Journal of Research in Information Technology, Volume 2, Issue 7, July 2014, Pg: 100-105. Vinitha Varghese ... In this paper, we address the ... Since cloud computing environment is constructed based on open architectures ...

Improving Face Recognition in Real Time for Different head Scales
Abstract. Face recognition is technique which is used to identify person by their face. Identifying the person is performed with different types of Biometric like iris scan, finger print, face recognition, Gait recognition; Signature etc. face recogn

Examples in Modular representation theory.
In particular we denote K for a field of characteristic. 0 and k for a field of characteristic p. Typically we have O, a DVR with maximal ideal p, K it's field of fractions ...

Improving Face Recognition in Real Time for Different head Scales
There are many problems are occurred when it perform face recognition like illumination, light intensity, blurred face, noisy image, tilted face, different head pose ...

Representation in case-based reasoning
e-mail: [email protected]. 2Georgia Institute of Technology, Atlanta, GA 30032-0280, USA; e-mail: [email protected] .... automated. The set of ...

PatientsLikeMe: Empowerment and Representation in ...
Online health communities, social networking, personal ... Additionally, the collection and sale of user data prompts ... Examples include patient websites.

Face-to-Face Tutorials in ODE and Student Satisfaction in Indonesia
Tutorial - a service in teaching and ... at least good services that can fulfill their ... Tutors need to have good ability .... they feel a service provider should offer ..... HERMAN is at the Universitas Terbuka, http://www.ut.ac.id, Indonesia. ema

PRODUCT REPRESENTATION FOR DEFAULT ...
ISP(4) = HSP(4) and that this is the equational class DB of distributive bilattices ...... piggyback dualities and applications to Ockham algebras, Houston J. Math.

Medication Verification Form for Physicians.pdf
There was a problem previewing this document. Retrying... Download. Connect more apps... Try one of the apps below to open or edit this item. Medication ...

Semantic Visualization for Spherical Representation
KDD'14, August 24–27, 2014, New York, NY, USA. Copyright is held by the ...... co re. Number of Topics Z. SSE. PLSV. SAM. LDA a. 20News b. Reuters8.

Employing structural representation for symbol ...
Jun 2, 2010 - Some remarks. Part2: Content based (focused) retrieval. Experimentation. Conclusion. - 2. Plan. Part 1. Representation and recognition of graphics content in line drawing document images. Part 2. Unsupervised indexation and content base

Verification Methodology for DEVS Models
that have been used in literature for verification and testing of DEVS models, ..... DEVS or TA models, but relies on exhaustive testing through simulation of ...

Face Recognition in Videos
5.6 Example of cluster containing misdetection . .... system which are mapped to different feature space that consists of discriminatory infor- mation. Principal ...

Face Authentication /Recognition System For Forensic Application ...
Graphic User Interface (GUI) is a program interface item that allows people to interact with the programs in more ways than just typing commands. It offers graphical icons, and a visual indicator, as opposed to text-based interfaces, typed command la