Recognizing Interaction Between Human Performers Using ‘Key Pose Doublet’∗ Snehasis Mukherjee



Electronics and Communication Sciences Unit, Indian Statistical Institute 203 B.T. Road, Kolkata

Sujoy Kumar Biswas

Dipti Prasad Mukherjee

Electronics and Communication Sciences Unit, Indian Statistical Institute 203 B.T. Road, Kolkata

Electronics and Communication Sciences Unit, Indian Statistical Institute 203 B.T. Road, Kolkata

[email protected] [email protected]

[email protected]

ABSTRACT In this paper, we propose a graph theoretic approach for recognizing interactions between two human performers present in a video clip. We watch primarily the human poses of each performer and derive descriptors that capture the motion patterns of the poses. From an initial dictionary of poses (visual words), we extract key poses (or key words) by ranking the poses on the centrality measure of graph connectivity. We argue that the key poses are graph nodes which share a close semantic relationship (in terms of some suitable edge weight function) with all other pose nodes and hence are said to be the central part of the graph. We apply the same centrality measure on all possible combinations of the key poses of the two performers to select the set of ‘key pose doublets’ that best represent the corresponding action. The results on standard interaction recognition dataset show the robustness of our approach when compared to the present state of the art method.

Categories and Subject Descriptors I.4 [Image Processing and Computer Vision]: Applications

General Terms Experimentation

Keywords graph centrality, bag of words, human poses, human interaction, key pose doublet

1. INTRODUCTION Detecting human activities in video is an active research area [5, 12, 15]. Semantic analysis of human activities in videos leads to various vision-based intelligent systems, including smart surveillance systems, intelligent robots, action∗Area chair: Wei Tsang Ooi †Corresponding author

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Copyright 20XX ACM X-XXXXX-XX-X/XX/XX ...$10.00.

(a) (b) (c) (d) Figure 1: Examples of some interactions (a)hugging, (b)punching, (c)pushing, (d)kicking taken from [14] based human computer interfaces, etc. For instance, a good method to automatically distinguish suspicious or unusual activities such as punching, pushing or kicking from normal activities can be useful in places like rail station, airport or shopping mal. Multiple activities must be recognized, even when the background is non-uniform (for pedestrians and/or other moving objects). Figure 1 shows examples of some interactions. Several approaches exist in recognizing human action focusing on either low and mid-level feature collection ([5, 12]) or modeling the high level interaction among the features [8, 7]. Following bag-of-words model [5, 12], Niebles et. al., automatically learns the probability distributions of the visual words using graphical models like probabilistic Latent Semantic Analysis (pLSA) and Latent Dirichlet Allocation (LDA) to form the dictionary of visual words. Also there are good efforts of combining global and local video features for action recognition [15]. There are approaches studying pose specific video features [2, 4] but modeling visual senses associated with poses in videos is largely an unexplored research area. In [9], a new pose descriptor is proposed using a gradient weighted optical flow feature combining both global and local features. They mine the pose descriptors using Max-diff kd-tree [10] to form the initial dictionary of poses. The initial pose dictionary contains both ambiguous and unambiguous (or key poses) poses. The poses in the initial dictionary are ranked using the centrality measure (a measure of ambiguity) of graph connectivity [11] to select the key poses. The approach made by [9] is limited to solo performer, and since they bank on the repetitive nature of the human actions by introducing action cycles, their methodology does not perform well (see Section 3 for results) in case of interactions like punching (once) or pushing away. However, if we consider each individual separately, one can observe a recurring pattern in the motion patterns of limbs. But it complicates the inferencing as action cues from two different actions need to be combined for semantic analysis. Our methodology tries to utilize this cue, i.e., motion and pose pattern of individual performer, in order to build a pose based recognition model

Figure 2: Bipartite graph representing key poses for two performers for ‘punching’ interaction. Green ellipses show examples of ‘key pose doublets’ concatenating key poses of left and right performers. for interaction types. We make a two-fold contribution to enhance the approach of [9] for recognizing interaction between multiple human performers. First, we apply the procedure of [9] on each of the performers present in the video to select the key poses for both the performers. Second, for each action we primarily take all possible combination of key poses of the two performers to form ‘pose doublets’. We again apply the centrality measure of graph connectivity [11] to rank the ‘pose doublets’ and select the ‘key pose doublets’ best representing the corresponding interaction. Figure 2 shows examples of some key poses for the two performers in the interaction ‘punching’. The key poses are represented as nodes in the bipartite graph shown in Figure 2. After ranking the ‘pose doublets’ we select the ‘key pose doublets’ shown by the green ellipses. We show the accuracy of our approach compared to the state-of-the-art in Section 3 followed by conclusions in Section 4. Before that we discuss the proposed approach in the next Section.

2. PROPOSED METHODOLOGY The proposed methodology consists of three steps. First we discuss the method for constructing the initial dictionary. Next we discuss the process of extracting key poses for each person in the video. Lastly we describe the method for constructing ‘key pose doublet’.

2.1 Forming Initial Pose Dictionary We use multidimensional pose descriptor corresponding to each performer of each frame of an action video as suggested in [9]. We have used the method of Gilbert et. al. [3] to draw the best fit rectangular window around each performer. The initial pose dictionary is constructed [9] from the multidimensional pose descriptors in two steps. First, the descriptors are derived by combining motion cue from the optical flow field (using Lucas-Kanade algorithm [6]) and pose cue from the gradient field of a particular video frame. Second, the pose descriptors, upon data condensation, result into a moderately compact representation and we call it the initial dictionary of visual poses. The initial pose dictionary contains pose clusters that may be equated with words in a document, where the document stands for a video in our version of the bag-of-words model. In the following, we briefly describe the descriptor extraction process as well as the formation of initial pose dictionary. The pose descriptor combines the benefit of motion information from optical flow field and pose information from the gradient field. The optical flow field is weighted with the strength of the gradient field to produce a resultant flow field. Next the resultant flow vectors are quantized into 8 bins of an angular histogram. Now this same process is car-

ried out for three different layers, where first layer is the whole image matrix, second layer splits the image into four equal blocks and the third layer splits it into 16 blocks each producing 8 bin vector resulting to a 168-dimensional pose descriptor after concatenation. Once we evaluate the pose descriptor, a kd-tree based data condensation technique eliminates much of the redundancies in the repetitive motion pattern of human actions and provides a moderately compact pose dictionary. The idea of data condensation is preferred to clustering because in data condensation one may afford to select multiple representatives from the same cluster, whereas in clustering the objective lies in identifying true number of clusters. The leaf nodes of the kd-tree denote one pose cluster or the visual pose word; one can choose multiple samples from each leaf node to construct the initial pose vocabulary Sj = {pi ∈ ℜd |i = 1, 2, ..., k} where j = 1, 2, and d denotes the dimensionality of the pose descriptors with k representing the cardinality of Sj , j = 1, 2. The algorithm to construct the kd-tree is explained in details in [10]. In our experiment we choose the mean of each leaf of the kd-tree as our pose word and learn the dictionaries S1 , S2 of poses of the two persons. Next we outline the scheme for ranking the poses of S1 , S2 using centrality theory of graph connectivity.

2.2

Formation of Compact Dictionary

The poses in S1 and S2 are often ambiguous and our goal is to identify the unambiguous poses (or key poses) from both S1 and S2 and produce compact dictionaries Ξj , j = 1, 2. The poses from Sj , j = 1, 2 are placed in a graph as nodes and the edge between each two poses stands for the dissimilarity in terms of a semantic relationship between them, measured using some form of weight function [9]. Let ρ(u, v) denote how many times the pose words u, v ∈ Sj both occur together in all video sequences of a particular action type. Then the edge weight ω(u, v) is given by { ω(u, v) =

1 ρ(u,v)



when ρ(u, v) ̸= 0 otherwise

(1)

According to (1), the lower the edge weight, the stronger is the semantic relationship between the pose words. Next we calculate the eccentricity measure [16] of graph connectivity to measure the semantic difference between the poses in a pose graph. Floyd-Warshall algorithm [1] computes all-pairshortest path between pose nodes. If the distance d(u, v) between two pose words u, v ∈ Sj is the sum of the edge weights ω(u, v) on a shortest path from u to v in the pose graph, then the eccentricity e(u) of a pose u is given by e(u) = max{d(u, v) | v ∈ Sj },

(2)

for j = 1, 2. Hence we have ranked the poses in Sj by their eccentricity measure, used as a measure of ambiguity. For each performer in video sequences of each action type, we choose the N -best poses by selecting poses with N -lowest eccentricity in a pose graph. We call the selected poses as key poses and include the key poses in the compact dictionaries Ξj . Once we identify the key poses for a particular kind of action we repeat the same process for all kinds of actions. The key poses p1 , p2 , . . . , pk extracted from all the action types are

grouped together to form Ξj . Ξj = {p1 , p2 , . . . , pk } ∀ pi ∈ Sj , i = 1, 2, . . . , k j = 1, 2. (3) Next we discuss the process of selecting ‘key pose doublet’.

2.3 Selecting ‘Key Pose Doublets’ We have two compact dictionaries Ξj , j = 1, 2 for each performers. These two dictionaries contain key poses of the corresponding action type. Each of the key poses can best represent the pose of a single performer during the corresponding interaction with the other performer. Our next task is to model these poses to represent the interaction between the two performers. An easy way for this is to make joint poses concatenating two 168-dimensional pose vectors ξ1 ∈ Ξ1 and ξ2 ∈ Ξ2 . Now the question is, which combination of poses (ξ1 , ξ2 ) we should take? To answer this question, we first take the all possible combinations (ξ1,i , ξ2,j ) where ξ1,i ∈ Ξ1 , i = 1, 2, . . . , m and ξ2,j ∈ Ξ2 , j = 1, 2, . . . , n, m and n being the cardinalities of Ξ1 and Ξ2 respectively. We call these concatenated combinations as ‘pose doublets’, where a set Y of mn ‘pose doublets’ represents a particular interaction type. For better recognition of the interaction, we have to choose a smaller set Ψ ⊆ Y of ‘pose doublets’. We call the elements of Ψ as ‘key pose doublets’. We represent the poses in Ξ1 and Ξ2 by a weighted undirected graph (Figure 2). The key poses are represented by nodes in the graph. There is no edge between the pose nodes of the same compact dictionary. Hence the weighted undirected graph constructed by poses in Ξ1 and Ξ2 is basically a bipartite graph. The two sets of poses may be thought of as a coloring of the graph with two colors: if we color all poses in Ξ1 red, and all poses in Ξ2 green, each edge has endpoints of differing colors, as is required in the graph coloring problem. But each pose node has an edge to all pose nodes of the other compact dictionary. In Figure 2, the two rows of nodes represent the pose nodes of the two different dictionaries Ξ1 and Ξ2 . If xi,j is the number of times the key poses ξ1,i ∈ Ξ1 and ξ2,i ∈ Ξ2 , i = 1, 2, . . . , m, j = 1, 2, . . . , n occur simultaneously in the video sequences of the same action, then the edge weight wi,j between ξ1,i and ξ2,i is defined as follows wi,j =

1 . xi,j

(4)

It is clear from (4) that the edge weight between any two poses indicate the likeliness of the two poses to occur simultaneously for the particular type of interaction. Better the likeliness of two poses of two different dictionaries, lower the edge weight between them. We rank the ‘pose doublets’ according to the edge weight between the corresponding pair of key poses. We choose the N -best ‘pose doublets’ to represent the corresponding interaction. We call them ‘key pose doublets’ for that interaction type and include them in Ψ. Hence learning phase of our approach is complete. Next we describe our experiments on the proposed approach and the results obtained from the experiments.

3. EXPERIMENTS AND RESULTS We get a complete set of ‘key pose doublets’ for all action types by taking the union of the Ψs of all types of interactions. For a training video, we just extract the ‘pose doublets’ and find the nearest ‘key pose doublet’ (in terms of Euclidean distance) from the union set of all ‘key pose dou-

Table 1: Confusion matrix of set 1 of UT-Interaction dataset (entries given in %) Hs Hg Kc Pt Pn Ps Hs 80 0 0 20 0 0 Hg 0 90 0 0 0 10 Kc 0 0 90 0 10 0 Pt 10 0 0 90 0 0 Pn 10 0 0 20 70 0 Ps 0 10 0 0 0 90 Table 2: Confusion matrix of set 2 of UT-Interaction dataset (entries given in %) Hs Hg Kc Pt Pn Ps Hs 70 0 0 10 20 0 Hg 0 80 0 0 0 20 Kc 0 0 80 0 20 0 Pt 10 0 0 70 20 0 Pn 10 0 10 20 60 0 Ps 0 20 0 0 0 80 blets’. We construct a histogram of the occurrences of the ‘key pose doublets’ in the video sequence. This histogram can be used as a signature of a particular interaction (as in bag of words model). The histogram is called the interaction descriptor. For a test video, we construct the interaction descriptor and find the closest match of the test descriptor among the descriptors of all the interactions. We test our method on a standard dataset known as UTInteraction dataset [14]. We find two sets of video data of 60 video sequences (two performers in each sequence) in each of them. Each set contains 10 video sequences of 6 different interactions ‘Handshaking (Hs)’, ‘Hugging(Hg)’, ‘Kicking(Kc)’, ‘Pointing(Pt)’, ‘Punching(Pn)’ and ‘Pushing(Ps)’. The only exception is, ‘Pt’, which is a single performer action. We consider the same performer as both left and right performers in case of ‘Pt’. Set 1 of the UT-Interaction dataset contains video with almost uniform background whereas set 2 contains more challenging video sequences with nonuniform background and poor lighting condition. We follow the leave-one-out experiment. We construct the interaction descriptors (as action descriptors in [9]) from the dictionary of ‘key pose doublets’ Ψ for recognition. The average time consumed for learning key poses and learning the pose doublet by each interaction amounts to little less than one minute. After the detection of ‘key pose doublets’, our approach takes only a few seconds for both learning and testing in our MATLABTM version 7.0.4 implementation in a machine with processor speed 2.37 GHz, 512MB RAM. We tested our method separately on both set 1 and set 2 of the UT-Interaction dataset. The confusion matrices for set 1 and set 2 are given in Tables 1 and 2 respectively. Table 3 shows the efficacy of the proposed approach over the method of Ryoo et. al. [13]. Accuracy is measured by the percentage of correct detections. The first and second rows of Figure 3 represent key poses of the left and right performers respectively for the interacTable 3: Recognition accuracy on UT-Interaction dataset compared to [13] (accuracy given in %) Interactions Hs Hg Kc Pt Pn Ps Overall [9] 50 60 30 60 20 40 43.33 [13] 75 87.5 75 62.5 50 75 70.8 Proposed 75 85 85 80 65 85 79.17

5.

(a)

(b)

(c)

(d)

(e) (f) (g) (h) Figure 3: Top and bottom rows show the key poses for left and right performer respectively for interaction ‘Hs’. (a)sequence 2, frame 53, (b)sequence 4, frame 48, (c)sequence 5, frame 41, (d)sequence 10, frame 48, (e)sequence 2, frame 54, (f ) sequence 6, frame 36, (g)sequence 8, frame 60 and (h)sequence 9, frame 33 from UT-Interaction dataset Table 4: ‘Key pose doublets’ (K) for interaction ‘Hs’ with their centrality values (C) obtained by (4) Figures of (K) (C) Figures of (K) (C) 3(a), 3(e) 0.009 3(b), 3(h) 0.018 3(d), 3(h) 0.014 3(a), 3(f) 0.019 3(d), 3(g) 0.015 tion ‘Hs’ of the UT-Interaction dataset. Table 4 shows the ‘key pose doublets’ selected for the interaction ‘Hs’ with their centrality values. For interaction ‘Hs’ we get 4 key poses for both the performers yielding 16 possible combinations of ‘pose doublets’. We rank them according to the centrality measure given by (4) and select the top 5 as ‘key pose doublets’. Figure 4 shows how the percentage of accuracy changes over the number of ‘key pose doublets’ per interaction. We get highest accuracy when number of ‘key pose doublets’ is 5 per interaction (i.e., N = 5).

4. CONCLUSIONS We develop an efficient method to recognize the interaction between human performers in video. From an initial vocabulary of poses, the proposed approach builds a small but highly discriminatory dictionaries of key poses for two different performers. By the notion of centrality theory of graph connectivity we extract the key poses for each performers. Among the all possible combinations of key poses for two different performers, we select most relevant pairs of poses called ‘key pose doublets’ to form a dictionary. In future, we want to fix a suitable threshold to select the ‘key pose doublets’. Presently our algorithm works for two performers; extending it to recognize multiple performers in the same scene may be another future research direction. 100 95 90

Accuracy (%)

85 80 75 70 65 60 55 50

1

2

3 4 5 6 Number of ‘key pose doublets’ per interaction

7

Figure 4: Accuracy plot with number of ‘key pose doublets’ per interaction for UT-Interaction dataset.

REFERENCES

[1] T. H. Cormen, C. E. Leiserson, R. L. Rivest, and C. Stein. Introduction to Algorithms. MIT Press, 2003. [2] L. Fengjun and R. Nevatia. Single view human action recognition using key pose matching and viterbi path seraching. In CVPR. IEEE Computer Society, 2007. [3] A. Gilbert and R. Bowden. Multi person tracking within crowded scenes. In conference on Human motion: understanding, modeling, capture and animation, pages 166–179. Springer, 2007. [4] N. Ikizler and P. Duygulu. Histogram of oriented rectangles: A new pose descriptor for human action recognition. Image and Vision Computing, 27(10):1515–1526, September 2009. [5] J. Liu, S. Ali, and M. Shah. Recognizing human actions using multiple features. In CVPR. IEEE Computer Society, July 2008. [6] B. D. Lucas and T. Kanade. An iterative image registration technique with an application to stereo vision. In International Joint Conference on Artificial Intelligence, pages 674–679. Morgan Kaufmann Publishers Inc., 1981. [7] G. Mori and J. Malik. Estimating human body configurations using shape context matching. In ECCV (volume 3) LNCS 2352, pages 666–680. Springer, January 2002. [8] G. Mori, X. Ren, A. Efros, and J. Malik. Recovering human body configurations: Combining segmentation and recognition. In CVPR (volume 2), pages 326–333. IEEE Computer Society, June 27-July 2 2004. [9] S. Mukherjee, S. K. Biswas, and D. P. Mukherjee. Modeling sense disambiguation of human pose: Recognizing action at a distance by key poses. In Asian Conference on Computer Vision, Volume 1, LNCS 6492, pages 244–255, November 2010. [10] B. L. Narayan, C. A. Murthy, and S. K. Pal. Maxdiff kd-trees for data condensation. Pattern Recognition Letters, 27(3):187–200, February 2006. [11] R. Navigli and M. Lapata. An experimental study of graph connectivity for unsupervised word sense disambiguation. IEEE T-PAMI, 32(4):678–692, April 2010. [12] J. C. Niebles, H. Wang, and L. Fei-Fei. Unsupervised learning of human action categories using spatial-temporal words. International Journal of Computer Vision, 79(3):299–318, June 2008. [13] M. S. Ryoo and J. K. Aggarwal. Spatio-temporal relationship match: Video structure comparison for recognition of complex human activities. In IEEE ICCV, 2009. [14] M. S. Ryoo and J. K. Aggarwal. UT-Interaction Dataset, ICPR contest on Semantic Description of Human Activities (SDHA). http://cvrc.ece.utexas.edu/ SDHA2010/ Human Interaction.html, 2010. [15] Y. Wang and G. Mori. Hidden part models for human action recognition: Probabilistic vs. max-margin. IEEE T-PAMI, 33(7):1310–1323, July 2011. [16] D. B. West. Introduction to Graph Theory. Prentice Hall, 2000.

Recognizing Interaction Between Human Performers ...

ABSTRACT. In this paper, we propose a graph theoretic approach for rec- ognizing interactions between two human performers present in a video clip.

1MB Sizes 0 Downloads 259 Views

Recommend Documents

Human - Computer Interaction - IJRIT
Human–computer interaction (HCI) is the study of interaction between ... disciplines, linguistics, social sciences, cognitive psychology, and human performance are relevant. ... NET, or similar technologies to provide real-time control in a separat

Human - Computer Interaction - IJRIT
disciplines, linguistics, social sciences, cognitive psychology, and human .... computing continue to grow, related HCI career opportunities will grow as .... but on expensive cameras this filter is usually applied directly to the lens and cannot be.

Interaction between monetary policy and macroprudential policies ...
Oct 6, 2015 - dialogue.html). ... 2. MACRO PRUDENTIAL TOOLS. 7. 3. INTERNATIONAL ... There are strong arguments to formally separate the two.

Understanding Interaction Differences between ...
interaction, tools can be built or tuned to support program- mers, such as recommending artifacts. For example, based on a Degree-Of-Interest model that favours recent interac- tion, Mylyn ... republish, to post on servers or to redistribute to lists

Interaction between equatorially symmetric and asymmetric tropical ...
Feb 3, 2010 - Interaction between equatorially symmetric and asymmetric tropical eastern Pacific SSTs. Soon-Il An & Jung Choi. Received: 23 September ...

[Read PDF] Interaction Design: Beyond Human-Computer Interaction ...
In this article we touch briefly on all aspects of Interaction Design : the deliverables, guiding principles, noted The following principles are fundamental to the design and implementation of effective interfaces, whether for traditional GUI environ

PDF Interaction Design: Beyond Human-Computer Interaction Best ...
Human-Computer Interaction Best EPUB Book By. #A#. Books detail. Title : PDF Interaction Design: Beyond q. Human-Computer Interaction Best EPUB Book By.

unit 11 human interaction with environment - eGyanKosh
instructional objectives using the locally available resources. At the end of ..... 3. Give at least two examples from primary occupations and tertiary occupations.