How Do We Use Our Hands? Discovering a Diverse Set of Common Grasps (Supplementary Material) De-An Huang, Minghuang Ma∗, Wei-Chiu Ma∗ , and Kris M. Kitani The Robotics Institute, Carnegie Mellon University {deanh, minghuam, weichium}@andrew.cmu.edu, [email protected]

In this supplementary material, we compare different hand regions harvesting techniques and feature representations. Additional clustering results and some pseudocodes of baselines in the paper are also included at the end.

(a) Random

(b) Center of view

1. Methods for Comparison Our goal is to discover the dominant modes of handobject interactions from first-person videos. The approach in our paper has four main components: (1) harvesting candidate hand-object regions, (2) extracting features from the candidate regions, (3) grouping the candidate regions to discover modes of hand-object interactions, and (4) modeling the temporal dynamics to capture higher-order relationships between sequences of candidate hand-object regions. In the paper, we use hand contour to harvest candidate hand-object regions and masked HOG descriptor as feature representation. In this section, we provide the details of the comparing methods.

(c) Objectness

(d) Eye gaze

(e) Hand

(f) Hand Eye

Figure 1. Candidate regions extracted by different methods.

1.1. Methods for Harvesting Hand-Object Regions

5. Hand contour: a bounding box around the hands is extracted. We detect hands at the pixel level with [2], using code obtained from the authors. It computes a hand probability value for each pixel based on the color and texture of a local surrounding image patch. It then thresholds the probability values and extracts a set of connect components from each frame. The bounding boxes are centered around the hand contour such that the top most pixel of the contour is at the top center position of the box. Regions are also adjusted (shifted inward) so that they never exceed the image boundaries. 6. Hand-fixation: a bounding box region is extracted from a hand contour only if there is an eye fixation detected in the potential bounding box.

In order to mine clusters of hand-object interactions, we first need a means to robustly extract the key frames and bounding boxes that capture important hand-object regions. To this end, we evaluate six different harvesting methods: 1. Random sampling: a bounding box region is randomly sampled from every frame. 2. Center focus: a bounding box centered in the image is extracted from every frame. 3. Objectness: high scoring bounding box(es) (confidence over 95%) are extracted using a categoryindependent object detector [1]. These candidate regions are likely to contain an object of any category. 4. Gaze fixation: a bounding box is extracted when an eye gaze fixation point is detected by an eye tracker. ∗ indicates

Methods (2), (4) and (6) are based on an attention-based perspective where we assume that a person’s focus of attention will help identify key hand-object regions. Method (3)

equal contribution

1

Hand Contour descriptor

Masked HOG descriptor

Figure 2. Visualization of the hand contour and masked HOG feature descriptor used to represent candidate regions.

is a purely object-centric approach, while methods (5) and (6) take a hand-centric approach. With the exception of the objectness detector, which generates a bounding box at its own scale, we use a fixed size bounding box region of 350 × 160 pixels. We selected the size heuristically by observing a few qualitative examples. In practice, we found it to encompass various objects in the environment (cups, plates, utensils) well. The box is also wide enough to capture two hands interacting with an object (e.g., peeling an onion with two hands). Examples of harvested object regions are given in Fig. 1. Our experimental result (see Table 1) confirms quantitatively our intuition, that the location of the hands are indeed important for learning about hand-object interactions. Based on this result, we use hand locations to harvest handobject regions in the paper.

1.2. Representations of Hand-Object Interactions Given an incoming stream of candidate hand-object regions we would like to group similar interactions into the same clusters. Before we can proceed to group the regions, we are faced with the challenge of representation. We explore six different feature representations to understand which representation best preserves the invariants of a hand-object interaction: 1. A bag-of-words (BOW) over sparsely detected SIFT features is accumulated using a codebook size of 1000 (learned using k-means). 2. A BOW of cuboids [3], a spatio-temporal descriptor often used for action recognition, is generated using spatial interest points [7] and quantized with a codebook size of 2000 (learned using k-means). 3. A global histogram-of-optical-flow (HOF) is used to spatially quantize the motion of the candidate region into a 3 × 3 grid, with a temporal window size of 3 frames (i.e. two temporal bin dimensions) and the optical flow for each bin is quantized into 5 directions. 4. A binary mask of the hand region (Hand Contour) is used to capture the outer shape of the hand (Fig. 2). 5. A large histogram-of-oriented-gradients template (Global HOG) is computed for the entire hand-object

Table 1. Evaluation of candidate hand-object region generation using different bounding box extraction methods. Harvest Tech. OP OM OU Hand 0.967 0.911 0.956 Hand eye 0.922 0.867 0.900 Eye gaze 0.838 0.465 0.717 Objectness 0.900 0.456 0.733 Center 0.717 0.131 0.626 Random 0.700 0.156 0.633

region including the hands, objects and background. 6. A large HOG template is generated only for a masked region (Masked HOG), inspired by work in object discovery [5]. This representation removes the effect of the background (i.e. non-hand regions, including interactive objects) and uses only the contour of the hand to group the regions. We use the following HOG template parameters: 8 × 8 cell, 8 × 8 stride, 16 × 16 blocks with 9 gradient orientation bins (see Fig. 2). We compared all of the proposed feature representations (Table 2) and found that a joint feature obtained by merging a large gradient histogram template with an extracted hand contour (Masked HOG) is the strongest representation to model the functional proximity between candidate hand-object interactions. This result is interesting because it implies that the appearance of the hand is more important than the appearance of the object. Based on this result, we use Masked HOG as feature representation in the paper.

2. Experimental Evaluation We compare harvesting techniques and feature representations using the publicly available ego-centric activities dataset of Fathi et al. [4]. It consists of first-person videos captured by 9 users preparing 8 different dishes (i.e. American breakfast, hamburger, Greek salad, pasta, pizza, snack, Tilapia, turkey sandwich) in a kitchen environment for a total of 40 videos. The videos range from 7 to 18 minutes depending on recipe, which amounts to a total of about 0.8 million frames and 7 hours of video. This data is particularly well-suited for our task since ego-centric videos contain many naturally occurring interactions with typical kitchen countertop objects.

2.1. Comparing Harvesting Techniques We first evaluate the ability that each harvesting approach from Section 1.1 has on generating candidate handobject interaction regions. Since we cannot exhaustively label all hand-object interaction regions from over 7 hours of video, we instead evaluate on a randomly selected set of 90 candidate regions for each harvesting technique, which serve to represent the larger dataset. We use three metrics: object presence (OP), object uniqueness (OU), and object manipulation (OM) defined as:

Table 2. Comparative performance of different features. Features Purity Diversity Coverage Masked HOG 0.727 0.727 141 Hand contour 0.478 0.444 264 Global HOG 0.206 0.206 254 BOW (ST cuboids) 0.133 0.133 65 Global HOF 0.122 0.122 24 BOW (SIFT) 0.091 0.091 82

(a) spatio-temporal cuboids

(b) global HOF

(c) global HOG

our video frames often contain many static objects that are not being manipulated. (d) Hand contour

2.2. Comparing Feature Representations (e) Masked HOG

Figure 3. A hand-object interaction cluster discovered using different feature representations (left most image is the cluster center).

# visible objects , # candidate regions # unique visible objects OU = , # candidate regions # objects being manipulated OM = . # visible object regions OP =

(1) (2) (3)

The OP rate measures the visibility of an object in the bounding box and is the number of regions that contain an object divided by the total number of candidate regions. The OU rate measures the number of unique objects in a cluster and is defined as the number of unique instances divided by the number of candidate regions. A high OU rate means that a single grasp is associated to a diverse set of objects. The objectness detector, for example, is highly sensitive to plates so the OU rate tends to be low. The OM rate is a measure of visible manipulation and is defined as the percentage of times an object region is captured while it is being manipulated by the camera-wearer’s hand(s). This is the most important metric as we would like clusters to contain hand interacting with objects. Table 1 shows that the hand-based region extraction approach generates the best candidate regions with respect to all three metrics. In particular, the hand-based region extraction has a very high OM rate, as the existence of the hands creates a strong prior over the existence of an object that is undergoing manipulation. Somewhat surprisingly, selecting hand detections that also have eye fixations (Hand-eye) does not help in harvesting the hand-object regions. Upon close inspection, we found that eye fixations do not consistently coincide with hand-object interaction. The other baseline approaches perform much worse since they do not search for regions in which a hand is present. For example, while the objectness detector [1] can find object-like regions (thus having high OP rate), it obtains a low OM rate since

We next evaluate the features from Section 1.2 used to represent each detected hand-object region, by evaluating cluster quality using hand-object regions harvested by hand locations. As the feature representation will have a large effect on the similarity measure, we would like to understand the strengths and limitations of different image representations when attempting to group the regions. In this experiment, we compare several standard approaches used for object and action recognition to represent and cluster the hand-object interaction object regions. The metrics used for evaluation are: # same grasped objects , # candidate regions # unique objects diversity = , # visible object regions purity =

coverage = # discovered groups.

(4) (5) (6)

Purity and diversity are computed in the following manner. First, each of the clusters are ranked according to the entropy PM over the video occurrence distribution p(v), E = − m=1 p(vm ) log p(vm ), where clusters that have object instances for many different videos v are ranked higher and clusters that only contain objects from a single video are ranked lower. The idea is to reward clusters that are likely to contain hand-object interactions of multiple users and objects. Additionally, the instances within a cluster are sorted by their distance to the cluster center. The top 9 instances for the top 10 clusters are scored by human annotation. When each of the objects being grasped are all unique, the purity and diversity score will be equivalent, but when there are near duplicates (same object, same video), only the unique instances are counted to compute the diversity score. Table 2 shows that the masked HOG features obtains the highest purity and diversity scores. The hand contour feature ranks second showing that the shape of the hand is a strong feature for representing objects in the context of manipulation, which is an expected result. Although the global HOG feature captures both the shape of the hand and commonalities in the shape of interactive objects, we

(a) tripod

(a) Different interaction with similar shapes clustered together

(b) Failed hand detection noise creates meaningless cluster

(b) extension type and tip pinch

(c) Repeated skin-like objects learned as cluster

(c) lateral pinch and lateral tripod (d) Difficulty capturing fingers of two handed interaction

(d) thumb-4 finger, extension type, lateral tripod

Figure 4. Clustering results of grasps observed multiple times in the UTG dataset.

(a) lateral pinch and lateral tripod

(b) thumb-2 finger and thumb-3 finger

(c) lateral pinch and thumb-3-finger

Figure 5. Clustering results of grasps observed multiple times in the YHG dataset.

also found that the global HOG was also greatly affected by background clutter and prevented groups from diversifying across object types. Although the coverage values of the hand contour feature and global HOG feature are high, the low purity values mean that the clusters contain a mix of interactions. The masked HOG feature strikes the best balance between purity and coverage. Fig. 3 shows examples of regions clustered together using different features. Fig. 3 (c) shows how the shape of the plate or pizza causes dissimilar hand-object interactions to be clustered together. Both the local motion-based spatiotemporal BOWs and the global HOF are the least discriminative as appearance is not explicitly modeled and constant ego-motion prevents the features from modeling subtle motions signatures. Our results confirm that the inclusion of hands as a feature is critical in retaining regions of handobject interaction.

3. Additional Clustering Results Figure 4 and Figure 5 shows additional clustering results of UTG and YHG datasets. Some failure cases of

Figure 6. Examples of failure cases caused by errors in hand detection and similar hand shapes across multiple interaction.

clustering result on GTEA+ dataset is shown in Figure 6. Figure 7 shows the learned taxonomy tree on the GTEA+ dataset by our DPP-based hierarchical clustering. While the results are affected by meaningless cluster centers caused by failed hand detection, our algorithm is still able to capture Napier’s [6] widely adopted 1956 categorizations of grasps into precision and power grasps. For example, the red box, consisting of palmar, index finger extension, lateral pinch and stick, corresponds to the power grasp; the blue box, composed of precision thumb-index finger, precision thumb-2 finger, precision thumb-3 finger, and precision thumb-4 finger, represents the precision grasp.

4. Baseline Pseudocodes The pseudocode for the k-means based hierarchical clustering (Section 4.4 in the paper) is shown in Algorithm 1, and pseudocode for nearest representive point clustering (NRP clustering in Section 4.2 of the paper) is shown in Algorithm 2.

References [1] B. Alexe, T. Deselaers, and V. Ferrari. What is an object? In CVPR, 2010. 1, 3 [2] L. Cheng and K. M. Kitani. Pixel-level hand detection in egocentric videos. In CVPR, 2013. 1 [3] P. Doll´ar, V. Rabaud, G. Cottrell, and S. Belongie. Behavior recognition via sparse spatio-temporal features. In VSPETS, 2005. 2 [4] A. Fathi, Y. Li, and J. M. Rehg. Learning to recognize daily actions using gaze. In ECCV, 2012. 2 [5] H. Kang, M. Hebert, and T. Kanade. Discovering object instances from scenes of daily living. In ICCV, 2011. 2 [6] J. R. Napier. The prehensile movements of the human hand. Journal of bone and Joint surgery, 38(4):902–913, 1956. 4 [7] J. Shi and C. Tomasi. Good features to track. In CVPR, 1994. 2

Algorithm 1 GetHierarchyCentersKmeans(X)

Power Grasp

L ← 1, Y1 ← {x1 } for i = 2 : τ : |X| do `←1 while ` ≤ L do y`∗ ← arg miny∈Y` d(y, xi ) if d(y`∗ , xi ) > θ` then Y` ← Y` ∪ x i `←`+1 else n∗` += 1 y`∗ += (xi − y`∗ )/n∗` break end if end while if ` == L then L ← L + 1, YL ← {xi } end if end for return {Y` }

Precision Grasp

Algorithm 2 NRPClustering(X) Y ← {x1 } for i = 2 : τ : |X| do y ∗ ← arg miny∈Y d(y, xi ) if d(y ∗ , xi ) > θ then Y ← Y ∪ xi end if end for return Y

Figure 7. Automatically learned taxonomy tree on the GTEA+ dataset by our DPP-based hierarchical clustering.

Supplementary Material

gaze fixation point is detected by an eye tracker. ∗indicates equal contribution. (a) Random ... 1) confirms quanti- tatively our intuition, that the location of the hands are in- deed important for learning about hand-object interactions. Based on this result, we use hand locations to harvest hand- object regions in the paper. 1.2.

4MB Sizes 0 Downloads 326 Views

Recommend Documents

Supplementary Material
Jan Heufer ∗. *TU Dortmund University, Department of Economics and Social Science, 44221 Dortmund,. Germany. .... 3 More Details on the Data Analysis.

Supplementary Material - Arkivoc
General Papers. ARKIVOC 2015 (vii) S1-S13. Page S3. ©ARKAT-USA, Inc. δ /ppm. 1. H Assignment. 8.60 (brs, 1H). 5.35 (t, J = 7.3 Hz, 1H). 5.08 (t, J = 6.1 Hz, ...

Supplementary Material
By the definition of conjunctive pattern, since S does not satisfy pc, then S ... Then there exists at least one instance of p o in S: {I1,I2,...,Im},. I1 ∈ I(p c. 1 ,S),...,Im ∈ I(p c m,S), such that ∀t1 ∈. I1,..., ∀tm ∈ Im,t1 < t2 < ...

Supplementary Material
and Business Cycles Conference, the Bank of Canada 2014 Fellowship ..... CGH exhibited an upward trend in the frequency of sales that could explain why they find ...... 12Fox and Sayed (2016) use the IRI scanner dataset to document the ...

Supplementary Material
The data are provided a for 10 subsectors, which I aggregate up to three sectors as follows. ... CAN Canada ...... Structural change in an open economy. Journal ...

Supplementary Material
SVK Slovakia. CYP Cyprus. ITA. Italy. SVN Slovenia. CZE Czech Republic JPN Japan. SWE Sweden. DEU Germany. KOR South Korea. TUR Turkey. DNK Denmark. LTU Lithuania. TWN Taiwan. ESP Spain. LUX Luxembourg. USA United States. EST Estonia. LVA Latvia. RoW

Efficient Repeated Implementation: Supplementary Material
strategy bi except that at (ht,θt) it reports z + 1. Note from the definition of mechanism g∗ and the transition rules of R∗ that such a deviation at (ht,θt) does not ...

Supplementary Material for
Aug 3, 2016 - alternatives are “autocracy and risk-sharing” and “democracy and investment.” In other words, the .... whether seizing power. A coup is .... Our main sources are Galasso (1976), Stearns (2001), and Ortu (2005). References.

Supplementary Online Material
Branstetter, M.G. (2009) The ant genus Stenamma Westwood (Hymenoptera: Formicidae) redefined, with a description of a new genus Propodilobus. Zootaxa,.

Supplementary Material for
Fujitsu Research & Development Center Co., Ltd, Beijing, China. {wangzhengxiang,rjliu}@cn.fujitsu.com. Abstract. This supplementary material provides the ...

Electronic supplementary material
Jun 22, 2009 - ... of two Hill functions (a-b) with sufficiently distant midpoints is equivalent to the Hill function with the largest midpoint (c). Namely: Hill(x, K 1) · Hill(x, K 2) ≈ Hill(x, K 2) where K 1

Supplementary Material - HEC Montréal
... the ONS website: http://www.ons.gov.uk/ons/datasets-and-tables/index.html. .... sitivity of sales to the unemployment rate for independent stores and chains.

Supplementary Material for
Sep 25, 2015 - Using archived and field-collected worker bumble bees, we tested for temporal changes in the tongue ... (Niwot Ridge Long Term Ecological Research Site, 40°3.567'N, 105°37.000'W), and by. P. A. Byron (17) ... these transects did not

Supplementary Material for ``Observations on ...
Nov 19, 2017 - C.2 Steady State in a Perturbed Environment. In this subsection we formally adapt the definitions of a consistent signal profile and of a steady state to perturbed environments. Let f((1−ϵ)·σ+ϵ·λ) : OS → OS be the mapping bet

Supplementary Material Methods Tables
αF was added (25nM final) to exponentially growing MATa bar1Δ cells in YPD (OD600 ≈ 0.3) and the culture incubated for two hours at. 30oC. >95% G1 arrest, as indicated by cells with small buds (schmoo formation), was confirmed by microscopy. Cell

Supplementary Material: Adaptive Consensus ADMM ...
inequalities together. We conclude. (Bvk+1 − Bvk)T (λk+1 − λk) ≥ 0. (S3). 1.2. Proof of ..... matrix Di ∈ Rni×d with ni samples and d features using a standard ...

Price Selection – Supplementary Material
†Email: Departamento de Economia, Pontifıcia Universidade Católica do Rio de .... and-tables/index.html, the Statistics Canada's Consumer Price Research ...

Electronic Supplementary Material- Optimizing ...
3Data Science Laboratory, Institute for Scientific Interchange (ISI), Turin, Italy. 4Istituto ..... The entropy function introduced in Equation 1 of the main text evaluates the level of frag- ..... In the case of a deterministic spreading, we recover

Price Selection – Supplementary Material
trade and search frictions in the goods market. An equilibrium pins down a unique relative price distribution G(p−1 | S−1) but does not pin down price changes. This distribution is invariant to monetary shocks. Hence there is full monetary neutra

SUPPLEMENTARY MATERIAL FOR “WEAK MONOTONICITY ...
This representation is convenient for domains with complete orders. 1 .... check dominant-strategy implementability of many classical social choice rules. In.

SUPPLEMENTARY MATERIAL FOR “WEAK MONOTONICITY ...
This representation is convenient for domains with complete orders. 1 ... v = (0,v2,0), v2 > 0, would want to deviate and misreport their type so as to get 3.

1 Supplementary online material: additional results ...
weakly higher winning frequencies for A and C in comparison to D. The corresponding tests reject the .... tives over the course of the experiment for PV elections.

Supplementary material for “Complementary inputs and ...
Aug 2, 2017 - statement of Theorem 2, and then complete the proof. ...... Jones, 1984), spaces of square-integrable functions (Harrison and Kreps, 1979; ...

Supplementary material to Mathematical models of ...
Mathematical models of the VEGF receptor and its role in cancer therapy ... restrict ourselves to the system of equations for the first moments (q1(t)) and for the ...