Video Stream Retrieval of Unseen Queries using Semantic Memory Spencer Cappallo [email protected]

Thomas Mensink [email protected]

Institute of Informatics University of Amsterdam Science Park 904 Amsterdam, The Netherlands

Cees G. M. Snoek [email protected]

Abstract

where s(q) returns a vector containing the cosine similarities between the embedding representation of the query q and those of the concepts C. If Retrieval of live, user-broadcast video streams is an under-addressed and the query q comprises multiple terms, we use the mean of the per term increasingly relevant challenge. The on-line nature of the problem ne- scores, which has been shown to hold semantic relevance [7]. cessitates temporal evaluation and the unforeseeable scope of potential queries motivates an approach which can accommodate arbitrary search 2.2 Memory for Stream Retrieval queries. To account for the breadth of possible queries, we adopt a noexample approach to query retrieval, which employs a query’s semantic Retrieved streams must still be relevant at the time of the query. For this relatedness to pre-trained deepnet classifiers. To adapt to shifting video reason, we explore three ways to incorporate a "memory", which prioricontent, we propose Memory Pooling and Memory Welling as methods tizes recent information, in a zero-example stream retrieval setting. which favor recent information over past, possibly unreliable content. Memory Pooling Two stream retrieval tasks are identified: Instantaneous Retrieval at any Temporal pooling of frame-based features or concepts over an entire video arbitrary time, and Continuous Retrieval of a single query over a prois used in state-of-the-art approaches for standard video retrieval tasks [9]. longed duration. Evaluation metrics suited for these two tasks are also This strategy could be adapted to an on-line setting by pooling among all developed. Three large scale video datasets are adapted to the challenge frames from time t = 0 to the present. However, this introduces problems of stream retrieval. We report results for the proposed approaches on the when the content of a stream changes, which is a particular concern with new stream retrieval tasks, as well as demonstrate their efficacy in a tradilonger streams. For this reason, we pool instead over a fixed temporal tional, non-streaming video task. This paper appeared at BMVC’16 [2] memory m, which is tethered to the present and offers a restricted view on the past:

1

Introduction

This work targets the challenge of searching among live streaming videos. This is a problem of growing importance as more video content is streamed via services like Meerkat, Periscope, and Twitch. Despite the popularity of live streaming video, searching in its content with state-of-the-art video search methods [6, 9] is nearly impossible as these typically assume the whole video is available for analysis before retrieval. We propose a new method that can search across live video streams, for any query, without analyzing the entire video. In live video, the future is unknowable thus one only has access to the past and present. It is therefore crucial to leverage knowledge of the (recent) past appropriately. Memory can be modeled with the aid of hidden Markov models or recurrent neural networks with long-short term memory. Through the ability to selectively remember and forget, recurrent neural networks have recently shown great potential for search in videos. Inspired by the success of supervised memory models, we propose a mechanism to incorporate memory and forgetting in video stream retrieval without learning from examples.

2

Video Stream Retrieval

The nature of live, user-broadcast video has two major implications. First, the full range of potential future queries cannot be known, necessitating the ability to respond to unanticipated queries. Second, the future content of live video is unknown, and might not relate to prior content within the same stream, therefore we propose several methods to emphasize recent stream content.

2.1

Ranking Unanticipated Queries

t

MPmax (xt ) = max xi i=t−m

MPmean (xt ) =

1 t ∑ xi m i=t−m

(2)

where xt denotes the features at time t, and we evaluate max pooling or mean pooling, denoted as MPmax and MPmean respectively. The contribution of low confidence concepts introduces noisy predictions and influences the retrieval performance, therefore we use only the highest-valued pooled concepts, as proposed in [8]. Memory Welling The need to capture both long term trends and short-duration confidence spikes motivates the development of what we term memory wells. Observations flow into the wells at every timestep, but the wells also leak at every timestep. In contrast to the memory pooling, where all observations are weighed equally and observations beyond the memory horizon are lost, the impact of past observations diminishes steadily over time for memory wells. A well is defined in the following manner:   1 m−1 w(xt−1 ) + xt − β , 0 , (3) w(xt ) = max m m where the current value of the well is based on the value at time t − 1, diminished by a tunable memory parameter m, and a fixed constant leaking term β . The β term creates sparseness in the representation, which ensures that only recent or consistently present concepts are used for prediction. We fix β = C1 , where C is the number of concepts, as this is the value given when all classes are considered equally likely. Enforcing sparseness, or rather, enforcing reliability of concept scores, means that the concept well values can be used directly in Equation 1, without the need for costly selection of the highest-confidence concepts. Max Memory Welling In the case of short streams and traditional video processing tasks, which are likely to have more consistent content, the short-term nature of memory welling can be a limitation, even if its properties are still effective for improving temporally local predictions. Memory welling is adapted to this task through execution of a max pooling of the query scores per stream:

The goal is to retrieve relevant streams for a textual query q. To be robust against unanticipated queries, we follow a zero-shot classification paradigm [1, 4, 8]. A deep neural network trained to predict image classes is applied to the frames of the video stream as a feature extractor. xt represents the softmax output of the deep network across the output classes C for a frame at time t. Some φ (xt ) encodes these concepts in a sparse t score(xt ) = max (s(q) · w(xi )| ) (4) manner. Both the concepts C as well as the query q are placed in a mutual i=0 embedding space (in our case, word2vec [7]), and video steams are scored This exploits local, high confidence predictions from the welling approach, based on the cosine similarities, using: without discarding past information. It is well-suited to single-topic conscore(xt ) = s(q) · φ (xt )| (1) tent such as short streams and traditional, full video retrieval tasks.

Table 1: Results of instantaneous and continuous retrieval tasks. m = 1 uses only the current frame, while m = t indicates pooling performed over We identify two evaluation settings for video stream retrieval: i) Instan- all past and present frames. Instantaneous Retrieval (% TAP) taneous Retrieval, which measures the retrieval performance at any given AN FCVS AN-L FCVS-L time t; and ii) Continuous Retrieval, where a succession of streams relevant to a single query are retrieved over a prolonged duration. Random 1.4 4.9 3.6 2.9

3

Tasks for Video Stream Retrieval

Instantaneous Retrieval The goal of instantaneous retrieval is to retrieve the most relevant stream for a query q at any arbitrary time t. This temporal assessment is important, given that a model which only performs well when a stream has ended is useless for discovery of live video streams. To incorporate the temporal domain, we use the mean of the average precision (AP) scores per time step t, which we coin Temporal Average Precision (TAP). Letting APt denote the AP score for some query at time t, the TAP then corresponds to the mean APt across all times for which there is at least one relevant stream. Continuous Retrieval The goal of the continuous retrieval task is to maximize the fraction of time spent watching relevant streams, while minimizing the number of times the stream is changed. Consider a viewer searching for coverage of the Olympics. When one stream stops showing the Olympics, she wants to switch to another stream showing the Olympics. However, switching between two streams every second, even if both relevant, provides a poor viewing experience. To evaluate this scenario, we consider the number of zaps. A zap is any change in the retrieved stream or its relevancy, including the move at time t = 0 to the first retrieved stream. We distinguish good zaps, which is any zap that moves from a currently irrelevant stream to a currently relevant stream, from all other (bad) zaps. The count of good zaps and bad zaps are represented by z+ and z− . To incorporate overall accuracy over time, we also reward an algorithm choosing to correctly remain on a relevant stream. Letting r+ track the number of times an algorithm remains on relevant stream, the zap precision ZP is ZP =

z+ + r+ ∑t yt

(5)

where yt again represents whether or not there is at least one relevant stream at time t.

4

Experiments

Datasets and Setup We evaluate our methods on three large scale video datasets: i) ActivityNet [3] (AN), a large action recognition dataset with 100 classes and 7200 labeled videos. Performance is evaluated on a test set composed of 60 classes randomly selected from the combined ActivityNet training and validation splits, and a validation set of the other 40 classes is used for parameter search; ii) A subset of the Fudan-Columbia Videos [5] (coined FCVS), composed of 25 videos for each of the 239 classes making up 250 hours of video, which we split into a validation set of 50 classes and a test set of 179 classes. FCVS annotations are more diverse (objects, locations, scenes, and actions), but lack temporal extent, so a class is assumed to be relevant for the duration of a video; iii) TRECVID MED 2013 [9] (MED), an event recognition dataset, used to evaluate the efficacy of our memory-based approach against published results. To facilitate comparison, the setting used by [4] is replicated: whole-video retrieval using only the event name. In addition to evaluating on short web videos themselves, we introduce AN-L and FCVS-L, which are adaptations to simulate longer streams with varied content. To accomplish this, individual videos are randomly concatenated until the simulated stream is at least 30 minutes long. Annotations from the original videos are propagated to these concatenated videos. We sample videos at a rate of two frames per second. Each frame is represented by 13k ImageNet concept confidence scores, using a pretrained deep neural network from [6]. Our semantic embedding is a 500dimensional skip-gram word2vec [7] model trained on the text accompanying 100M Flickr images [1, 10]. To simulate streams, we process all videos sequentially, as though they were concurrent live video streams. We report the TAP and ZP measures averaged over all test classes. For our memory based methods, the

Mean Memory Pooling m=1 m=t m = m∗ Max Memory Pooling m=t m = m∗ Memory Welling Max Memory Welling

16.9 18.4 21.7

21.4 30.7 28.8

25.1 8.5 29.3

24.8 9.3 30.0

20.0 21.0 22.5 24.6

27.4 27.5 30.5 35.9

9.0 29.7 30.1 11.0

9.5 30.3 30.6 15.9

AN-L

FCVS-L

Continuous Retrieval (% ZP) Random

1.3

1.1

Mean Memory Pooling m=1 m=t m = m∗

21.9 5.9 27.5

21.6 6.3 27.7

Max Memory Pooling m=t m = m∗

5.9 27.3

6.0 27.5

Memory Welling Max Memory Welling

28.3 5.6

28.4 10.9

optimal value of m = m∗ is determined on the validation set. Two extremes of memory pooling are used as baselines: m = 1, which simply relies on the current frame of a video to make a prediction; and m = t, which corresponds to pooling over the entirety of the stream up to the present time. Results Table 1 shows the results for the instantanous and continuous stream retrieval tasks. We observe that memory-based approaches shine when query relevance is temporally limited, as in the AN, AN-L, and FCVSL datasets. For a setting like FCVS, where each annotation covers an entire stream, the baselines become more competitive. In a scenario where streams are guaranteed to be short in duration and focus on a single topic, then a max memory welling approach makes the most sense. For streams of indeterminate length and content, the memory welling approach offers the best results and flexibility to cover any situations that may arise. To compare memory-based methods against published results, we report mAP results on the MED dataset, following the setting from [4]: multimedia event retrieval based solely on the event-name. The Max Memory Welling approach outperforms [4] (4.7% mAP to 4.2% mAP), despite [4] using a more advanced Fisher Vector event-name encoding. Note, such an event-name encoding could also be used alongside our method, but the focus of our work is stream retrieval. Max Memory Welling leverages the short-term, high-confidence predictions generated through memory welling, signals which may be averaged away in whole-video pooling. [1] S. Cappallo, T. Mensink, and C. Snoek. Query-by-emoji video search. In MM, 2015. [2] S. Cappallo, T. Mensink, and C. Snoek. Video stream retrieval of unseen queries using semantic memory. In BMVC, 2016. [3] F. Heilbron, V. Escorcia, B. Ghanem, and J. Niebles. Activitynet: A large-scale video benchmark for human activity understanding. In CVPR, 2015. [4] M. Jain, J. van Gemert, T. Mensink, and C. Snoek. Objects2action: Classifying and localizing actions without any video example. In ICCV, 2015. [5] Y. G. Jiang, Z. Wu, J. Wang, X. Xue, and S. F. Chang. Exploiting feature and class relationships in video categorization with regularized deep neural networks. arXiv preprint arXiv:1502.07209, 2015. [6] P. Mettes, D. Koelma, and C. Snoek. The imagenet shuffle: Reorganized pre-training for video event detection. In ICMR, 2016. [7] T. Mikolov, I. Sutskever, K. Chen, G. Corrado, and J. Dean. Distributed representations of words and phrases and their compositionality. In NIPS, 2013. [8] M. Norouzi, T. Mikolov, S. Bengio, Y. Singer, J. Shlens, A. Frome, G. Corrado, and J. Dean. Zero-shot learning by convex combination of semantic embeddings. ICLR, 2014. [9] P. Over, G. Awad, M. Michel, J. Fiscus, W. Kraaij, A. Smeaton, G. Quénot, and R. Ordelman. Trecvid 2015 – an overview of the goals, tasks, data, evaluation mechanisms and metrics. In TRECVID, 2015. [10] B. Thomee, D. Shamma, G. Friedland, B. Elizalde, K. Ni, D. Poland, D. Borth, and L. Li. YFCC100M: The new data in multimedia research. CACM, 59(2), 2016.

Video Stream Retrieval of Unseen Queries using ...

Retrieval of live, user-broadcast video streams is an under-addressed and increasingly relevant challenge. The on-line nature of the problem ne- cessitates temporal evaluation and the unforeseeable scope of potential queries motivates an approach which can accommodate arbitrary search queries. To account for the ...

117KB Sizes 0 Downloads 190 Views

Recommend Documents

Video Retrieval Based on Textual Queries
Center for Visual Information Technology,. International Institute of Information Technology,. Gachibowli ... There are two important issues in Content-Based Video Ac- cess: (a) A .... matching algorithm then computes the degree of similarity be-.

Video Retrieval Based on Textual Queries
Audio and the textual content in videos can be of immense use ... An advanced video retrieval solution could identify the text ..... Conference on Computer.

Fast Forensic Video Event Retrieval Using Geospatial ...
It also makes the online video search possible by filtering tremendous amount of data ... H.3.3 [Information Storage and Retrieval]: Informa- tion Search and ...

Compressed Domain Video Retrieval using Object and ...
The object features such as speed, area and trajectory are then obtained after ... With the advent of broadband networks, high-powered workstations and ... (3) With the availability of hardware encoders and decoders for MPEG, such an.

Segmented Trajectory based Indexing and Retrieval of Video Data.
Indexing and Retrieval of Video. Data. Multimedia Systems Lab, UIC. 1. Faisal I. Bashir, Ashfaq A. Khokhar, Dan Schonfeld. University of Illinois at Chicago,.

Fast C1 Proximity Queries using Support Mapping of ...
STP-BV construction steps, from left to right: point clouds (vertices of an object), building ..... [Online]. Available: http://www.math.brown.edu/∼dan/cgm/index.html.

Approximate Rewriting of Queries Using Views
gorithm called Build-MaxCR, for constructing a UCQAC size-limited MCR ... information integration [4], data warehousing [10], web-site design [23], and query.

Rewriting queries using views in the presence of ...
bDepartment of Computer Science, University of California, Irvine, CA 92697-3435, USA ... +302102232097; fax: +302107722499. ...... [13] S. Chaudhuri, M.Y. Vardi, On the equivalence of recursive and nonrecursive datalog programs, ...

an effective video retrieval system by combining visual ...
May 2, 2009 - COMBINING VISUAL AND TEXTUAL MINING .... Totally, the duration of video data is about 20 hours ... Recall = 5 / 20 * 100% = 25%. %100*.

CONTENT-FREE IMAGE RETRIEVAL USING ...
signed value of one is denoted by XE = {Xj1 , ..., XjE } and is called the ... XE is in practice much smaller than. XH. .... Fourteen trademark examiners in the Intel-.

Retrieval of 3D Articulated Objects using a graph-based representation
called protrusion function, pf() [APP. ∗ ... the minimum cost paths between the representative salient .... The optimal cost of the optimization process is the EMD.

A Motion Trajectory Based Video Retrieval System ...
learning and classification tool. In this paper, we propose a novel motion trajectory based video retrieval system. For feature space representation, we use two ...

Enhancing Image and Video Retrieval: Learning via ...
School of Computer Science and Engineering and the Center for Neural Computation. The Hebrew ... on the clustering and retrieval of surveillance data. Our.

Automated Detection of Engagement using Video-Based Estimation of ...
Abstract—We explored how computer vision techniques can be used to detect ... supervised learning for detection of concurrent and retrospective self-reported engagement. ...... [49] P. Ekman and W. V. Friesen, Facial Action Coding System: A ... [On

Automated Detection of Engagement using Video-Based Estimation of ...
Abstract—We explored how computer vision techniques can be used to detect engagement while ... supervised learning for detection of concurrent and retrospective self-reported engagement. ...... [Online]. Available: http://msdn.microsoft.com/en-us/l

In-Stream Video Advertising with DoubleClick for Advertisers
In-stream advertisers can overcome the challenges of video advertising today by managing in-stream campaign implementation and measurement using the ...

Wireless video transmission Evaluation using Lossless video ... - IJRIT
support streaming from live video sources, such as cable/satellite ..... Solomon code is widely used from optical media storages, such as DVD and CD, .... S. Priyadarshi, and J. Villasenor "H. 264 Parameter Optimizations for Internet Based.

Wireless video transmission Evaluation using Lossless video ... - IJRIT
open or closed-loop translating or even ... several parallel data stream . ..... the condition of transmission without degradation and error free sequence is possible ...

Rewriting queries using views with negation - IOS Press
AI Communications 19 (2006) 229–237. 229. IOS Press. Rewriting queries using views with negation. Foto Afrati and Vassia Pavlaki. Department of Electrical ...

Bayesian Active Learning Using Arbitrary Binary Valued Queries
Liu Yang would like to extend her sincere gratitude to Avrim Blum and Venkatesan. Guruswami for several enlightening and highly stimulating discussions. Bibliography. Bshouty, N. H., Li, Y., & Long, P. M. (2009). Using the doubling dimension to analy