Webly-supervised Video Recognition by Mutually Voting for Relevant Web Images and Web Video Frames Chuang Gan

IIIS, Tsinghua University

[email protected]

Chen Sun

Google Research

[email protected]

Lixin Duan

Amazon

[email protected]

Boqing Gong

CRCV, University of Central Florida

[email protected]

...

... (a) Basketbal Dunk

...

... (b) Bench Press

...

... (c) Pizza Tossing

Figure 1: To utilize Web images and videos for video classification, our key observation is that the query-relevant images and frames typically appear in both domains with similar appearances, while the irrelevant images and videos have their own distinctiveness. Here we show Web images (top) and video frames (bottom) retrieved by keywords basketball dunk, bench press and pizza tossing from search engines. The relevant ones are marked in red.

the irrelevant images and videos have their own distinctiveness. In Figure 1, we show the Web images (top) and video frames (bottom) retrieved by keywords basketball dunk, bench press and pizza tossing. We can see that for the basketball dunk example, non-slam-dunk frames in the video are mostly about a basketball game. The irrelevant Web images are more likely to be cartoons. Similar observation also holds for bench press and pizza tossing, where the irrelevant images include cartoons and product shots. This observation indicates that selecting training examples from Web images and videos can be made easier, if they could be mutually filtered to keep those in common! Our algorithm to mutually filtering Web images and video frames goes as follows: we first jointly choose images and video frames and try to match them aggressively. A good match between the subset of images and the subset of video frames occurs when both subsets are relevant to the action name, since “each irrelevant image or frame is irrelevant in its own way”. We then impose a passive constraint over the video frames to be selected, such that they are collectively not too far from the original videos. We would like to be passive on the videos, in contrast to the images, because our ultimate goal is for video action recognition. Otherwise, the aggressive matching mechanism may end up with too few frames and causes a domain adaptation problem between the training set and test videos. Once the Web images and video frames are selected for the actions or events of interest, they can be readily used to train action or event classifiers with a wide range of tools. Some examples include SVM, CNN and LSTM.

2 1

Proposed Approach

Introduction

This work aims to classify actions and events in user-captured videos without human labeling. Video recognition in the wild is a very challenging task: videos from the same categories could vary greatly in lighting conditions, video resolutions, camera movements, etc. Meanwhile, those from different categories could be inherently similar (e.g. “apply eye makeup” and “apply lipstick”). State-of-the-arts approaches require and implicitly assume the existence of large-scale labeled training data. Manually labeling large amount of video examples is time-consuming and difficult to scale up. On the other hand, there are abundant image and video examples on the Web that can be easily retrieved by querying action or event names from image/video search engines. These two observations motivate us to focus on Webly-supervised video recognition by exploiting Web images and Web videos. Using video frames in addition to images not only adds more diverse examples for training better appearance models, but also allows us to train better temporal models. However, there are two key difficulties that prevent us from using Web data directly. First, the images and videos retrieved from Web search engines are typically noisy. They may contain irrelevant results, or relevant results from a completely different domain than users’ interest (e.g. cartoons or closeup shots of objects). To make the problem worse, Web videos are usually untrimmed and could be several minutes to hours long. Even for a correctly tagged video, the majority of its frames could be irrelevant to the actual action or event. Our goal then becomes to identify query-relevant images and video frames from the Web data which are both noisily and weakly labeled, in order to train good machine learning models for action and event classification. Our proposed method is based on the following observation: the relevant images and video frames typically exhibit similar appearances, while

In this section, we present the details of our approach to jointly selecting video frames and images from the Web data, for the purpose of Weblysupervised video recognition. Our algorithm is built upon the motivating observation that “all relevant images and frames to an action name are alike; each irrelevant image or frame is irrelevant in its own way.” We firstly give the overall formulation, and then describe an alternative optimization procedure for solving the problem.

2.1

Joint selection of action/event relevant Web video frames and Web images

For the ease of presentation, we first define the following notations. For each class (of an action or event), we denote by I = {xm }M m=1 the set of Web images, and by V = {vn }N n=1 the set of video frames, both returned by some search engines in response to the query of the class name. The Web data are quite noisy; there are both relevant items and outliers for the class. In order to filter out the relevant items, we introduce M indicator variables α = [α1 , . . . , αM ]> , where αm ∈ {0, 1} for each image xm , and N indicator variables β = [β1 , . . . , βN ]> , where βn ∈ {0, 1} for each video frame vn . If αm = 1 (resp., βn = 1), the corresponding image xm (resp., video frame vn ) will be identified as a relevant item to the class. 2.1.1

Aggressive matching.

If we conduct a pairwise comparison between a subset of the images I with a subset of the video frames V, any class-irrelevant images or frames would decrease the similarity between the two subsets, because the irrelevant items are likely different from each other and also different from the relevant items. Therefore, we can let the images and video

frames mutually vote for class-relevant items, by matching all possible Table 1: Webly-supervised action recognition results on UCF101, by finepairwise subsets of them, respectively. Such a pair can be expressed tuning VGGNET19 using both Web images and Web video frames. (x%: percentage abandoned) N by ({αm xm }M m=1 , {βn vn }n=1 ). The pairs with high matching scores have Method # Number of training data Acc (%) lower chance of containing irrelevant images or video frames. All crawled data 426K 64.7 Since simplicity and effectiveness of the maximum mean discrepanValidation 368K 66.5 cy (MMD) criterion [3], we adopt it in this work to measure the degree One-class SVM (5%) 405K 65.4 N One-class SVM (10%) 384k 65.9 of matching between any images and frames ({αm xm }M m=1 , {βn vn }n=1 ). One-class SVM (15%) 363k 65.9 We propose to minimize the square of MMD such that the true negative Unsupervised One-class SVM (5%) 405K 66.6 images and video frames are expected to be filtered out (i.e., the correUnsupervised One-class SVM (10%) 384k 66.9 sponding αm ’s or βn ’s will tend to be zeros). In other words, the remain363k 66.4 Unsupervised One-class SVM (15%) ing images and video frames are expected to be the true positive items for Landmarks (5%) 405K 67.9 Landmarks (10%) 384k 68.3 the class. Formally, we formulate the following optimization problem: min

αm ,βn ∈{0,1}

M 1 1

α φ (xm ) − N

M ∑

∑m=1 αm m=1 m ∑n=1 βn

2

∑ βn φ (vn )

, n=1 N

(1)

H

where φ (·) is a mapping function which maps a feature vector from its original space into a Reproducing Kernel Hilbert Space H. The above is an integer programming problem, which is very computationally expensive to solve. Following [2], we relax Eq. (1) by introducing αˆ m = Mαm and βˆn = Nβn . Then, we arrive at the following ∑m=1 αm ∑n=1 βn optimization problem:

min

αˆ ∈[0,1]M ,βˆ ∈[0,1]N

  > αˆ > , βˆ

KI −KV>I

−KIV KV



αˆ βˆ

 ,

(2)

where αˆ = [α1 , . . . , αM ]> , βˆ = [β1 , . . . , βN ]> , KI ∈ RM×M and KV ∈ RN×N are the kernel matrices computed over the images and video frames respectively, and KV>I = KIV ∈ RM×N denotes the kernel matrix computed between the images and video frames, respectively. We use a Gaussian RBF kernel in our experiments. 2.1.2

Passive video frame selection.

Landmarks (15%) Ours (5%) Ours (10%) Ours (15%)

363k 405K 384k 363k

67.7 68.7 69.3 68.9

• Validation: For each action class, we split the crawled data U into K equal and disjoint subsets. Each subset is scored by a binary SVM classifier trained on the rest K − 1 subsets as positive and some random images of the other classes as negative. Every data point in U is predicted once. Negative-scored data are considered as noise and rejected. We use the implementation of LibSVM [1] with default hyper parameter λ = 1 to conduct experiments. In our experiment, we set K as 5. • One-class SVM: We use LibSVM [1] to conduct the experiment. • Unsupervised One-class SVM: We implemented this method [4] ourselves and followed the suggested details for tuning the hyperparameters (e.g. using Gaussian kernels, soft labels and the number of neighbors). • Landmarks: The concept of landmarks [2] is originally defined as a subset of data point from source domain that match the target domain. In our problem, we first treat Web images as the source domain (and Web video frames as the target domain) to select “landmark” Web images. Then we reverse the source and target domains to select video frames. We use the code provided by authors for the experiments. Table 1 reports the Webly-supervised action recognition results when our and the baseline approaches are used to select both Web images and video frames for fine-tuning CNNs.

Note that Eq. (1) matches a subset of images with a subset of video frames very aggressively. While there could be many pairs of subsets whose images and frames are all relevant to the class, Eq. (1) only choose the one with the best matching (in terms of the MMD measure). This strategy is effective in removing true negative images and frames. However, it may also abandon many relevant ones in order to reach the best matching. We thus introduce a passive term to balance the aggressive matching. 4 Conclusions Since our eventual task is video recognition, we propose to impose a passive regularization over the selected video frames, such that they are In this paper, we investigated to what extent the Web images and Wecollectively not too far from the original videos: b videos could be leveraged to conduct Webly-supervised video recognition. To distill useful data from the noisy Web ones, we proposed a

2

min (3) unified approach to jointly removing irrelevant Web images and (also re V −V · diag(βˆ ) ·W , F βˆ ∈[0,1]M ,W dundant) video frames. We developed an efficient alternative optimization procedure to solve our proposed formulation. Extensive experiments, for where V = [v1 , . . . , vN ], and the variable W is a linear transformation ma- both action recognition and event detection, validate that our framework trix which linearly reconstructs V from all the selected video frames, i.e., not only outperforms competing baselines, but also beats existing systems V · diag(βˆ ). In order to have a low reconstruction error, one cannot keep which also exploit Web data for event detection. We expect this work to too few video frames selected by the variables β . On the other hand, it benefit future research on large-scale video recognition tasks. is fine to remove redundant frames from the candidate set V. Our experiments show that removing the redundant frames incurs little loss on the overall performance, and even improves the performance of an LSTM- References based classifier. [1] Chih-Chung Chang and Chih-Jen Lin. LIBSVM: A library for supCombining Eq. (2) and Eq. (3), we present our overall optimization port vector machines. ACM Transactions on Intelligent Systems and problem as follows: Technology, 2011.     K [2] Boqing Gong, Kristen Grauman, and Fei Sha. Connecting the dots > αˆ −KIV I min αˆ > , βˆ + λ kV −V · diag(βˆ ) ·W k2F , with landmarks: Discriminatively learning domain-invariant features > ˆ M −KIV KV β αˆ ∈[0,1] , βˆ ∈[0,1]N ,W for unsupervised domain adaptation. In ICML, pages 222–230, 2013. (4) [3] Jiayuan Huang, Arthur Gretton, Karsten M Borgwardt, Bernhard Schölkopf, and Alex J Smola. Correcting sample selection bias by where λ > 0 is a pre-defined tradeoff parameter to balance these two unlabeled data. In NIPS, pages 601–608, 2006. terms. [4] Wei Liu, Gang Hua, and John R Smith. Unsupervised one-class learning for automatic outlier removal. In Proceedings of the IEEE Confer3 Experiment Results on UCF101 ence on Computer Vision and Pattern Recognition, pages 3826–3833, 2014. To evaluate our framework, we compare against several state of the art noise removal approaches as baselines:

Webly-supervised Video Recognition by Mutually ...

appear in both domains with similar appearances, while the irrelevant im- ages and ... vant results from a completely different domain than users' interest (e.g..

636KB Sizes 0 Downloads 179 Views

Recommend Documents

Human Action Recognition in Video by 'Meaningful ...
condensation technique alleviates the curse of dimensional- ity by mining the multi-dimensional pose descriptors into a kd-tree data structure. The leaf nodes of the kd-tree ...... Morgan Kaufmann. Publishers Inc., 1981. [13] G. Mori and J. Malik. Es

Human Action Recognition in Video by 'Meaningful ...
Bag-of-word based action recognition tasks either seek right kind of features for ... The emphasis on the pose specific details is in accordance with the theme of this ... supervised fashion, i.e. this set of poses is constructed sep- arately for eac

Human Activity Recognition for Video Surveillance
of unusual event recognition with lack of training data. Zelnik-Manor et al. .... a GMM classifier Ci for each CFV Fi with MAP (Maximum a. Posteriori) principle, as ...

A statistical video content recognition method using invariant ... - Irisa
class detection in order to understand object behaviors. ... to the most relevant (nearest) class. ..... using Pv equal to 95% gave the best efficiency so that in ...... Activity representation and probabilistic recognition methods. Computer. Vision

recognition of 48 human behaviors from video
integrated into an application. Many reasonably ... of an item, this is the best clue that some item is being carried ... On a standard desktop pc, the processing time per clip ..... We have proposed a framework to recognize a set of 48 complex ...

License Plate Recognition From Still Images and Video Sequences-A ...
License Plate Recognition From Still Images and Video Sequences-A survey.pdf. License Plate Recognition From Still Images and Video Sequences-A survey.

Automated recognition of human activities in video ...
10.1117/2.1201405.005489. Automated recognition of human activities in video streams in real-time. Sebastiaan Van den Broek, Johan-Martijn ten Hove,.

A statistical video content recognition method using invariant ... - Irisa
scene structure, the camera set-up, the 3D object motions. This paper tackles two ..... As illustration, examples of a real trajectories are showed in. Fig. 4, their ..... A tutorial on support vector machines for pattern recognition. Data Mining and

License Plate Recognition From Still Images and Video Sequences ...
License Plate Recognition From Still Images and Video Sequences-A survey.pdf. License Plate Recognition From Still Images and Video Sequences-A survey.

VERSA – Video Event Recognition for Surveillance ...
Nov 3, 2008 - streamed by a lower-level video analytics service that provides basic ..... Similarly, supervised learning methods ...... flatten([R|Relsx],Rels).

DaMN – Discriminative and Mutually Nearest: Exploiting ... - Stanford CS
Chang, C.C., Lin, C.J.: LIBSVM: A library for support vector machines. ACM .... Yu, F., Cao, L., Feris, R., Smith, J., Chang, S.F.: Designing category-level at-.

DaMN – Discriminative and Mutually Nearest: Exploiting ... - CS Stanford
information from known categories to enable existing classifiers to generalize to novel categories for which there ... is how best to identify such categories and then how to combine information from the different classifiers in a ... trains from all

DaMN – Discriminative and Mutually Nearest: Exploiting ... - CS Stanford
was originally created by domain experts in the cognitive science [18] and AI [9] com- munities. Traditionally, semantic ... attributes by mining text and images on the web. Ferrari and Zisserman [7] learn ... 2) unlike the attributes described in Se

pdf-1432\video-game-optimization-text-only-by-epreiszbgarney-by ...
Connect more apps... Try one of the apps below to open or edit this item. pdf-1432\video-game-optimization-text-only-by-epreiszbgarney-by-eric-preisz.pdf.

Simulation of Mutually Coupled Oscillators Using ...
Modern RF chips for mobile devices, for instance, typically .... Using this simple and numerically cheap method, one can do many kinds of analysis ... Floor plan with relocation option that was considered after nonlinear phase noise analysis ...

DaMN – Discriminative and Mutually Nearest: Exploiting ... - CS Stanford
6. Rui Hou, Amir Roshan Zamir, Rahul Sukthankar, Mubarak Shah. −3. −2. −1 .... pairs that involve category yi plus one (to count for its one-vs-rest classifier).

Spatiotemporal dynamics in a ring of N mutually ...
sates the dissipation in the system. Such a system can then be described by a stable limit cycle. A network of such systems is of interest for technological ...

Notes Mutually Exclusive and Overlapping Events.pdf
Sign in. Loading… Whoops! There was a problem loading more pages. Retrying... Whoops! There was a problem previewing this document. Retrying.

Synchronization dynamics in a ring of four mutually ...
linear dynamical systems, two main classes are to be distinguished: the regular and chaotic one. ..... An illustration of such a behavior is represented in Fig.

Synchronization dynamics in a ring of four mutually ...
Corresponding author. Tel.: +229 97471847; fax: +229 20222455. ...... Adaptive synchronization of chaos for secure communication. Phys Rev E 1997 ...

Selective Promoter Recognition by Chlamydial σ28 Holoenzyme †
Jul 11, 2006 - Quantity One software (Bio-Rad, Hercules, CA). Primer extension analysis. Primer ..... As a control, we ana- lyzed the σ66-dependent promoter ...

Mechanism of Ubiquitin Recognition by the CUE Domain of Vps9p
May 29, 2003 - CUE domain is a dimeric domain-swapped variant of .... 100%. 99.5 a Statistics are shown for the peak wavelength of the Selenomethionine MAD datasets. Statistics for ..... from the CUE dimer and free ubiquitin are identical to.

Mechanism of Ubiquitin Recognition by the CUE Domain of Vps9p
May 29, 2003 - Only the ubiquitin making functionally rele- vant contacts with the ..... (F) Schematic illustration of the lattice relationships. (G) of the second .... Department of Energy, Basic Energy Sciences, Office of Science, novel ubiquitin .

pdf-0738\face-detection-and-recognition-on-mobile-devices-by ...
pdf-0738\face-detection-and-recognition-on-mobile-devices-by-haowei-liu.pdf. pdf-0738\face-detection-and-recognition-on-mobile-devices-by-haowei-liu.pdf.