Semantic Video Event Search for Surveillance Video Tae Eun Choe1, Mun Wai Lee2, Feng Guo1, Geoffrey Taylor1, Li Yu1, Niels Haering1 1 2 ObjectVideo Inc. Intelligent Automation Inc. Reston, VA, USA Rockville, MD, USA {tchoe, fguo, gtylor, liyu, nhaering}@objectvideo.com

Abstract We present a distributed framework of understanding, indexing, and searching complex events from large amounts of surveillance video content. Video events and relationships between scene entities are represented by Spatio-Temporal And-Or Graphs (ST-AOG) and inferred in a distributed computing system using a bottom-up top-down strategy. We propose a method for sub-graph indexing of ST-AOGs of the recognized events for robust retrieval and quick search. Plain text reports of the scene are automatically generated to describe scene entities’ relationships, contextual information, as well as events of interest. When a query is provided as keywords, plain text, voice, or a video clip, the query is parsed and the closest events are extracted utilizing text description and sub-graph matching.

1. Introduction Video surveillance cameras are deployed worldwide and automatic scene understanding from those camera feeds is a necessary yet challenging problem. Current systems lack a general visual knowledge framework and efficient computational algorithms to represent complex events and to handle large volumes of video data.

[email protected]

commonly used to find similar scenes. However, since we know the locations of surveillance cameras, our interest is more in retrieving events (e.g. approach a vehicle, meet) than in finding similar images. Snoek et al. retrieve video events using time interval [17] and also propose video retrieval concept detectors which handle multi-modal queries (query-by-text, query-by-ontology and query-by-example) and fuse them to find best matching videos [16]. We propose a method for automatic scene understanding and indexing hierarchical semantic data from a large imagery dataset utilizing existing distributed processors. We adopted the And-Or graph [18] to represent and infer scene events and extract semantics and contextual content. The graph provides a principled mechanism to list visual elements and objects in the scene and describe how they are related (see Figure 1). These relationships can be spatial, temporal or compositional. A text generation system then converts inferred information to plain text reports to describe these relationships, contextual information, as well as events of interest [10]. The inferred graphical video events are indexed as a set of sub-graphs and saved in a distributed database.

This study aims to address the following questions: a) How do we integrate the inference of multiple image processing components in a modular architecture? b) What is the suitable representation for video content to facilitate exploitation? c) How do we index and search large amount of video content? Recent works on scene understanding include Markov Logic Networks (MLN) [15] and attribute grammar [21]. Attribute grammars build a scene parse tree using a stochastic model. Similarly to attribute grammars, the And-Or graph (AoG) [18] is introduced for scene understanding, and has a more flexible topological graph structure. For video retrieval, color (histogram or correlogram) and feature-based (e.g. HOG[3], SIFT[11]) methods are

Figure 1. Representation of loading event using Spaio-Temporal And-Or graph. The graphical data is indexed by sub-graphs for efficient and robust search.

2. Video Event Detection Spatio-Temporal And-Or Graph serves as a framework for analysis, extraction, and representation of the visual elements and structure of the scene, such as the ground plane, sky, buildings, moving vehicles, humans, and

interactions between entities. Image content extraction is now formulated as a graph parsing process to find a specific configuration produced by the grammar that best describes the image. The inference algorithm finds the best configuration by integrating bottom-up detection and top-down hypotheses. As illustrated in Figure 1, using a traffic scene as an example, bottom-up detection includes classification of image patches (such as road, land, and vegetation), detection of moving objects, and representation of events, which generate data-driven candidates for scene content. Top-down hypotheses, on the other hand, are driven by scene models and contextual relations represented by the attribute grammar, such as the port scene model and ship-wake model. The fusion of both the bottom-up and top-down approaches results in a more robust video content extraction.

a cloud database to extract and store events (See Figure 2-(b)).

2.1. Scene Element Extraction

We use the MapReduce framework [4] for event detection in the semantic analyzer in a distributed system. A large network of cameras can produce voluminous amounts of primitive data, particularly for crowded scenes with many targets, which must be processed in real time so that events of interest can be detected in a timely manner. Event detection is a prime candidate for parallel processing, due to the large amounts of data involved and the fact that each target trajectory is processed independently.

Analysis of urban scenes benefits greatly from knowledge of the locations of buildings, roads, sidewalks, vegetation, and land areas. Maritime scenes similarly benefit from knowledge of the locations of water regions, berthing areas, and sky/cloud regions. From video feeds, a background image is periodically learned and it is processed to extract scene elements. We first perform over-segmentation to divide the image into super-pixels using the mean-shift color segmentation method. Since adjacent pixels are highly correlated, analyzing scene elements at the super-pixel level reduces the computational complexity. For each super-pixel, a set of local features is extracted and super-pixels are grouped by Markov Random Field and Swanson Cut [1]. The extracted background scene element boosts classification and tracking of a target in the scene after transferred to an event detection routine. Figure 9 shows an example of extracted scene element with sky, vegetation, and road.

2.2. Event Detection in Distributed System Conventional video analytic systems typically use one processor per sensor. The video from the calibrated sensor is processed and metadata of target information is transferred to a semantic analyzer [2][14]. The metadata consists of a set of primitives, each representing the classification type, target ID, timestamp, bounding box and other associated data for a single detection in a video frame. The semantic analyzer track targets across multiple cameras and infers complex events in a common map-based coordinate system. In this framework, all cross-camera processing is concentrated in the semantic analyzer, resulting in a bottleneck and scalability issue for analyzing events with multiple cameras (See Figure 2-(a)). We also observe that the workloads of video analytic processors are not evenly distributed. Therefore, we utilize pre-existing processors as distributed computing nodes and

Video Analyzer Video Analyzer

… Video Analyzer

Video VideoSemantic Analyzer Analyzer Analyzer

Semantic Analyzer

DB

Video VideoSemantic Analyzer Analyzer Analyzer Cloud DB



Processors

Video VideoSemantic Analyzer Analyzer Analyzer

Processors

(a) (b) Figure 2: Comparison between (a) classic video analytics and (b) proposed distributed video analytics framework

Figure 3: Hadoop implementation of event detection. MapReduce Implementation Figure 3 shows the mapping of the event detection analytics to the MapReduce framework using Hadoop [23]. The input to the process is a set of primitive files, each containing a set of primitives collected over a period of time from a single camera. The set of primitive files may correspond to data collected from a different camera over different windows of time, and/or concurrent data from different cameras. These files are transferred into HDFS for processing. The output is a single event file, also written in to HDFS, which is used for cloud database system. This is a text file in which each line lists the parameters of the low-level events detected over the entire input data set. The gain of a use of distributed computing is discussed in Section 5.1.

2.3. Complex Event Recognition From the semantic analyzer, more meaningful video information is generated, including: (i) agent (human, vehicle, or general agent), (ii) properties of agent such as UTC time and location (in latitude/longitude), and (iii) low-level events of agent (appear, disappear, move, stationary, stop, start-to-move, turn, accelerate, decelerate, etc.) Spatial (far, near, beside) and temporal (before, after, during, etc.) relationships are also estimated for the Spatio-Temporal And-Or graph representation of complex events. From training data, threshold values of location and time are learned and Spatio-Temporal And-Or-graphs of the following events are built especially for video surveillance applications: • Single agent composite events: stop / start-to-move, turn, accelerate / decelerate. •

2-agent complex events: approach / move-away, lead / follow, catch-up, over-take, meet.



human-vehicle interaction events:



ƒ

embark / disembark,

ƒ

park (a person disembarks a vehicle and the vehicle remains stationary) / ride (a vehicle was stationary, a person embarks the vehicle, and the vehicle drives away),

ƒ

drop-passenger (a person disembarks a vehicle and the vehicle drives away) / pickup-passenger (a vehicle arrives, a person embarks, and the vehicle drives away) ,

ƒ

loiter-around,

ƒ

load / unload.

multi-vehicle-human complex events: switch-vehicle, switch-driver, exchange-vehicle, convoy, queuing.

Figure 4. Inference of a complex event (pick-up) using AoG.

Figure 4 illustrate inference of a pick-up event. When a vehicle appears in the scene and stops, a human approaches the vehicle and disappears, and then the vehicle leaves the scene, we define this event as a pick-up event and represent it in the Spatio-Temporal And-Or graph. We use the simplified Earley-Stolcke parsing algorithm [5] to infer an event based on a particular event grammar iteratively.

2.4. Automatic Text Generation Automatic generation of text descriptions is the key to video search from query-by-keyword/text/voice. After the complex video events are recognized through the semantic analysis engine, those events are transformed to a plain text report. The text generation process adopts a common architecture for natural language generation systems [20], including a pipeline of two distinct tasks: text planning and text realization. The text planner selects the content to be expressed, specifies hard sentence boundary, and provides information about the content. Based on this formation; the text realizer generates the sentences by determining grammatical form and performing word substitution. Figure 5 shows the two-step diagram of text generator. Text Generation Engine

Scene Analysis Result

Document Structure

Phrase Structure Grammar

Text Planner

Text Realizer

Lexical Structure

Symbolic Generator

Natural Language Reports

Figure 5. Flow diagram of text generation Text planner The Text Planner module translates the semantic representation to a representation that can readily be used by the Text Realizer to generate text. This immediate step is used to convert a representation that is semantic and ontology-based to a representation based on lexical structure. Text Planner converts scene analysis results to a functional description (FD) which has a feature-value pair structure. This structure is commonly used in text generation input schemes such as the HALogen [9] and Sentence Planning Language (SPL) [7]. The functional description is stored in a file called text-planner-xml. We use XML as the output format so that standard xml parsing tool can be used, but it essentially follows the same labeled feature-value structure of the logic notation used in HALogen [9]. For each sentence, the functional description language specifies the details of the text we want to generate, such as the process

(or event), actor, agent, predicates, and other functional properties. Text realizer A simplified head-driven phrase structure grammar (HPSG) [13] is used to generate text sentences. HPSG consists of two main components (i) a highly-structured representation of grammatical categories, (ii) a set of descriptive constraints for phrasal construction. The input is the functional description with a collection of features. HPSG consists of a generation grammar representing the structure of features with production rules. For video surveillance applications, generated texts are mostly indicative or declarative sentences and this simplifies the grammar structure significantly. A unification process [8] matches the input features with the grammar in a recursively manner. The derived lexical tree is then linearized to form sentence output. We associate the generated text with corresponding video clips and geo-tracks, and save them in Video Event Markup Language (VEML) [12]. An example of an automatically generated text description of a video is shown in Figure 9.

3. Video Event Indexing using Sub-Graphs To efficiently search a semantic video event from large amounts of video content, we propose an indexing scheme based on semantic labels and graphical data. A robust search method is required to handle noisy data, especially when a query video clip with the event-of-interest is provided. From ST-AoG, inferred scene structure and complex events are represented by relational graphs. Therefore, querying this data becomes a graph similarity search problem. This involves matching nodes with similar attributes and topological structure. In spectral graph theory, spectral decomposition is used to represent graphs in a vector space which encodes important structural properties of the graphs. Two graphs are compared using vector differencing. However, for video content, not all nodes are relevant to the query. Therefore, inspired by a substructure similarity search approach [19], we propose a new graph indexing method for content-based retrieval. A visual scene can be characterized by a set of sub-graph structures. A feature vector is used to represent the scene where each entry counts the occurrences of a specific substructure in the graph. Matching is based on vector comparison. As feature vectors are pre-computed for all candidate graphs, searching is very efficient. A schematic illustration of this framework is shown in Figure 6.

Figure 6. Scalable content indexing and retrieval using graph substructure similarity search. Indexed features are used to represent semantic labels and sub-graphs.

3.1. Definition of sub-graph indexing All events that occurred in a video are parsed and saved as ST-AoG. A graph G(V, E) is defined by a set of nodes V and a set of edges E. In the graph, each node v∈V represents an agent or an event, and each edge e∈E correspond to the relationship between an agent and event (has) or between two events (spatio-temporal relationships). A query Q is also a graph from the same definition, for example, “a vehicle stops and a human comes out” query is illustrated in Figure 7. Now, the video event searching problem is reduced to finding a sub-graph S∈G from the video to match the query: S=Q.

Move

Slow

during has

before

has

Vehicle has

before

before

Stop near Appear before

Move

has Human

has

Figure 7. A graph for a complex event (disembark event) For efficient sub-graph matching, we build a graph index. A feature f∈G contains several nodes on the graph and the corresponding relationship edges, as shown in Figure 8. Obviously if the query contains a feature: f0∈Q, the graph must also contains the same feature f0∈G if it has a sub-graph that matches the query. Using this concept, we generate a set of features. Given a video, we count the number of these features appearing in the graph.

Move

during

before

Slow

has has Vehicl before before has

Appear

Appear

before Huma

Stop near before

has Human

Appear

Vehicl has Stop

Move

has

has Vehicle

Slow

Vehicle

Move has

before

Move

has Human

has

Move has

has has Stop

Appear

near Slow

during

near

has

Stop before

Stop

before before

Vehicle has Stop

Figure 8. Example of a set of features from the graph in Figure 7. Whenever a query is given, we check which features the query contains, and whether they are contained in the candidate videos. Because it is just a lookup table, the process can be done very quickly as shown in Figure 6. Another advantage is the relaxation of conditions. Suppose the query has a total of n features f0, f1, f2 ,…, fn, but the video has m < n of them. Although G does not contain Q, it may contain a sub-graph very similar to Q, if m ≈ n. We conclude the video contains events similar to the query and use the missing features to describe the difference between the query and sub-graph we found.

(ii) given a plain text query, it is parsed [6] and keywords such as action verb, subject, object, time, and location are extracted and renamed using WordNet for keyword-based query; (iii) a voice query is transformed to a text query using COTS Text-to-voice engine [22] and then follows the query-by-text process; (iv) when a video clip is presented as a query, the system provides two ways to handle query-by-example. First, the query video is analyzed by the semantic analyzer and the names of detected events are extracted. Then, using keyword-based search, corresponding events are listed. Secondly, when a user does not want text-based results or the query video contains an unknown event, the user can choose to run the sub-graph matching method. When corresponding events are retrieved, the system generates dynamic web pages containing video clips, text descriptions of events, and geo-locations as shown in Figure 9.

3.2. Implementation of substructure indexing In implementation, the agent types used include: (LAND_VEHICLE, HUMAN, OTHERS); event types used include: (appear, move, turn, decelerate, stop, stay stationary, start moving, accelerate, disappear), and relation types used include: (has Agent, has Patent, has SubEvent, interval Before, interval Meets, interval Overlaps, interval Starts, interval Finishes, interval During, interval Equals, spatial Equals, spatial Near). To collect the features, we first generate triples containing two nodes and one edge from one or two events, and index them based on their node/edge types. Given a pair of triples, we generate features based on their temporal and spatial relationship. We count the number of features in the video and save it for query. Because a video may have a long duration, while a query is generally short, we divide the video into clips and perform the index query on each clip. This helps determine a more accurate time stamp for the event.

4. Query-by-Keyword / Text / Voice / Video We implemented a web-based video event search system which handles multi-modal queries as follows: (i) for a keywords query, the classic frequency based indexing is applied to search for the corresponding event utilizing automatically generated text description of video events;

Figure 9. An example of retrieved webpage. Automatically generated text describes overall video information, scene type, and detected events. Hyperlinks directs further detailed explanation of each target and event.

5. Experimental Results 5.1. Gain from distributed computing To evaluate distributed event detection system experiment, the input data was the primitive file from a test video of approximately 2 minutes duration, showing pedestrians and vehicles in a parking lot which contains 15 targets and 146 low-level events. To simulate a larger data set, the primitive file was replicated to generate 2 minutes to 7 hours dataset. To fully explore the impact of cluster size and data size on processing performance, the MapReduce task was run on input varying from 1 to 200 datasets and with the cluster size varying from 1 to 4 nodes. For comparison, the linear

Matching Score

(non-Hadoop) analytics implementationn was applied to the same datasets.

Query

Time

Figure 10: Hadoop processing times ffor event detection as a function of the size of the datasett. Red dot indicates break-even points when distributed ssystem overpasses single processor. Figure 10 plots the processing time aggainst both the input size and number of Hadoop nodes. For a cluster of 4 nodes, the Hadoop implementation is more effiicient than the linear implementation only when the inpuut size exceeds 60 datasets (equivalent to approximagely 990 MB of total data, 180 000 primitives or 2 hours of video). This represents the point at which the additional overheaad (data duplication network communication) imposed by H Hadoop is offset by the availability of additional processingg power.

5.2. Query-by-example using sub--graph indexing To evaluate sub-graph indexing, we buuilt graph indices for 27 surveillance videos. There weree more than 9000 sub-graph features. We evaluate eaach feature by its frequency in the training data. A moree common feature is assigned less weight and a rarer featuree is assigned higher weight for the comparison between tho se features. For this test, a short video clip contaaining a disembark event is selected as a query as show wn in Figure 12-(a). Sub-graph features from the query clipp are extracted and those features are compared with subb-graphs of all the stored videos for sub-graph matchinng. Because stored videos typically have long duration while a query is generally short, we divide the stored vvideos into multiple clips and perform the index query on eeach clip. This helps determine more accurate time stamp s for the retrieved events. The estimation of the weighted ssub-graph similarity measure for each time slot is plotted inn Figure 11, where a higher bar indicates a better match.

Time

Figure 11. Matching score forr sub-clips in the two test videos, where 1 means perfectt matching, i.e. all features are found in the sub-clip. Prim mitive events are plotted to show the time correspondencee. The left graph contains both the query (blue, Figu ure 12-(a)) and highest matching event (red, Figure 12-(b)). The right graph contains a second matching ev vent (red, Figure 12-(c)). The query (Figure 12-(a)) inclludes the disembark event with some noise. For instance, a vehicle passes in front of the disembarking event at a third frame. Sub-graph indexing and matching helped to handle noisy features in both the query and retrieved cliips. Except the query itself, Figure 12-(b) is the best matchin ng video clip from the same sensor with the query. The diseembarking event happens at the upper deck of the parking lo ot as shown in the red box. The second rank video with disembark event from another sensor is shown in Figure 12-(c)).

(a) Query viideo clip

(b)The retrieved first f rank clip

(c) The retrieved seecond rank clip Figure 12. Snapshots of a qu uery and retrieved video clips.

5.3. Evaluation of event rettrieval Several detection results on human-vehicle interaction nd park) and 2-agent events events (ride, drop passenger, an (meeting) are shown below. Th he detection result of drop passenger event is shown as snaapshots in Figure 13. In this case, we detect an event that a specific asset (a vehicle) stops, human disembarks, and d the vehicle start moving away is detected as drop passen nger event.

An example of the re-directed page describing the event is shown in Figure 16.

Figure 13. An example of detected drop passenger event. The detection result for meeting event is shown in Figure 14. In case that two persons move and get closer each other and remain stationary is identified as meeting event.

Figure 14. An example of detected meeting event. We tested on 12 videos, which contain 25 complex events. The recall of the detection was 76% and precision was 90.5%. The missed detection was mostly caused by wrong classification of targets (human and vehicle) and their occlusion. However, with reasonable detection and tracking results, the most of the complex events are accurately retrieved.

Figure 16. The re-directed page after approach link is selected.

6. Query and Retrieval Tool

6.1. Video event search on smartphone

The semantic search engine is developed to retrieve the events from all the processed videos based on the user request. This search engine accepts four different kinds of user requests, keywords, plain texts, voice, and a video clip. User can search for a specific event or a specific event in a particular video or all the events in a particular video or specific events associated with the targets. Users can view videos and plain text as search results.

The emergence of commercial wireless handheld devices now enables users to access and visualize rich geospatial data, even in remote locations. The potential benefits of this new technology are enormous. However, to make a smart phone application usable, we need to overcome challenges like limited bandwidth and small display screen size. In this study, a web-based smart phone application has been developed for users to submit queries and display query result in both text and map modes (See Figure 17).

An example of search results using keyword along with spatio-temporal information is shown in Figure 15.

(a)

(b)

(c)

Figure 17. Query UI on smartphone (a) Initial map view (b) Query result. The text descriptions can be spoken by text-to-speech converter[22]. (c) A smartphone with video search app. Figure 15. An example of search results with spatial and temporal parameters.

7. Conclusion A novel distributed architecture for automatic visual event analysis and search has been developed. Based on Spatio-Temporal And-Or graph, sub-graph indexing, and a Hadoop implementation, our approach is able to analyze, store, index, and search for complex visual events from different cameras in a distributed and scalable framework. Future works includes automatic building of an And-Or-Graph from training data and better inference using intent/prediction on-the-fly through an improved Earley algorithm. In sub-graph indexing, as the number of graph nodes increases, the number of sub-graphs increases exponentially. The learning process of sub-graph indexing should be optimized to discard less important features. In addition, better weighting of each sub-graph feature is also required.

Acknowledgment This material is based upon work supported by the Office of Naval Research under Contract number N00014-11-C-0308.

References [1]

A. Barbu, S.C. Zhu, “Graph partition by Swendsen-Wang cut,” ICCV, 2003. [2] T. E. Choe, M. Lee, N. Haering, “Traffic Analysis with Low Frame Rate Camera Network”, First IEEE Workshop on Camera Networks (WCN2010), held in conjunction with CVPR, 2010. [3] N. Dalal, B. Triggs, “Histograms of oriented gradients for human detection,” CVPR, 2005. [4] J. Dean and S. Ghemawat, “MapReduce: Simplified Data Processing on Large Clusters,” Sixth Symposium on Operating System Design and Implementation, 2004. [5] J. Earley, “An efficient context-free parsing algorithm”, Communications of the Association for Computing Machinery, 13:2:94-102, 1970. [6] C. Fellbaum, “Wordnet: An Electronic Lexical Database,” Cambridge: MIT Press. [7] H. Jhuang, et al., “A Biologically Inspired System for Action Recognition.” ICCV 2007. [8] K. Knight, “Unification: A Multidisciplinary Survey.” ACM Computing Surveys. 21 (1) (1989). [9] I. Langkilde-Geary, K. Knight, "HALogen Input Representation", http://www.isi.edu/publications/licensed-sw/halogen/ interlingua.html [10] M. W. Lee, A. Hakeem, N. Haering, S.C. Zhu, “SAVE: A framework for Semantic Annotation of Visual Events”, IEEE First Workshop on Internet Vision, June 2008.

[11] D. Lowe, “Distinctive image features from scale-invariant keypoints,” IJCV, 60(2):91–110, 2004. [12] R. Nevatia, J. Hobbs, B. Bolles, “An Ontology for Video Event Representation,” IEEE Workshop on Event Detection and Recognition, June 2004. [13] C. Pollard, I. A. Sag, (1994). Head-Driven Phrase Structure Grammar. Chicago, IL: University of Chicago Press. [14] Z. Rasheed, G. Taylor, L. Yu, M. W. Lee, T. E. Choe, F. Guo, A. Hakeem, K. Ramnath, M. Smith, A. Kanaujia, D. Eubanks, N. Haering, "Rapidly Deployable Video Analysis Sensor Units for Wide Area Surveillance," First IEEE Workshop on Camera Networks (WCN2010), held in conjunction with CVPR 2010, 14 June, 2010. [15] M. Richardson, P. Domingos “Markov logic networks.” Mach. Learn., 62:107–136, 2006. [16] C. G. M. Snoek, B Huurnink, L Hollink, M.D. Rijke, G. Schreiber, M. Worring, “Adding semantics to detectors for video retrieval,” IEEE Trans. on Multimedia, 2007. [17] C. G. M. Snoek, M. Worring, "Multimedia Event-Based Video Indexing Using Time Intervals," IEEE Trans. on Multimedia, Vol.7, NO.4, AUGUST 2005. [18] T. Wu, S. C. Zhu “A Numeric Study of the Bottom-up and Top-down Inference Processes in And-Or Graphs,” ICCV, 2009. [19] X. Yan, P. S. Yu, and J. Han, “Substructure Similarity Search in Graph Databases,” SIGMOD, June 2005. [20] B. Yao, X. Yang, L. Lin, M. W. Lee, S.C. Zhu, “I2T: Image Parsing to Text Description”, Proceedings of IEEE, Vol 98, no.8, pp 1485-1508, August, 2010. [21] S. C. Zhu, D. Mumford, “Quest for a stochastic grammar of images”, Foundations and Trends of Computer Graphics and Vision, vol.2, no.4, pp259-362, 2006. [22] Google Text To Speech, http://desktop.google.com/ plugins/i/ texttospeech_bijoy.html [23] Hadoop: http://hadoop.apache.org/

Semantic Video Event Search for Surveillance Video

2.2. Event Detection in Distributed System. Conventional video analytic systems typically use one processor per sensor. The video from the calibrated sensor.

2MB Sizes 2 Downloads 277 Views

Recommend Documents

Group Event Detection for Video Surveillance
and use an Asynchronous Hidden Markov Model (AHMM) to model the .... Based on the description of SAAS, before detecting the symmetric activity of each ...

VERSA – Video Event Recognition for Surveillance ...
Nov 3, 2008 - streamed by a lower-level video analytics service that provides basic ..... Similarly, supervised learning methods ...... flatten([R|Relsx],Rels).

Outdoor Video Surveillance System
system just transmits the acquired video to a base station where most of the processing is ..... Barnard Communications, IEEE conference on UAVs, 2007.

Human Activity Recognition for Video Surveillance
of unusual event recognition with lack of training data. Zelnik-Manor et al. .... a GMM classifier Ci for each CFV Fi with MAP (Maximum a. Posteriori) principle, as ...

Video Surveillance for Biometrics: Long-Range Multi-Biometric System
recognition system for long-range human identification. The system is capable of ...... International Conference on Audio- and Video- based. Biometric Person ...

[PDF] Digital Video Surveillance and Security, Second ...
Online PDF Digital Video Surveillance and Security, Second Edition, Read .... this revised edition addresses the newest technologies and solutions available on ...

Interactive Exploration of Surveillance Video through ...
provide tools for spatial and temporal filtering based on regions of interest. This allows analysts to filter .... therefore challenging for an automatic vision algorithm, and (b) need ...... feedback on the final deployment of our tools. We also tha

Time-Series Linear Search for Video Copies Based on Compact ...
use a sequence-level signature as a compact representation of a video sequence based ... etecting copies of digital documents, such as text, image, and video ...

Time-Series Linear Search for Video Copies Based on Compact ...
In recent years, the maturity of hardware and software has generated enormous amounts of ... age/video content for infringement detection and data mining. [25]; blog operators to identify near-duplicate images/videos for piracy removal, search result

Semantic Video Trailers - Research at Google
Sep 7, 2016 - hours on YouTube (Youtube Blog Statistics, 2008). With the rapid growth of available ..... resentation, 24(6):729–738, 2013. Bengio, Yoshua ...

Automatic, Efficient, Temporally-Coherent Video ... - Semantic Scholar
Enhancement for Large Scale Applications ..... perceived image contrast and observer preference data. The Journal of imaging ... using La*b* analysis. In Proc.

Online Video Recommendation Based on ... - Semantic Scholar
Department of Computer Science and Technology, Tsinghua University, Beijing 100084, P. R. ... precedented level, video recommendation has become a very.

Time-Series Linear Search for Video Copies Based on Compact ...
supported in part by the National Science Council of Taiwan under Grants NSC. 99-2221-E-415-011 and NSC 99-2631-H-001-020. C. Y. Chiu is with the Department of Computer Science and Information. Engineering ...... University, Taiwan, in 1997, and the

PATTERN BASED VIDEO CODING WITH ... - Semantic Scholar
quality gain. ... roughly approximate the real shape and thus the coding gain would ..... number of reference frames, and memory buffer size also increases.

Scalable Video Summarization Using Skeleton ... - Semantic Scholar
the Internet. .... discrete Laplacian matrix in this connection is defined as: Lij = ⎧. ⎨. ⎩ di .... video stream Space Work 5 at a scale of 5 and a speed up factor of 5 ...

Scalable Video Summarization Using Skeleton ... - Semantic Scholar
a framework which is scalable during both the analysis and the generation stages of ... unsuitable for real-time social multimedia applications. Hence, an efficient ...

A Topic-Motion Model for Unsupervised Video ... - Semantic Scholar
Department of Electrical and Computer Engineering. Carnegie Mellon University ..... Locus: Learning object classes with unsupervised segmentation. In IEEE ...