Video content analysis


Video content analysis is the capability of automatically analyzing video to detect and determine temporal and spatial events.
This technical capability is used in a wide range of domains including entertainment, video retrieval and video browsing, health-care, retail, automotive, transport, home automation, flame and smoke detection, safety and security. The algorithms can be implemented as software on general purpose machines, or as hardware in specialized video processing units.
Many different functionalities can be implemented in VCA. Video Motion Detection is one of the simpler forms where motion is detected with regard to a fixed background scene. More advanced functionalities include video tracking and egomotion estimation.
Based on the internal representation that VCA generates in the machine, it is possible to build other functionalities, such as identification, behavior analysis or other forms of situation awareness.
VCA relies on good input video, so it is often combined with video enhancement technologies such as video denoising, image stabilization, unsharp masking and super-resolution.

Functionalities

Several articles provide an overview of the modules involved in the development of video analytic applications.
This is a list of known functionalities and a short description.
FunctionDescription
Dynamic maskingBlocking a part of the video signal based on the signal itself, for example because of privacy concerns.
Flame and smoke detectionIP cameras with intelligent video surveillance technology can be used to detect flame and smoke in 15–20 seconds or even less because of the built-in DSP chip. The chip processes algorithms that analyzes the videos captured for flame and smoke characteristics such as color chrominance, flickering ratio, shape, pattern and moving direction.
Egomotion estimationEgomotion estimation is used to determine the location of a camera by analyzing its output signal.
Motion detectionMotion detection is used to determine the presence of relevant motion in the observed scene.
Shape recognitionShape recognition is used to recognize shapes in the input video, for example circles or squares. This functionality is typically used in more advanced functionalities such as object detection.
Object detectionObject detection is used to determine the presence of a type of object or entity, for example a person or car. Other examples include fire and smoke detection.
RecognitionFace recognition and Automatic Number Plate Recognition are used to recognize, and therefore possibly identify, persons or cars.
Style detectionStyle detection is used in settings where the video signal has been produced, for example for television broadcast. Style detection detects the style of the production process.
Tamper detectionTamper detection is used to determine whether the camera or output signal is tampered with.
Video trackingVideo tracking is used to determine the location of persons or objects in the video signal, possibly with regard to an external reference grid.
Video error level analysisVideo scene content tamper analysis using free software. Video Error level analysis
Object co-segmentationJoint object discovery, classification and segmentation of targets in one or multiple related video sequences

Commercial applications

VCA is a new technology. New applications are frequently found, however the track record of different types of VCA differs widely. Functionalities such as motion detection and people counting and gun detection are believed to be available as commercial off-the-shelf products with a decent track-record, even freeware such as dsprobotics Flowstone can handle movement and color analysis.
In many domains VCA is implemented on CCTV systems, either distributed on the cameras or centralized on dedicated processing systems. Video Analytics and Smart CCTV are commercial terms for VCA in the security domain. In the UK the BSIA has developed an introduction guide for VCA in the security domain. In addition to video analytics and to complement it, audio analytics can also be used.
Independent video management software manufacturers are constantly expanding the range of the video analytics modules available. With the new suspect tracking technology, it is then possible to track all of this subject's movements easily: where they came from, and when, where, and how they moved. Within a particular surveillance system, the indexing technology is able to locate people with similar features who were within the cameras’ viewpoints during or within a specific period of time. Usually, the system finds a lot of different people with similar features and presents them in the form of snapshots. The operator only needs to click on those images and subjects which need to be tracked. Within a minute or so, it’s possible to track all the movements of a particular person, and even to create a step-by-step video of the movements.
Kinect is an add-on peripheral for the Xbox 360 gaming console that uses VCA for part of the user input.
In retail industry, VCA is used to track shoppers inside the store. By this way, a heatmap of the store can be obtained, which is beneficial for store design and marketing optimisations.
The quality of VCA in the commercial setting is difficult to determine. It depends on many variables such as use case, implementation, system configuration and computing platform. Typical methods to get an objective idea of the quality in commercial settings include independent benchmarking and designated test locations.
VCA has been used for crowd management purposes, notably at The O2 Arena in London and The London Eye.

Law enforcement

Police and forensic scientists analyse CCTV video when investigating criminal activity. Police use software, such as Kinesense, which performs video content analysis to search for key events in video and find suspects. Surveys have shown that up to 75% of cases involve CCTV. Police use video content analysis software to search long videos for important events.

Academic research

Video content analysis is a subset of computer vision and thereby of artificial intelligence. Two major academic benchmark initiatives are TRECVID, which uses a small portion of i-LIDS video footage, and the PETS Benchmark Data. They focus on functionalities such as tracking, left luggage detection and virtual fencing. Benchmark video datasets such as the enables action recognition researches incorporating temporal and spatial visual attention with convolutional neural network and long short-term memory. Video analysis software is also being paired with footage from body-worn and dashboard cameras in order to more easily redact footage for public disclosure and to identify events and people in videos.
The EU is funding a FP7 project called P-REACT to integrate video content analytics on embedded systems with police and transport security databases.