Sign In to Follow Application
View All Documents & Correspondence

A System For Automatic Detection And Classification Of Video Frames Of Videos

Abstract: A system for automatic detection and classification of video frames of videos is disclosed. The system extracts block features based similarities from each frame of the input video and uses the statistical moments of up to the second order, over an accumulator to measure the motion present in the frames. Further, the system generates feature vectors for each frame using a sliding window over time and these feature vectors are used to train a SVM for classifying the frames as cuts and normal scenes of the video.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
29 August 2008
Publication Number
31/2010
Publication Type
INA
Invention Field
COMMUNICATION
Status
Email
Parent Application
Patent Number
Legal Status
Grant Date
2019-11-29
Renewal Date

Applicants

TATA CONSULTANCY SERVICES LIMITED
BOMBAY HOUSE, 24, SIR HOMI MODY STREET, MUMBAI,

Inventors

1. GOSWAMI KAUSTAV
TATA CONSULTANCY SERVICES LTD., BENGAL INTELLIGENT PARK, BUILDING-D, PLOT NO A2 M2 & N2, BLOCK-EP, SALT LAKE ELECTRONICS COMPLEX, SECTOR-V, KOLKATA-700091,
2. BHOWMIK BROJESHWAR
TATA CONSULTANCY SERVICES LTD., BENGAL INTELLIGENT PARK, BUILDING-D, PLOT NO A2 M2 & N2, BLOCK-EP, SALT LAKE ELECTRONICS COMPLEX, SECTOR-V, KOLKATA-700091,

Specification

FORM -2
THE PATENTS ACT, 1970
(39 of 1970)
&
THE PATENTS RULES, 2003
PROVISIONAL
Specification


(See Section 10 and rule 13)
ANALYZING VIDEO CONTENT

TATA CONSULTANCY SERVICES LTD.,
an Indian Company of Bombay House, 24, Sir Homi Mody Street, Mumbai 400 001
Maharashtra, India
THE FOLLOWING SPECIFICATION DESCRIBES THE INVENTION


Field of the Invention:
This invention relates to the field of analyzing video content.
Background of the Invention:
With increased availability and an exponential growth in the number of image and video applications, the issue of analyzing, automatically indexing and summarizing video content by retrieving highly representative information is of paramount importance and an active area of research. A shot in a video sequence is a contiguous sequence of one or more video frames recorded by a single camera. The nature of transitions or boundaries between shots is varied. Depending on the nature, shot boundaries may be classified as being a cut, a fade, a wipe or a dissolve.
A cut is an abrupt shot change that occurs in a single frame. A fade is a slow change in brightness usually resulting in or starting with a black frame. A dissolve occurs when the frames of the first shot get dimmer and the frames of the second get brighter, with frames within the transition showing one image superimposed on the other. A wipe occurs when pixels from the second shot replace those of the first shot in a regular pattern such as in a line from the left edge of the frames. Of course, other types of gradual transitions are possible.
Video shot boundary detection techniques are an active area of research in recent years and have to challenge the difficulty of finding shot boundaries in the presence of camera, object motion and illumination variations.
2

Several approaches have been reported in the literature for automatic shot
boundary detection. Of them, one class of algorithms exploit color intensity
variations, the second class of algorithms exploits classification based on
mathematical and statistical models; other methods exploit edge difference
and edge comparisons between successive frames. Most of the cut detection
schemes are threshold based. They identify a cut if the frame difference
between consecutive frames exceeds a certain threshold. Similarity between
frames is expressed by extracting features between frames. Various types of
features have been reported in literature. The features are based on:
Luminance/color - The average grayscale luminance is used as a feature but
it is susceptible to illumination changes.
Color Histogram - Because of its insensitivity to camera motion, this
technique is widely used.
Transform Coefficients- They can be used to describe texture which can be
used as a feature.
Edges- Edge features have been used to detect cuts as they are invariant to
illumination and motion.
The size of the region from which features are extracted plays a significant
role on the performance of any shot detection algorithm. If the ROI is large,
there is a tendency to miss transitions between similar shots. The spatial
features used in literature include the following:
Single frame pixel per picture e.g. luminance, edge strength.
Arbitrary shaped region- In order to extract discontinuous regions better,
arbitrary regions have been made the ROI for each frame to exploit
homogeneity.
Whole frame- They have the advantages to be resistant to motion but the
performance may be poor in detecting transition between shots.

The proposed shot boundary detection technique extracts block features from each frame of the video. Motion based features are extracted by taking the statistical moments over an accumulator. Accumulator represents motion histogram over the entire frame. A feature vector is generated using a sliding window over time. These features are trained using a Support Vector Machine (supervised learning) to classify the cut and normal scenes of the video.
An object of this invention is the detection of cuts (shot boundaries) of video sequences.
In particular, this invention envisages a novel way of estimating the motion of frames of input videos using the concept of motion histogram.
In particular, this invention envisages a novel way of shot boundary detection (cut) of a video using Support Vector Machine (SVM).
Further, this invention provides means for automatic detection of cuts from an input video sequence.
In particular, this invention discloses a non-threshold based technique for detection of cuts in an input sequence.
Summary of the Invention:
Browsing video databases has become an extremely difficult job with the ever increasing magnitude of video data. To facilitate the user's access to
4

this huge data, indexing, retrieval and summarization have become an active area of research. Temporal video segmentation is a fundamental step in that direction. The objective of shot boundary detection is to partition the video into meaningful, basic structural units called shots.
In this invention, a shot boundary detection technique is disclosed for cuts. The method extracts block feature based similarities from the frames of the input video. Statistical moments of second order are used to measure the total motion of the frames. Feature vectors are generated using a sliding window over time and are trained by a SVM to identify the cuts.
Brief Description of the Accompanying Drawings:
The invention will now be described with reference to the accompanying drawings, in which
Figure 1: Spiral Search for use in the method in accordance with this
invention
Figure 2: Accumulator for use in the method in accordance with this
invention
Detailed Description of Invention:
The invention envisaged uses both block based minimum features as well as statistical moment based motion histogram approach to extract features. Feature vector has been constructed using a sliding window over time. These features have been trained using Support Vector Machine (SVM). The SVM classifies the video frames as a cut or a normal scene.
5

The frames of the input video have been decomposed into rectangular blocks of sizeMxN. This was done because block based computation tends to reduce detection invariance while a large region tends to miss transitions between similar shots. This approach is invariant to small camera and object motion.
Each block of the current frame has been compared with the previous frame blocks by choosing a window around the corresponding previous frame block. For comparison of blocks, normalized SAD (NSAD) in R, G, and B planes have been used separately. The distance between two blocks in each of R, G and B plane is given by the following equation:
255 xMxN where P is either of R, G or B;(x,y) and (x'yf) denotes the co-ordinates of the
top left of the current and previous blocks respectively. distp is the error between the current block at and previous block at position (.*>'). To get the best fit block, a spiral search [Figure 1] has been made around the (x\y') position in the previous frame. The search window has been defined to be between (-M/2, M/2) along the row direction and between (-N/2, N/2) along
the column direction thereby giving (M+l) (N+l) values for"'^. The minimum error among all these distances was computed as

This range of spiral search may change depending on the fps of the input video.
6

Equation 1 gives the minimum error in two blocks of current and previous frames. The normalized total error in R, G and B planes for the entire frame was computed as:

Here i represents either of R, G or B.
Now we get a total of three errors for a current frame from previous frame in
each of R, G, and B plane. Along with the errors in the R, G and B planes,
the motion of the block also played an important role in determining the cuts
from the normal scenes. We have introduced a novel motion estimation
technique for the entire frame.
Corresponding to the best match, the block movement in both the x and y
directions have been measured for each block in the current frame. The x
and y directional movements have been calculated for each block. If the
increments are represented as Ax and Ay, then
Ax = x — x'
and Ay = y- y' where x' and y' are the best matched block position in
previous frame.
An accumulator [Figure 2] has been designed to capture the motion
histogram of a frame. Let A denote the set of integers{0,l,....M-l}and B
denote the set of integers{0,1....,N-1}. Accumulator f(i ,j) is a function
f :AxB → Z+ {o} where Z+ denotes the set of positive integers. For each
block, (Δx, Δy) is computed and the accumulator f(Δx, Δy) is updated by the
following equation.
f(Δx, Δy) = f(Δx, Δy) ÷ 1;
7

The function f{i,j) is the motion histogram of the whole frame. The index of the accumulator ranges from -M/2 to M/2 along row direction and -N/2 to N/2 along column direction.
Statistical moments provide an ideal feature to capture the amount of dispersion of blocks between frames. Dispersion of order (p, q) is given by:
These moments on motion histogram accumulator will give the amount of block dispersion in each frame.
Four normalized central moments are used as additional features to the three errors in R, G and B plane to prepare the feature vector. A sliding window over time has been constructed in order to prepare the
feature vectors. Corresponding to the current frame F', the
frames have been taken and each provides for 7 features
making a total of 35 features.
Having prepared the feature vector for cuts and normal frames of the video,
it is needed to classify the frames as positive (cuts) and negative (normal
frames).For classification, the well known concept of Support Vector
Machine (SVM) has been used.
The SVM, introduced by Vapnik is a powerful supervised learning algorithm
which is primarily a two class classifier. The optimization criterion of an
SVM is the width of the margin between the positive and the negative
8

examples. An SVM with a large margin yields a good generalization performance. The inputs to a SVM are n-dimensional vectors and the machine classifies the vectors by using a hyperplane. The SVM works in the following manner: Given a training dataset of the form:

Each c' indicates to which class the p-dimensional vector belongs to. Any hyperplane is a set of points x satisfying the equation w x-f-6=0; all training points should satisfy

The parameter b/ || w || determines the offset of the hyperplane from the origin in the direction of the normal vector w. w and b are to be chosen so as to maximize the margin. Since it has been proven mathematically that the
2
margin is , so finding the hyperplane which separates the data with
|| w ||
maximum margin leads to the following optimization problem:
2 II W II
Minimize 2 Subject to

Writing the classification rule in its dual form reveals that the maximum
margin hyperplane and therefore the classification task is only a function of
the support vectors which is nothing but the training data that lie on the
margin.
The dual of the SVM is as follows:
9


In order to handle non-separable data, equation (A) becomes

The optimization problem is reformulated as
Minimize 2
are the upper bounds on training
error and c is the regularization parameter. The dua] formulation is
Maximize
Subject t0
For non-linear SVM, the training vectors are mapped to some higher
dimensional data using a transformation function' in Eqn (3)
becomes
By using a suitable kernel function we do
not need to calculate For our work, we used Gaussian kernel of the form
In order to avoid overfitting, a four fold cross validation of training data was done. The training dataset was partitioned into four sub sets and one of them was tested iteratively with the SVM being trained with the remaining three subsets.
10

While considerable emphasis has been placed herein on the specific steps and elements of the preferred embodiment, it will be appreciated that many alterations can be made and that many modifications can be made in the preferred embodiment without departing from the principles of the invention. These and other changes in the preferred embodiment as well as other embodiments of the invention will be apparent to those skilled in the art from the disclosure herein, whereby it is to be distinctly understood that the foregoing descriptive matter is to be interpreted merely as illustrative of the invention and not as a limitation.
Dated this 29th day of August 2008.

APPLICANTS' PATENT ATTORNEY

Documents

Orders

Section Controller Decision Date
Section 15 DEVASENA AB 2019-10-25
Section 43(1) DEVASENA AB 2019-11-29

Application Documents

# Name Date
1 1825-MUM-2008-RELEVANT DOCUMENTS [28-09-2023(online)].pdf 2023-09-28
1 Other Patent Document [05-10-2016(online)].pdf 2016-10-05
2 abstract1.jpg 2018-08-09
2 1825-MUM-2008-RELEVANT DOCUMENTS [26-09-2022(online)].pdf 2022-09-26
3 1825-MUM-2008-RELEVANT DOCUMENTS [30-09-2021(online)].pdf 2021-09-30
3 1825-mum-2008-power of attorney.pdf 2018-08-09
4 1825-MUM-2008-RELEVANT DOCUMENTS [29-03-2020(online)].pdf 2020-03-29
4 1825-mum-2008-form 5.pdf 2018-08-09
5 1825-MUM-2008-IntimationOfGrant29-11-2019.pdf 2019-11-29
5 1825-MUM-2008-FORM 5(27-8-2009).pdf 2018-08-09
6 1825-MUM-2008-PatentCertificate29-11-2019.pdf 2019-11-29
6 1825-mum-2008-form 2.pdf 2018-08-09
7 1825-MUM-2008-Written submissions and relevant documents (MANDATORY) [04-11-2019(online)].pdf 2019-11-04
8 1825-MUM-2008-HearingNoticeLetter-(DateOfHearing-05-11-2019).pdf 2019-11-01
8 1825-mum-2008-form 2(title page).pdf 2018-08-09
9 1825-MUM-2008-Written submissions and relevant documents (MANDATORY) [25-10-2019(online)].pdf 2019-10-25
9 1825-MUM-2008-FORM 2(TITLE PAGE)-(27-8-2009).pdf 2018-08-09
10 1825-MUM-2008-ExtendedHearingNoticeLetter_11-10-2019.pdf 2019-10-11
10 1825-mum-2008-form 2(27-8-2009).pdf 2018-08-09
11 1825-MUM-2008-FORM 18(4-11-2010).pdf 2018-08-09
11 1825-MUM-2008-HearingNoticeLetter04-10-2019.pdf 2019-10-04
12 1825-MUM-2008-ABSTRACT [29-01-2019(online)].pdf 2019-01-29
12 1825-MUM-2008-FORM 13(27-8-2009).pdf 2018-08-09
13 1825-MUM-2008-CLAIMS [29-01-2019(online)].pdf 2019-01-29
13 1825-mum-2008-form 1.pdf 2018-08-09
14 1825-MUM-2008-COMPLETE SPECIFICATION [29-01-2019(online)].pdf 2019-01-29
14 1825-MUM-2008-FORM 1(18-3-2009).pdf 2018-08-09
15 1825-MUM-2008-FER.pdf 2018-08-09
15 1825-MUM-2008-FER_SER_REPLY [29-01-2019(online)].pdf 2019-01-29
16 1825-mum-2008-drawing.pdf 2018-08-09
16 1825-MUM-2008-OTHERS [29-01-2019(online)].pdf 2019-01-29
17 1825-MUM-2008-DRAWING(27-8-2009).pdf 2018-08-09
17 1825-MUM-2008-PETITION UNDER RULE 137 [29-01-2019(online)].pdf 2019-01-29
18 1825-MUM-2008-FORM-26 [17-01-2019(online)].pdf 2019-01-17
18 1825-mum-2008-description(provisional).pdf 2018-08-09
19 1825-MUM-2008-ABSTRACT(27-8-2009).pdf 2018-08-09
20 1825-MUM-2008-CLAIMS(27-8-2009).pdf 2018-08-09
20 1825-MUM-2008-DESCRIPTION(COMPLETE)-(27-8-2009).pdf 2018-08-09
21 1825-MUM-2008-CORRESPONDENCE(18-3-2009).pdf 2018-08-09
21 1825-mum-2008-correspondence.pdf 2018-08-09
22 1825-MUM-2008-CORRESPONDENCE(27-8-2009).pdf 2018-08-09
22 1825-MUM-2008-CORRESPONDENCE(4-11-2010).pdf 2018-08-09
23 1825-MUM-2008-CORRESPONDENCE(27-8-2009).pdf 2018-08-09
23 1825-MUM-2008-CORRESPONDENCE(4-11-2010).pdf 2018-08-09
24 1825-MUM-2008-CORRESPONDENCE(18-3-2009).pdf 2018-08-09
24 1825-mum-2008-correspondence.pdf 2018-08-09
25 1825-MUM-2008-DESCRIPTION(COMPLETE)-(27-8-2009).pdf 2018-08-09
25 1825-MUM-2008-CLAIMS(27-8-2009).pdf 2018-08-09
26 1825-MUM-2008-ABSTRACT(27-8-2009).pdf 2018-08-09
27 1825-mum-2008-description(provisional).pdf 2018-08-09
27 1825-MUM-2008-FORM-26 [17-01-2019(online)].pdf 2019-01-17
28 1825-MUM-2008-DRAWING(27-8-2009).pdf 2018-08-09
28 1825-MUM-2008-PETITION UNDER RULE 137 [29-01-2019(online)].pdf 2019-01-29
29 1825-mum-2008-drawing.pdf 2018-08-09
29 1825-MUM-2008-OTHERS [29-01-2019(online)].pdf 2019-01-29
30 1825-MUM-2008-FER.pdf 2018-08-09
30 1825-MUM-2008-FER_SER_REPLY [29-01-2019(online)].pdf 2019-01-29
31 1825-MUM-2008-COMPLETE SPECIFICATION [29-01-2019(online)].pdf 2019-01-29
31 1825-MUM-2008-FORM 1(18-3-2009).pdf 2018-08-09
32 1825-MUM-2008-CLAIMS [29-01-2019(online)].pdf 2019-01-29
32 1825-mum-2008-form 1.pdf 2018-08-09
33 1825-MUM-2008-ABSTRACT [29-01-2019(online)].pdf 2019-01-29
33 1825-MUM-2008-FORM 13(27-8-2009).pdf 2018-08-09
34 1825-MUM-2008-FORM 18(4-11-2010).pdf 2018-08-09
34 1825-MUM-2008-HearingNoticeLetter04-10-2019.pdf 2019-10-04
35 1825-MUM-2008-ExtendedHearingNoticeLetter_11-10-2019.pdf 2019-10-11
35 1825-mum-2008-form 2(27-8-2009).pdf 2018-08-09
36 1825-MUM-2008-Written submissions and relevant documents (MANDATORY) [25-10-2019(online)].pdf 2019-10-25
36 1825-MUM-2008-FORM 2(TITLE PAGE)-(27-8-2009).pdf 2018-08-09
37 1825-mum-2008-form 2(title page).pdf 2018-08-09
37 1825-MUM-2008-HearingNoticeLetter-(DateOfHearing-05-11-2019).pdf 2019-11-01
38 1825-MUM-2008-Written submissions and relevant documents (MANDATORY) [04-11-2019(online)].pdf 2019-11-04
39 1825-MUM-2008-PatentCertificate29-11-2019.pdf 2019-11-29
39 1825-mum-2008-form 2.pdf 2018-08-09
40 1825-MUM-2008-IntimationOfGrant29-11-2019.pdf 2019-11-29
40 1825-MUM-2008-FORM 5(27-8-2009).pdf 2018-08-09
41 1825-MUM-2008-RELEVANT DOCUMENTS [29-03-2020(online)].pdf 2020-03-29
41 1825-mum-2008-form 5.pdf 2018-08-09
42 1825-mum-2008-power of attorney.pdf 2018-08-09
42 1825-MUM-2008-RELEVANT DOCUMENTS [30-09-2021(online)].pdf 2021-09-30
43 1825-MUM-2008-RELEVANT DOCUMENTS [26-09-2022(online)].pdf 2022-09-26
43 abstract1.jpg 2018-08-09
44 1825-MUM-2008-RELEVANT DOCUMENTS [28-09-2023(online)].pdf 2023-09-28
44 Other Patent Document [05-10-2016(online)].pdf 2016-10-05

Search Strategy

1 search_13-06-2018.pdf

ERegister / Renewals

3rd: 11 Feb 2020

From 29/08/2010 - To 29/08/2011

4th: 11 Feb 2020

From 29/08/2011 - To 29/08/2012

5th: 11 Feb 2020

From 29/08/2012 - To 29/08/2013

6th: 11 Feb 2020

From 29/08/2013 - To 29/08/2014

7th: 11 Feb 2020

From 29/08/2014 - To 29/08/2015

8th: 11 Feb 2020

From 29/08/2015 - To 29/08/2016

9th: 11 Feb 2020

From 29/08/2016 - To 29/08/2017

10th: 11 Feb 2020

From 29/08/2017 - To 29/08/2018

11th: 11 Feb 2020

From 29/08/2018 - To 29/08/2019

12th: 11 Feb 2020

From 29/08/2019 - To 29/08/2020

13th: 11 Feb 2020

From 29/08/2020 - To 29/08/2021

14th: 28 Aug 2020

From 29/08/2021 - To 29/08/2022

15th: 01 Aug 2022

From 29/08/2022 - To 29/08/2023

16th: 01 Aug 2023

From 29/08/2023 - To 29/08/2024

17th: 31 Jul 2024

From 29/08/2024 - To 29/08/2025

18th: 20 Aug 2025

From 29/08/2025 - To 29/08/2026