Sign In to Follow Application
View All Documents & Correspondence

Method And System Of Auto Tagging Brands Of Television Advertisements

Abstract: The present disclosure provides a computer-implemented method and system for auto-tagging of one or more advertisements broadcasted on a channel in real time. The computer-implemented method includes a fetching of a set of prominent frames and a pre-defined section of an audio clip corresponds to a detected advertisement. Further, the computer-implemented method includes a retrieving of a plurality of features corresponds to the set of prominent frames and the pre-defined section of the audio clip. Further, the computer-implemented method includes a comparison of each of the plurality of features with corresponding pre-defined set of features. Furthermore, the computer-implemented method includes a tagging of the detected advertisement with a unique tag.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
09 March 2016
Publication Number
37/2017
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
nishantk@ediplis.com
Parent Application

Applicants

Silveredge Technologies Pvt. Ltd.
Plot No. 131, 2nd Floor, Sector 44, Gurgaon

Inventors

1. Debasish Mitra
Plot No. 131, 2nd Floor, Sector 44, Gurgaon 122002, Haryana
2. Hitesh Chawla
G1701, Bestech Park View Spa, Sector 47, Gurgaon – 122002, Haryana

Specification

TECHNICAL FIELD
 The present invention relates to the field of digital fingerprinting of media content and, in particular, relates to auto-tagging of one or more advertisements broadcasted in real time.
5
BACKGROUND
 A television broadcast essentially consists of scheduled programs and sponsored advertisements. Each advertisement is generally scheduled to run for 10 to 35 seconds approximately on multiple channels. The advertisements are provided by advertisers to run in between the scheduled broadcast of the program 10 on each channel. These advertisements generate revenues for the company (advertisers) who is into advertising. Also, these advertisements are important source of revenue and marketing for the channels. As the revenue model of each advertiser is closely associated with airing of their own advertisement or their competitor’s advertisements, there is an increased competition between each 15 channel to rope in more advertisers and advertisements. This has created a need for detecting airing, frequency and duration of each advertisement broadcasted on their channel and their competitive channels.
 To detect the airing, frequency and duration of each advertisement, 20 advertisement detection systems are used. In these advertisement detection systems, manual interference is required during the new advertisement detection. In various prior arts, such advertisement verification or collection procedures were manually performed by human beings during scheduled broadcast time, or by visually searching (fast forwarding, rewinding, etc.) a tape or other record of 25 an earlier broadcast. As can be appreciated, waiting for the advertisement to broadcast, setting up recording equipment to record a broadcast, and/or searching records of broadcast content to verify advertisement content broadcastings can be time consuming, laborious, and costly undertakings.
30
Page 3 of 40
 These advertisements can be primarily detected through an unsupervised machine learning based approach and a supervised machine learning based approach. The unsupervised machine learning based approach focuses on detection of advertisements by extracting and analyzing digital fingerprints of each advertisement. Similarly, the supervised machine learning based approach 5 focuses on mapping and matching digital fingerprints of each advertisement with a known set of digital fingerprints of corresponding advertisement. Furthermore, these advertisements can be tagged automatically with a unique tag (herein “brand name”) for unsupervised detection using predictive analysis.
10
 In US patent application US 13/832,083 a method and system for broadcast ad identification is presented. The method includes the steps of providing fingerprint signatures of each frame in a broadcast video; and designating at least two repeat fingerprint signatures upon detecting at least one fingerprint-signature match from the signatures. Preferably, methods further 15 include: prior to the designating, determining whether the fingerprint signatures correspond to a known ad based upon detecting at least one fingerprint-signature match of the fingerprint signatures with pre-indexed fingerprint signatures of pre-indexed ads. Preferably, method further include creating segments of the fingerprint signatures, ordered according to a timeline temporal proximity of the 20 fingerprint signatures, by grouping at least two fingerprint signatures based on a repeat temporal proximity of at least two repeat fingerprint signatures respective of at least two fingerprint signatures. Preferably, methods further include detecting at least one ad candidate based on an occurrence of at least one repeat segment. 25
 In another US patent application US 11/613,822 the method and system for automated auditing of advertising is presented. The timing and placement of advertising on TV, radio or other broadcast media are automatically verified or audited by monitoring and recording channels of TV, radio or 30 broadcast media by storing and tagging discrete portions of segments of the
Page 4 of 40
broadcast signals in a database. The system includes a controller, or “dispatcher” server for dispatching the files to an analysis server for performing various mathematical comparisons and statistical correlations on the audio and video signals for positively identifying one or more advertisements of interest. Further, a report is generated, providing particulars about the airing times of the 5 advertisement of interest and whether its content exactly matches the content of a reference advertisement used as the basis for the mathematical comparisons and correlations.
 The present systems and methods have several disadvantages. Most 10 of the methods and system rely on manual tagging of new advertisements. This is somewhat slow and requires 24 hour staff support. In addition, manual tagging may be flawed due to shear negligence of any staff member. These prior arts are time consuming, laborious, and costly undertakings. In addition, these prior arts lack the precision and accuracy to one advertisement from another. These prior 15 arts lack any approach and technique for an automated unsupervised detection of any new advertisements.
 In light of the above stated discussion, there is a need for a method and system which overcomes the above stated disadvantages. 20
SUMMARY
 In an aspect, the present disclosure provides a computer-implemented method for detecting one or more advertisements broadcasted on a channel in real time. The computer-implemented method includes a fetching of a set of 25 prominent frames and a pre-defined section of an audio clip. The set of prominent frames and the pre-defined section of the audio clip corresponds to a detected advertisement. Further, the computer-implemented method includes a retrieving of a plurality of features. The plurality of features corresponds to the set of prominent frames and the pre-defined section of the audio clip. Further, the 30 computer-implemented method includes a comparison of each of the plurality of features with corresponding pre-defined set of features. Furthermore, the
Page 5 of 40
computer-implemented method includes a tagging of the detected advertisement with a unique tag.
 In an embodiment of the present disclosure, the plurality of features includes a brand logo displayed in one or more prominent frames of the set of 5 prominent frames. In addition, the plurality of features includes a brand tagline displayed in the one or more prominent frames of the set of prominent frames. Moreover, the plurality of features include a brand tagline recited corresponding to the pre-defined section of the audio clip.
10
 In an embodiment of the present disclosure, the pre-defined set of features are stored in a reference database.
 In an embodiment of the present disclosure, the tag is a brand name corresponding to the detected advertisement. 15
 In an embodiment of the present disclosure, the computer-implemented method further includes an extraction of a first set of audio fingerprints and a first set of video fingerprints. The first set of audio fingerprints and the first set of video fingerprints corresponds to a media content broadcasting 20 on the channel. The first set of audio fingerprints and the first set of video fingerprints are extracted sequentially in the real time. Moreover, the extraction of the first set of video fingerprints is done by sequentially extracting one or more prominent fingerprints. The one or more prominent fingerprints corresponds to the one or more prominent frames of a pre-defined number of frames present in 25 the media content for a pre-defined interval of broadcast.
 In an embodiment of the present disclosure, the computer-implemented method further includes a generation of a set of digital signature values. The digital signature values corresponds to an extracted set of video 30 fingerprints. The generation of each digital signature value of the set of digital signature values is done by dividing each prominent frame of the one or more
Page 6 of 40
prominent frames into a pre-defined number of blocks. Further, each block of each prominent frame of the one or more prominent frames is gray scaled. Furthermore, the generation of each digital signature value of the set of digital signature values is done by calculating a first bit value and a second bit value for each block of the prominent frame. In addition, the generation of each digital 5 signature value of the set of digital signature values is done by obtaining a 32 bit digital signature value corresponding to each prominent frame. Each block of the pre-defined number of block has a pre-defined number of pixels. The first bit value and the second bit value is calculated from comparison of a mean and a variance for the pre-defined number of pixels in each block of the prominent 10 frame with a corresponding mean and variance for a master frame. The corresponding mean and variance for the master frame is present in the master database. The 32 bit digital signature value is obtained by sequentially arranging the first bit value and the second bit value for each block of the pre-defined number of blocks of the prominent frame. 15
 In an embodiment of the present disclosure, the first bit value and the second bit value are assigned a binary 0 when the mean and the variance for each block of the prominent frame is less the corresponding mean and variance of each master frame. 20
 In another embodiment of the present disclosure, the first bit value and the second bit value are assigned a binary 1 when the mean and the variance for each block of the prominent frame is greater than the corresponding mean and variance of each master frame. 25
 In an embodiment of the present disclosure, the computer-implemented method further includes a detection of the one or more advertisements broadcasted on the channel. The detection of the one or more advertisements includes a supervised detection and an unsupervised detection. 30
Page 7 of 40
 In an embodiment of the present disclosure, the computer-implemented method further includes storage of a generated set of digital signature values, the first set of audio fingerprints and the first set of video fingerprints in a first database and a second database.
5
 In an embodiment of the present disclosure, the computer-implemented method further includes updating of a first metadata comprising the set of digital signature values and the first set of video fingerprints corresponding to a detected advertisement in a master database for an unsupervised detection.
10
 In another aspect, the present disclosure provides a computer program product. The computer program product includes a non-transitory computer readable medium storing a computer readable program. The computer readable program when executed on a computer causes the computer to perform one or more steps. The one or more steps include a step of fetching a set of 15 prominent frames and a pre-defined section of an audio clip corresponding to a detected advertisement. Further, the one or more steps include a step of retrieving a plurality of features corresponding to the set of prominent frames and the pre-defined section of the audio clip. Furthermore, the one or more steps include a step of comparing each of the plurality of features with a corresponding pre-20 defined set of features. Moreover, the one or more steps include a step of tagging the detected advertisement with a unique tag.
 In an embodiment of the present disclosure, the plurality of features include a brand logo displayed in one or more prominent frames of the set of 25 prominent frames. In addition, the plurality of features include a brand tagline displayed in the one or more prominent frames of the set of prominent frames. Moreover, the plurality of features include a brand tagline recited corresponding to the pre-defined section of the audio clip.
30
 In an embodiment of the present disclosure, the tag is a brand name corresponding to the detected advertisement.
Page 8 of 40
 In an embodiment of the present disclosure, the computer-implemented method further includes a detection of the one or more advertisements broadcasted on the channel. The detection of the one or more advertisements includes a supervised detection and an unsupervised detection.
5
 In yet another aspect, the present disclosure provides an auto-tagging system for tagging the one or more advertisements broadcasted on a channel in real time. The auto-tagging system includes a fetching module in a processor. The fetching module fetch a set of prominent frames and a pre-defined section of an audio clip corresponding to a detected advertisement. Further, the auto-10 tagging system includes a retrieving module in the processor. The retrieving module retrieves a plurality of features corresponding to the set of prominent frames and the pre-defined section of the audio clip. Furthermore, the auto-tagging system includes a tagging module in the processor. The tagging module tags the detected advertisement with a unique tag. 15
 In an embodiment of the present disclosure, the plurality of features include a brand logo displayed in one or more prominent frames of the set of prominent frames. In addition, the plurality of features include a brand tagline displayed in the one or more prominent frames of the set of prominent frames. 20 Moreover, the plurality of features include a brand tagline recited corresponding to the pre-defined section of the audio clip.
 In an embodiment of the present disclosure, the auto-tagging system further includes a generation module in the processor. The generation module 25 generates a set of digital signature values. The digital signature values corresponds to an extracted set of video fingerprints. The generation of each digital signature value of the set of digital signature values is done by dividing each prominent frame of the one or more prominent frames into a pre-defined number of blocks. Further, each block of each prominent frame of the one or 30 more prominent frames is gray scaled. Furthermore, the generation of each digital
Page 9 of 40
signature value of the set of digital signature values is done by calculating a first bit value and a second bit value for each block of the prominent frame. In addition, the generation of each digital signature value of the set of digital signature values is done by obtaining a 32 bit digital signature value corresponding to each prominent frame. Each block of the pre-defined number of 5 block has a pre-defined number of pixels. The first bit value and the second bit value is calculated from comparison of a mean and a variance for the pre-defined number of pixels in each block of the prominent frame with a corresponding mean and variance for a master frame. The corresponding mean and variance for the master frame is present in the master database. The 32 bit digital signature value 10 is obtained by sequentially arranging the first bit value and the second bit value for each block of the pre-defined number of blocks of the prominent frame.
 In an embodiment of the present disclosure, the auto-tagging system further includes a storage module in the processor. The storage module stores a 15 generated set of digital signature values, the first set of audio fingerprints and the first set of video fingerprints in a first database and a second database.
 In an embodiment of the present disclosure, the auto-tagging system further includes an updating module in the processor. The updating module 20 updates a first metadata comprising the set of digital signature values and the first set of video fingerprints corresponding to a detected advertisement in a master database for an unsupervised detection.
BRIEF DESCRIPTION OF THE FIGURES 25
 Having thus described the invention in general terms, reference will now be made to the accompanying drawings, which are not necessarily drawn to scale, and wherein:
 FIG. 1A illustrates a system for an auto-tagging of one or more advertisements broadcasted on a channel, in accordance with an embodiment of 30 the present disclosure;
Page 10 of 40
 FIG. 1B illustrates a system for the unsupervised detection and the auto-tagging of the one or more advertisements broadcasted on the channel, in accordance with another embodiment of the present disclosure;
 FIG. 1C illustrates a system for the supervised detection of the one or more advertisements broadcasted on the channel, in accordance with yet 5 another embodiment of the present disclosure;
 FIG. 2 illustrates a block diagram of an auto-tagging system, in accordance with various embodiments of the present disclosure;
 FIG. 3 illustrates a flow chart for the auto-tagging of the one or more advertisements broadcasted on the channel, in accordance with various 10 embodiments of the present disclosure; and
 FIG. 4 illustrates a block diagram of the portable communication device, in accordance with various embodiments of the present disclosure.
 It should be noted that the accompanying figures are intended to present illustrations of exemplary embodiments of the present disclosure. These 15 figures are not intended to limit the scope of the present disclosure. It should also be noted that accompanying figures are not necessarily drawn to scale.
Page 11 of 40
DETAILED DESCRIPTION  Reference will now be made in detail to selected embodiments of the present disclosure in conjunction with accompanying figures. The embodiments described herein are not intended to limit the scope of the disclosure, and the present disclosure should not be construed as limited to the embodiments 5 described. This disclosure may be embodied in different forms without departing from the scope and spirit of the disclosure. It should be understood that the accompanying figures are intended and provided to illustrate embodiments of the disclosure described below and are not necessarily drawn to scale. In the drawings, like numbers refer to like elements throughout, and thicknesses and 10 dimensions of some components may be exaggerated for providing better clarity and ease of understanding.
 It should be noted that the terms "first", "second", and the like, herein do not denote any order, quantity, or importance, but rather are used to distinguish 15 one element from another. Further, the terms "a" and "an" herein do not denote a limitation of quantity, but rather denote the presence of at least one of the referenced item.
 FIG. 1A illustrates a system 100 for an unsupervised and a 20 supervised detection of one or more advertisements broadcasted on a channel, in accordance with an embodiment of the present disclosure. The system 100 describes an environment suitable for an interactive reception and processing of a channel broadcast. The system 100 is configured to provide a setup for detection of the one or more advertisements. Moreover, the system 100 is configured to tag 25 each of the one or more advertisements automatically with a brand name.
 The system 100 includes a broadcast reception device 102, an auto-tagging system 104 and a master database 112. The above stated elements of the system 100 operate coherently and synchronously to detect the one or more 30 advertisements present in media content broadcasted on the channel. In addition,
Page 12 of 40
the above stated elements of the system 100 operate coherently and synchronously to tag each of the one or more advertisements automatically.
 The broadcast reception device 102 is a channel feed receiving and processing device. The broadcast reception device 102 is attached directly or 5 indirectly to a receiving antenna or dish. The receiving antenna receives a broadcasted signal carrying one or more channel feeds. The one or more channel feeds are encoded in a pre-defined format. In addition, the one or more channel feeds have a set of characteristics. The set of characteristics includes a frame rate, an audio sample rate, one or more frequencies and the like. 10
 The broadcasted signal carrying the one or more channel feeds is initially transmitted from a transmission device. In an embodiment of the present disclosure, the broadcasted signal carrying the one or more channel feeds is a multiplexed MPEG-2 encoded signal having a constant bit rate. In another 15 embodiment of the present disclosure, the broadcasted signal carrying the one or more channel feeds is a multiplexed MPEG-2 encoded signal having a variable bit rate. In yet another embodiment of the present disclosure, the broadcasted signal carrying the one or more channel feeds is any digital standard encoded signal. The bit rate is based on complexity of each frame in each of the one or more 20 channel feeds. The quality of the multiplexed MPEG-2 encoded signal will be reduced when the broadcasted signal is too complex to be coded at a constant bit-rate. The bit rate of the variable bit-rate MPEG-2 streams is adjusted dynamically as less bandwidth is needed to encode the images with a given picture quality. In addition, the broadcasted signal is encrypted for a conditional access to a 25 particular subscriber. The encrypted broadcast signal is uniquely decoded by the broadcast reception device 102 uniquely. In an embodiment of the present disclosure, the broadcast reception device 102 receives media content corresponding to the broadcasted content having audio in the pre-defined regional language or the standard language. The media content corresponds to another 30 channel.
Page 13 of 40
 In an example, a digital TV signal is received on the broadcast reception device 102 as a stream of MPEG-2 data. The MPEG-2 data has a transport stream. The transport stream has a data rate of 40 megabits/second for a cable or satellite network. Each transport stream consists of a set of sub-streams. The set of sub-streams is defined as elementary streams. Each elementary stream 5 includes an MPEG-2 encoded audio, an MPEG-2 encoded video and data encapsulated in an MPEG-2 stream. In addition, each elementary stream includes a packet identifier (hereinafter “PID”) that acts as a unique identifier for corresponding elementary stream within the transport stream. The elementary streams are split into packets in order to obtain a packetized elementary stream 10 (hereinafter “PES”).
 In an embodiment of the present disclosure, the broadcast reception device 102 is a digital set top box. In another embodiment of the present disclosure, the broadcast reception device 102 is a hybrid set top box. In yet 15 another embodiment of the present disclosure, the broadcast reception device 102 is an internet protocol television (hereinafter IPTV) set top box. In yet another embodiment of the present disclosure, the broadcast reception device 102 is any standard broadcast signal processing device. Moreover, the broadcast reception device 102 may receive the broadcast signal from any broadcast signal medium. 20
 In an embodiment of the present disclosure, the broadcast signal medium is an ethernet cable. In another embodiment of the present disclosure, the broadcast signal medium is a satellite dish. In yet another embodiment of the present disclosure, the broadcast signal medium is a coaxial cable. In yet another 25 embodiment of the present disclosure, the broadcast signal medium is a telephone line having DSL connection. In yet another embodiment of the present disclosure, the broadcast signal medium is a broadband over power line (hereinafter “BPL”). In yet another embodiment of the present disclosure, the broadcast signal medium is an ordinary VHF or UHF antenna. 30
Page 14 of 40
 The broadcast reception device 102 primarily includes a signal input port, an audio output port, a video output port, a de-multiplexer, a video decoder, an audio decoder and a graphics engine. The broadcast signal carrying the one or more channel feeds is received at the signal input port. The broadcast signal carrying the one or more channel feeds is de-multiplexed by the de-multiplexer. 5 The video decoder decodes the encoded video and the audio decoder decodes the encoded audio. The video and audio corresponds to a channel selected in the broadcast reception device 102. In general, the broadcast reception device 102 carries the one or more channel feeds multiplexed to form a single transporting stream. The broadcast reception device 102 can decode only one channel in real 10 time.
 Further, the decoded audio and the decoded video are received at the audio output port and the video output port. Further, the decoded video has a first set of features. The first set of features includes a frame height, a frame width, a 15 frame rate, a video resolution, an aspect ratio, a bit rate and the like. Moreover, the decoded audio has a second set of features. The second set of features includes a sample rate, a bit rate, a bin size, one or more data points, one or more prominent frequencies and one or more prominent amplitudes. Further, the decoded video may be of any standard quality. In an embodiment of the present 20 disclosure, the decoded video signal is a 144p signal. In another embodiment of the present disclosure, the decoded video signal is a 240p signal. In yet another embodiment of the present disclosure, the decoded video signal is a 360p signal. In yet another embodiment of the present disclosure, the decoded video signal is a 480p signal. In yet another embodiment of the present disclosure, the decoded 25 video signal is a 720p video signal. In yet another embodiment of the present disclosure, the decoded video signal is a 1080p video signal. In yet another embodiment of the present disclosure, the decoded video signal is a 1080i video signal. In yet another embodiment of the present disclosure, the decoded video signal is a 1440p video signal. In yet another embodiment of the present 30
Page 15 of 40
disclosure, the decoded video signal is a 2160p video signal. Here, p and i denotes progressive scan and interlace scan techniques.
 Further, the decoded video and the decoded audio (hereinafter “media content”) are transferred to the auto-tagging system 104 through a transfer 5 medium. The transfer medium can be a wireless medium or a wired medium. Moreover, the media content includes one or more television programs, the one or more advertisements, one or more channel related data, subscription related data, operator messages and the like. The media content has a pre-defined frame rate, a pre-defined number of frames and a pre-defined bit rate for a pre-defined interval 10 of broadcast. In an embodiment of the present disclosure, the media content broadcasted on the channel uses a pre-defined regional language in the audio. In another embodiment of the present disclosure, the media content broadcasted on the channel uses a standard language accepted nationally. Moreover, the auto-tagging system 104 includes a first processing unit 106 and a second processing 15 unit 108. The auto-tagging system 104 has a built in media splitter configured to copy and transmit the media content synchronously to the first processing unit 106 and the second processing unit 108 in the real time. The first processing unit 106 includes a first central processing unit and associated peripherals for unsupervised detection of the one or more advertisements (as shown in FIG. 1B). 20 The first processing unit 106 is connected to a first database 106a.
 The first processing unit 106 is programmed to perform extraction of a first set of audio fingerprints and a first set of video fingerprints corresponding to the media content broadcasted on the channel. The first set of video 25 fingerprints and the first set of audio fingerprints are extracted sequentially in the real time. The extraction of the first set of video fingerprints is done by sequentially extracting one or more prominent fingerprints corresponding to one or more prominent frames present in the media content. The one or more prominent frames correspond to the pre-defined interval of broadcast. 30
Page 16 of 40
 For example, let the media content be related to a channel say, A. The channel A broadcasts a 1 hour reality show between 9 PM to 10 PM. Suppose the media content is broadcasted on the channel A with a frame rate of 25 frames per second (hereinafter “fps”). Again let us assume that the channel A administrator has placed 10 advertisements in between 1 hour broadcast of the 5 reality show. The first processing unit 106 separates audio and video from the media content corresponding to the reality show in the real time. Further, the first processing unit 106 sets a pre-defined range of time to approximate duration of play of every advertisement. Let us suppose the pre-defined range of time is between 12 seconds to 38 seconds. The first processing unit 106 processes each 10 frame of the pre-defined number of frames of the 1 hour long reality show. The first processing unit 106 filters and selects prominent frames having dissimilar scenes. The first processing unit 106 extracts relevant characteristics corresponding to each prominent frame. The relevant characteristics constitute a digital video fingerprint. Similarly, the first processing unit 106 extracts the first 15 set of audio fingerprints corresponding to the media content.
 Furthermore, each of the one or more prominent fingerprints corresponds to a prominent frame having sufficient contrasting features compared to an adjacent prominent frame. For example, let us suppose that the first 20 processing unit 106 select 5 prominent frames per second from 25 frames per second. Each pair of adjacent frames of the 5 prominent frames will have evident contrasting features. The first processing unit 106 generates a set of digital signature values corresponding to an extracted set of video fingerprints. The first processing unit 106 generates each digital signature value of the set of digital 25 signature values by dividing each prominent frame of the one or more prominent frames into a pre-defined number of blocks. In an embodiment of the present disclosure, the predefined number of block is 16 (4X4). In another embodiment of the present disclosure, the pre-defined number of blocks is any suitable number. Each block of the pre-defined number of blocks has a pre-defined 30 number of pixels. Each pixel is fundamentally a combination of red (hereinafter
Page 17 of 40
“R”), green (hereinafter “G”) and blue (hereinafter “B”) colors. The colors are collectively referred to as RGB. Each color of a pixel (RGB) has a pre-defined value in a pre-defined range of values. The predefined range of values is 0-255.
 In an example, the RGB for the pixel has value of 000000. The color 5 of pixel is black. In another example, the RGB for the pixel has a value of FFFFFF (255; 255; 255). The color of the pixel is white. Here, FF is hexadecimal equivalent of decimal, 255. In yet another example, the RGB for the pixel has a value of FF0000 (255, 0, 0). The color of the pixel is red. In yet another example, the RGB for the pixel has a value of 0000FF (0, 0, 255). The 10 color of the pixel is blue. In yet another example, the RGB for the pixel has a value of 008000 (0, 128, 0). The color of the pixel is green.
 The first processing unit 106 gray-scales each block of each prominent frame of the one or more prominent frames. The gray-scaling of each 15 block is a conversion of RGB to monochromatic shades of grey color. Here 0 represents black and 255 represents white. Further, the first processing unit 106 calculates a first bit value and a second bit value for each block of the prominent frame. The first bit value and the second bit value are calculated from comparing a mean and a variance for the pre-defined number of pixels in each block of the 20 prominent frame with a corresponding mean and variance for a master frame in the master database 112. The first processing unit 106 assigns the first bit value and the second bit with a binary 0 when the mean and the variance for each block of the prominent frame is less the corresponding mean and variance of each master frame. The first processing unit 106 assigns the first bit value and the 25 second bit value with a binary 1 when the mean and the variance for each block is greater than the corresponding mean and variance of each master frame.
 Furthermore, the first processing unit 106 obtains a 32 bit digital signature value corresponding to each prominent frame. The 32 bit digital 30 signature value is obtained by sequentially arranging the first bit value and the
Page 18 of 40
second bit value for each block of the pre-defined number of blocks of the prominent frame. The first processing unit 106 stores each digital signature value corresponding to each prominent frame of the one or more prominent frames in the first database 106a. The digital signature value corresponds to the one or more programs and the one or more advertisements. The first processing unit 106 5 utilizes a temporal recurrence algorithm to detect the one or more advertisements. In temporal recurrence algorithm, the first processing unit 106 probabilistically matches a first pre-defined number of digital signature values with a stored set of digital signature values present in the first database 106a.
10
 In an example, let us suppose that the first processing unit 106 generates 100 digital signature values corresponding to 100 prominent frames in the first database 106a. The first processing unit 106 probabilistically matches 20 digital signature values corresponding to 101st to 121st prominent frame with each 20 digital signature values corresponding to 100 previously stored prominent 15 frames.
 The probabilistic match of the first pre-defined number of digital signature values sequentially for each of the prominent frame is performed by utilizing a sliding window algorithm. In an embodiment of the present disclosure, 20 the first pre-defined number of digital signature values of the set of digital signature values for the unsupervised detection of the one or more advertisements is 20. The first processing unit 106 determines a positive probabilistic match of the pre-defined number of prominent frames based on a pre-defined condition. The pre-defined condition includes a pre-defined range of positive matches 25 corresponding to probabilistically match digital signature values and a pre-defined duration of media content corresponding to the positive match. In addition, the pre-defined condition includes a sequence and an order of the positive matches and a degree of match of a pre-defined range of number of bits of the first pre-defined number of signature values. In an embodiment of the present disclosure, 30 the pre-defined range of probabilistic matches corresponding to the positive match lies in a range of 40 matches to 300 matches. In another embodiment of the
Page 19 of 40
present disclosure, the pre-defined range of probabilistic matches corresponding to the positive match lies in a suitable duration of each advertisement running time. In an embodiment of the present disclosure, the first processing unit 106 discards the probabilistic matches corresponding to less than 40 positive matches.
5
 Further, the pre-defined duration of media content corresponding to the positive match has a first limiting duration bounded by a second limiting duration. In an embodiment of the present disclosure, the first limiting duration is 10 seconds and the second limiting duration is 25 seconds. In another embodiment of the present disclosure, the first limiting duration is 10 seconds and 10 the second limiting duration is 35 seconds. In yet another embodiment of the present disclosure, the first limiting duration is 10 seconds and the second limiting duration is 60 seconds. In yet another embodiment of the present disclosure, the first limiting duration is 10 seconds and the second limiting duration is 90 seconds. In yet another embodiment of the present disclosure, the first limiting 15 duration and the second limiting duration may have any suitable limiting durations.
 In an example, suppose 100 digital signature values from 1000th prominent frame to 1100th prominent frame gives a positive match with a stored 20 100th frame to 200th frame in the first database 106a. The first processing unit 106 checks whether the number of positive matches is in the pre-defined range of positive matches. In addition, the first processing unit 106 checks whether the positive matches correspond to media content is in the first limiting duration and the second limiting duration. Moreover, the first processing unit 106 checks 25 whether the positive matches of 100 digital signature values for unsupervised detection of the one or more advertisements is in a required sequence and order.
 The first processing unit 106 checks for the degree of match of the pre-defined range of number of bits of the first pre-defined number of signature 30 values. In an example, the degree of match of 640 bits (32 Bits X 20 digital signature values) of the generated set of digital signature values with stored 640
Page 20 of 40
digital signature values is 620 bits. In such case, the first processing unit 106 flags the probabilistic match as the positive match. In another example, the degree of match of 640 bits of the generated set of digital signature values with stored 640 digital signature values is 550 bits. In such case, the first processing unit 106 flags the probabilistic match as the negative match. In an embodiment of 5 the present disclosure, the pre-defined range of number of bits is 0-40.
 The first processing unit 106 generates one or more prominent frequencies and one or more prominent amplitudes from extracted first set of audio fingerprints. The first processing unit 106 fetches a sample rate of first set 10 of audio fingerprints. The sample rate is divided by a pre-defined bin size set for the audio. The division of the sample rate by the pre-defined bin size provides the data point. Further, the first processing unit 106 performs fast fourier transform (hereinafter “FFT”) on each bin size of the audio to obtain the one or more prominent frequencies and the one or more prominent amplitudes. The first 15 processing unit 106 compares the one or more prominent frequencies and the one or more prominent amplitudes with a stored one or more prominent frequencies and a stored one or more prominent amplitudes.
 Going further, the first processing unit 106 fetches the corresponding 20 video and audio clip associated to the probabilistically matched digital signature values. The first database 106a and the first processing unit 106 are associated with an auto-tagging system 104. Furthermore, the auto-tagging system 104 retrieves a plurality of features associated with the video clip and the audio clip of the corresponding advertisement. Further, the auto-tagging system 104 compares 25 the plurality of features with a pre-defined set of features. In an embodiment of the present disclosure, the pre-defined set of features are stored in a reference database 110.
 In an embodiment of the present disclosure, the plurality of features 30 include a brand logo displayed in one or more prominent frames of the set of prominent frames. In another embodiment of the present disclosure, the plurality
Page 21 of 40
of features include a brand tagline displayed in the one or more prominent frames of the set of prominent frames. In yet another embodiment of the present disclosure, the plurality of features include a brand tagline recited corresponding to the pre-defined section of the audio clip. Moreover, the auto-tagging system 104 decides whether the audio clip and the video clip correspond to a new 5 advertisement. Further, the auto-tagging system 104 tags each audio clip and the video clip with a unique tag. In an embodiment of the present disclosure, the unique tag is a brand name associated with a detected advertisement.
 In an embodiment of the present disclosure, the auto-tagging system 10 104 tags each audio clip through a brand tagline identification in the real time. In an example of a product B, while advertising, an actor recites a unique tagline of the product B. The auto-tagging system 104 automatically converts the unique tagline of the product B recited by the actor into the corresponding text by a speech-to-text analysis. The auto-tagging system 104 compares the text 15 associated with the unique tagline of the product B with the pre-defined taglines present in the reference database 110. The auto-tagging system 104 tags the advertisement with the corresponding brand name of the product B.
 In an embodiment of the present disclosure, the auto-tagging system 20 104 tags each video clip through a brand logo identification in the real time. In another embodiment of the present disclosure, the auto-tagging system 104 tags each video clip through a brand tagline identification in the real time. In yet another embodiment of the present disclosure, the auto-tagging system 104 tags each video clip through any suitable brand element identification in the real time. 25
 In an example of an advertisement of a product C, the brand logo is displayed during the streaming of the advertisement associated with the product C in the real time. The auto-tagging system 104 retrieves the one or more prominent frames containing the brand logo associated with the product C. In 30 addition, the auto-tagging system 104 compares the brand logo of the product C with a plurality of brand logo stored in the reference database 110. Moreover, the
Page 22 of 40
auto-tagging system 104 tags the advertisement with the corresponding brand name of the product C. In another example of the advertisement of product D, the brand tagline is displayed in the video clip of the advertisement associated with the product D in the real time. The auto-tagging system 104 fetches the one or more prominent frames containing the brand tagline associated with the 5 advertisement of the product D. Furthermore, the auto-tagging system 104 retrieves the brand tagline associated with the product D and co-relates the brand tagline of the product D with a plurality of brand taglines stored in the reference database 110. Moreover, the auto-tagging system 104 tags the advertisement with the corresponding brand name of the product D. 10
 In an embodiment of the present disclosure, the first processing unit 106 extracts the first set of audio fingerprints and the first set of video fingerprints corresponding to another channel. The first processing unit 106 extracts the pre-defined number of prominent frames and generates pre-defined number of digital 15 signature values. The first processing unit 106 performs the temporal recurrence algorithm to detect a new advertisement. In an embodiment of the present disclosure, the first processing unit 106 generates prominent frequencies and prominent amplitudes of the audio. In another embodiment of the present disclosure, the first processing unit 106 discards the audio from the media content. 20 In an embodiment of the present disclosure, the first processing unit 106 probabilistically matches the one or more prominent frequencies and the one or more prominent amplitudes with stored prominent frequencies and stored prominent amplitudes in the first database. The stored prominent frequencies and the stored prominent amplitudes correspond to a regional channel having audio in 25 the pre-defined regional language or standard language. In an embodiment of the present disclosure, the standard language is English. In another embodiment of the present disclosure, the first processing unit 106 gives precedence to results of probabilistic match of video fingerprints than to the audio fingerprints. Moreover, the auto-tagging system 104 automatically tags the detected advertisement 30 broadcasted in the pre-defined regional language or the standard language.
Page 23 of 40
 Further, the auto-tagging system 104 stores the plurality of digital fingerprints of the advertisement for determining one or more advertisements associated with a corresponding product. In an example, a product E may have one or more advertisements. Each of the one or more advertisement associated 5 with the product E may have different duration and fingerprints. The auto-tagging system 104 compares the stored fingerprints of each advertisement of the product E. In addition, the auto-tagging system 104 determines the difference in the fingerprints associated with each advertisements of the product E. Simultaneously, the auto-tagging system 104 compares the brand logo and the 10 brand tagline of the product E with the brand logo and the brand tagline stored in the reference database 110. The auto-tagging system 104 treats the advertisements as the new advertisement of the product E after obtaining positive match results.
15
 Going further, the first processing unit 106 reports a positively matched digital signature values corresponding to each detected advertisement in a reporting database present in the first database 106a. The first processing unit 106 discards any detected advertisement already reported in the reporting database. 20
 The second processing unit 108 includes a second central processing unit and associated peripherals for supervised detection of the one or more advertisements (also shown in FIG. 1C). The second processing unit 106 is connected to a second database 108a. The second processing unit 108 is 25 programmed to perform the extraction of the first set of audio fingerprints and the first set of video fingerprints corresponding to the media content broadcasted on the channel. The first set of video fingerprints and the first set of audio fingerprints are extracted sequentially in the real time. The extraction of the first set of video fingerprints is done by sequentially extracting the one or more 30 prominent fingerprints corresponding to the one or more prominent frames for the pre-defined interval of broadcast.
Page 24 of 40
 Furthermore, each of the one or more prominent fingerprints corresponds to the prominent frame having sufficient contrasting features compared to the adjacent prominent frame. For example, let us suppose that the second processing unit 108 selects 6 prominent frames per second from 25 frames 5 per second. Each pair of adjacent frames of the 6 prominent frames will have evident contrasting features. The second processing unit 108 generates the set of digital signature values corresponding to the extracted set of video fingerprints. The second processing unit 108 generates each digital signature value of the set of digital signature values by dividing each prominent frame of the one or more 10 prominent frames into the pre-defined number of blocks. In an embodiment of the present disclosure, the predefined number of block is 15 (4X4). In another embodiment of the present disclosure, the pre-defined number of blocks is any suitable number. Each block of the pre-defined number of blocks has the pre-defined number of pixels. Each pixel is fundamentally the combination of R, G 15 and B colors. The colors are collectively referred to as RGB. Each color of the pixel (RGB) has the pre-defined value in the pre-defined range of values. The predefined range of values is 0-255.
 The second processing unit 108 gray-scales each block of each 20 prominent frame of the one or more prominent frames. The second processing unit 108 calculates the first bit value and the second bit value for each block of the prominent frame. The first bit value and the second bit value are calculated from comparison of the mean and the variance for the pre-defined number of pixels with the corresponding mean and variance for the master frame. The master 25 frame is present in the master database 112. The second processing unit 108 assigns the first bit value and the second bit with the binary 0 when the mean and the variance for each block is less the corresponding mean and variance of each master frame. The second processing unit 108 assigns the first bit value and the second bit value with the binary 1 when the mean and the variance for each block 30 is greater than the corresponding mean and variance of each master frame.
Page 25 of 40
 The second processing unit 108 obtains the 32 bit digital signature value corresponding to each prominent frame. The 32 bit digital signature value is obtained by sequentially arranging the first bit value and the second bit value for each block of the pre-defined number of blocks of the prominent frame. The 5 second processing unit 108 stores each digital signature value corresponding to each prominent frame of the one or more prominent frames in the second database 108a. The digital signature value corresponds to the one or more programs and the one or more advertisements.
10
 The second processing unit 108 performs the supervised detection of the one or more advertisements. The second processing unit 108 probabilistically matches a second pre-defined number of digital signature values with the stored set of digital signature values present in the master database 112. The second pre-defined number of digital signature values corresponds to the second pre-defined 15 number of prominent frames of the real time broadcasted media content. The probabilistic match is performed for the set of digital signature values by utilizing a sliding window algorithm. The second processing unit 108 determines the positive match in the probabilistically matching of the second pre-defined number of digital signature values with the stored set of digital signature values. The 20 stored set of digital signature values is present in the master database 112. In an embodiment of the present disclosure, the second pre-defined number of digital signature values of the set of digital signature values for the supervised detection of the one or more advertisements is 6. In another embodiment of the present disclosure, the second pre-defined number of digital signature values is selected 25 based on optimal processing capacity and performance of the second processing unit 108.
 In an example, let us suppose that the second processing unit 108 stores 300 digital signature values corresponding to 300 prominent frames in the 30 second database 108a for 10 seconds of the media content. The second processing unit 108 probabilistically matches 6 digital signature values
Page 26 of 40
corresponding to 101st to 107nth prominent frame with each 6 digital signature values corresponding to 300 previously stored prominent frames. The 300 previously stored prominent frames are present in the master database 112.
 In another example, suppose 300 digital signature values from 500th 5 prominent frame to 800th prominent frame gives a positive match with a stored 150th frame to 450th frame in the master database 112. The second processing unit 108 checks whether the number of positive matches is in the pre-defined range of positive matches and the positive matches correspond to media content in the first limiting duration and the second limiting duration. In addition, the 10 second processing unit 108 checks whether the positive matches of 300 digital signature values for supervised detection of the one or more advertisements is in the required sequence and order.
 The second processing unit 108 checks for the degree of match of the 15 pre-defined range of number of bits of the second pre-defined number of signature values. In an example, the degree of match of 192 bits of the generated set of digital signature values with stored 192 digital signature values is 185 bits. In such case, the second processing unit 108 flags the probabilistic match as the positive match. In another example, the degree of match of 192 bits of the 20 generated set of digital signature values with stored 192 digital signature values is 179 bits. In such case, the second processing unit 108 flags the probabilistic match as the negative match. In an embodiment of the present disclosure, the pre-defined range of number of bits is 0-12.
25
 The second processing unit 108 compares the one or more prominent frequencies and the one or more prominent amplitudes with the stored one or more prominent frequencies and the stored one or more prominent amplitudes. The one or more prominent frequencies and the one or more prominent amplitudes corresponding to the extracted first set of audio fingerprints. In an 30 embodiment of the present disclosure, the auto-tagging system 104 automatically checks whether each supervised advertisement detected is an advertisement or a
Page 27 of 40
program. In an embodiment of the present disclosure, the auto-tagging system 104 reports a frequency of each advertisement broadcasted for a first time and a frequency of each advertisement broadcasted repetitively.
 Further, the master database 112 is present in a master server. The 5 master database 112 includes a plurality of digital video and audio fingerprint records and every signature value corresponding to each previously detected and newly detected advertisement. The master database 112 is connected to the auto-tagging system 104. In an embodiment of the present disclosure, the master server is present in a remote location. In another embodiment of the present 10 disclosure, the master server is present locally with the auto-tagging system 104.
 In an embodiment of the present disclosure, the second processing unit 108 extracts the first set of audio fingerprints and the first set of video fingerprints corresponding to another channel. The second processing unit 108 15 extracts the pre-defined number of prominent frames and generates pre-defined number of digital signature values. The second processing unit 108 performs probabilistic matching of digital signature values corresponding to the video with the stored digital signature values in the master database detect a repeated advertisement. In an embodiment of the present disclosure, the second processing 20 unit 108 generates the one or more prominent frequencies and the one or more prominent amplitudes of the audio. In another embodiment of the present disclosure, the second processing unit 108 discards the audio from the media content. In an embodiment of the present disclosure, the master database 112 includes the one or more advertisements corresponding to a same advertisement 25 in every regional language. In another embodiment of the present disclosure, the master database 112 includes the advertisement in a specific national language. In embodiment of the present disclosure, the second processing unit 108 probabilistically matches the one or more prominent frequencies and the one or more prominent amplitudes with stored prominent frequencies and stored 30 prominent amplitudes. The stored prominent frequencies and the stored prominent amplitudes correspond to a regional channel having audio in the pre-
Page 28 of 40
defined regional language or standard language in the master database 112. In an embodiment of the present disclosure, the standard language is English. In another embodiment of the present disclosure, the second processing unit 108 gives precedence to results of probabilistic match of video fingerprints than to the audio fingerprints. 5
 Further, the auto-tagging system 104 stores the generated set of digital signature values, the first set of audio fingerprints and the first set of video fingerprints in the first database 106a and the second database 108a. Furthermore, the auto-tagging system 104 updates the first metadata manually in the master 10 database 112 for the unsupervised detection of the one or more advertisements. The first metadata includes the set of digital signature values and the first set of video fingerprints.
 It may be noted that in FIG. 1A, FIG. 1B and FIG. 1C, the system 15 100 includes the broadcast reception device 102 for decoding one channel; however, those skilled in the art would appreciate the system 100 includes more number of broadcast reception devices for decoding more number of channels. It may be noted that in FIG. 1A, FIG. 1B and FIG. 1C, the system 100 includes the auto-tagging system 104 for the supervised and the unsupervised detection of the 20 one or more advertisement corresponding to one channel; however, those skilled in the art would appreciate that the auto-tagging system 104 detects the one or more advertisements corresponding to more number of channels.
 FIG. 2 illustrates a block diagram 200 of the auto-tagging system 25 104, in accordance with various embodiments of the present disclosure. The block diagram 200 describes the auto-tagging system 104 configured for the unsupervised and the supervised detection of the one or more advertisements.
 The block diagram 200 of the auto-tagging system 104 includes an 30 extraction module 202, a generation module 204, a storage module 206, a detection module 208, a fetching module 210 and a retrieving module 212. In
Page 29 of 40
addition, the auto-tagging system 104 includes a comparison module 214, a tagging module and an updating module 214. The extraction module 202 extracts the first set of audio fingerprints and the first set of video fingerprints corresponding to the media content broadcasted on the channel. The first set of audio fingerprints and the first set of video fingerprints are extracted sequentially 5 in the real time (as described above in detailed description of FIG. 1A, FIG. 1B and FIG. 1C).
 Further, the generation module 204 generates the set of digital signature values corresponding to the extracted set of video fingerprints. The 10 generation module 204 generates each digital signature value of the set of digital signature values by dividing and grayscaling each prominent frame into the pre-defined number of blocks. Further, the generation module 204 calculates and obtains each digital signature value corresponding to each block of the prominent frame (as discussed above in the detailed description of FIG. 1A, FIG. 1B and 15 FIG. 1C). The generation module 204 includes a dividing module 204a, a grayscaling module 204b, a calculation module 204c and an obtaining module 204d. The dividing module 204a divides each prominent frame of the one or more prominent frames into the pre-defined number of blocks (as discussed above in the detailed description of FIG. 1A). The grayscaling module 204b grayscales 20 each block of each prominent frame of the one or more prominent frames. The calculation module 204c calculates the first bit value and the second bit value for each block of the prominent frame (as described above in the detailed description of FIG. 1A). The obtaining module 204d obtains the 32 bit digital signature value corresponding to each prominent frame (as described above in detailed 25 description of FIG. 1A, FIG. 1B and FIG. 1C).
 The storage module 206 stores the generated set of digital signature values, the first set of audio fingerprints and the first set of video fingerprints in the first database 106a and the second database 108a (as described above in 30 detailed description of FIG. 1A, FIG. 1B and FIG. 1C). Further, the detection
Page 30 of 40
module 208 detects the one or more advertisements broadcasted on the channel. The detection module 208 includes an unsupervised detection module 208a and the supervised detection module 208b. The unsupervised detection module 208a detects the new advertisement through unsupervised machine learning (as discussed in the detailed description of FIG. 1A, FIG. 1B and FIG. 1C). 5 Moreover, the supervised detection module 208b detects the advertisements broadcasted previously during the broadcasting of the media content (as described above in the detailed description of FIG. 1A, FIG. 1B and FIG. 1C).
 The fetching module 210 fetches the set of prominent frames and the 10 pre-defined section of the audio clip. The set of prominent frames and the pre-defined section of the audio clip corresponds to the detected advertisement (as discussed above in the detailed description of FIG. 1A, FIG. 1B and FIG. 1C). Further, the retrieving module 212 retrieves the plurality of features. The plurality of features corresponds to the set of prominent frames and the pre-defined section 15 of the audio clip (as discussed above in the detailed description of FIG. 1A, FIG. 1B and FIG. 1C).
 Going further, the comparison module 214 compares each of the plurality of features with the corresponding pre-defined set of features. In 20 addition, the pre-defined set of features are stored in the reference database (as described above in the detailed description of FIG. 1A, FIG. 1B and FIG. 1C). Further, the tagging module 216 tags the detected advertisement with the unique tag. The unique tag is the brand name associated with the detected advertisement (as discussed above in the detailed description of FIG. 1A, FIG. 1B and FIG. 1C). 25 Furthermore, the updating module 218 updates the first metadata manually in the master database 112 for the unsupervised detection of the one or more advertisements. The first metadata includes the set of digital signature values and the first set of video fingerprints corresponding to the detected advertisement (as described in the detailed description of FIG. 1A). 30
Page 31 of 40
 FIG. 3 illustrates a flow chart 300 for auto-tagging the one or more advertisements broadcasted on the channel, in accordance with various embodiments of the present disclosure. It may be noted that to explain the process steps of the flowchart 300, references will be made to the interactive messaging system elements of the FIG. 1A, FIG. 1B, FIG. 1C and FIG. 2. 5
 The flowchart 300 initiates at step 302. At step 304, the fetching module 210 fetches the set of prominent frames and the pre-defined section of an audio clip corresponding to the detected advertisement. At step 306, the retrieving module 212 retrieves the plurality of features corresponding to the set 10 of prominent frames and the pre-defined section of the audio clip. At step 308, the comparison module 214 compares each of the plurality of features with the corresponding pre-defined set of features. At step 310, the tagging module 216 tags the detected advertisement.
15
 It may be noted that the flowchart 300 is explained to have above stated process steps; however, those skilled in the art would appreciate that the flowchart 300 may have more/less number of process steps which may enable all the above stated embodiments of the present disclosure.
20
 FIG. 4 illustrates a block diagram of a communication device 400, in accordance with various embodiments of the present disclosure. The communication device 400 enables host process of the auto-tagging system 104. The communication device 400 includes a control circuitry module 402, a storage module 404, an input/output circuitry module 406, and a communication circuitry 25 module 408. The communication device 400 includes any suitable type of portable electronic device. The communication device 400 includes but may not be limited to a personal e-mail device (e.g., a Blackberry.TM. made available by Research in Motion of Waterloo, Ontario), a personal data assistant ("PDA"), a cellular telephone. In addition, the communication device 400 includes a 30 smartphone, the laptop, computer and the tablet. In another embodiment of the present disclosure, the communication device 400 can be a desktop computer.
Page 32 of 40
 From the perspective of this disclosure, the control circuitry module 402 includes any processing circuitry or processor operative to control the operations and performance of the communication device 400. For example, the control circuitry module 402 may be used to run operating system applications, 5 firmware applications, media playback applications, media editing applications, or any other application.
 In an embodiment of the present disclosure, the control circuitry module 402 drives a display and process inputs received from the user interface. 10 From the perspective of this disclosure, the storage module 404 includes one or more storage mediums. The one or more storage medium includes a hard-drive, solid state drive, flash memory, permanent memory such as ROM, any other suitable type of storage component, or any combination thereof. The storage module 404 may store, for example, media data (e.g., music and video files), 15 application data (e.g., for implementing functions on the communication device 400).
 From the perspective of this disclosure, the I/O circuitry module 406 may be operative to convert (and encode/decode, if necessary) analog signals and 20 other signals into digital data. In an embodiment of the present disclosure, the I/O circuitry module 406 may convert the digital data into any other type of signal and vice-versa. For example, the I/O circuitry module 406 may receive and convert physical contact inputs (e.g., from a multi-touch screen), physical movements (e.g., from a mouse or sensor), analog audio signals (e.g., from a microphone), or 25 any other input. The digital data may be provided to and received from the control circuitry module 402, the storage module 404, or any other component of the communication device 400.
 It may be noted that the I/O circuitry module 406 is illustrated in FIG. 30 4 as a single component of the communication device 400; however those skilled
Page 33 of 40
in the art would appreciate that several instances of the I/O circuitry module 406 may be included in the communication device 400.
 The communication device 400 may include any suitable interface or component for allowing the user to provide inputs to the I/O circuitry module 406. 5 The communication device 400 may include any suitable input mechanism. Examples of the input mechanism include but may not be limited to a button, keypad, dial, a click wheel, and a touch screen. In an embodiment, the communication device 400 may include a capacitive sensing mechanism, or a multi-touch capacitive sensing mechanism. 10
 In an embodiment of the present disclosure, the communication device 400 may include specialized output circuitry associated with output devices such as, for example, one or more audio outputs. The audio output may include one or more speakers built into the communication device 400, or an 15 audio component that may be remotely coupled to the communication device 400.
 The one or more speakers can be mono speakers, stereo speakers, or a combination of both. The audio component can be a headset, headphones or ear buds that may be coupled to the communication device 400 with a wire or 20 wirelessly.
 In an embodiment, the I/O circuitry module 406 may include display circuitry for providing a display visible to a user. For example, the display circuitry may include a screen (e.g., an LCD screen) that is incorporated in the 25 communication device 400.
 The display circuitry may include a movable display or a projecting system for providing a display of content on a surface remote from the communication device 400 (e.g., a video projector). In an embodiment of the 30 present disclosure, the display circuitry may include a coder/decoder to convert digital media data into the analog signals. For example, the display circuitry may include video Codecs, audio Codecs, or any other suitable type of Codec.
Page 34 of 40
 The display circuitry may include display driver circuitry, circuitry for driving display drivers or both. The display circuitry may be operative to display content. The display content can include media playback information, application screens for applications implemented on the electronic device, 5 information regarding ongoing communications operations, information regarding incoming communications requests, or device operation screens under the direction of the control circuitry module 402. Alternatively, the display circuitry may be operative to provide instructions to a remote display.
10
 In addition, the communication device 400 includes the communication circuitry module 408. The communication circuitry module 408 may include any suitable communication circuitry operative to connect to a communication network. In addition, the communication circuitry module 408 may include any suitable communication circuitry to transmit communications 15 (e.g., voice or data) from the communication device 400 to other devices. The other devices exist within the communications network. The communications circuitry 408 may be operative to interface with the communication network through any suitable communication protocol. Examples of the communication protocol include but may not be limited to Wi-Fi, Bluetooth RTM, radio 20 frequency systems, infrared, LTE, GSM, GSM plus EDGE, CDMA, and quadband.
 In an embodiment, the communications circuitry module 408 may be operative to create a communications network using any suitable communications 25 protocol. For example, the communication circuitry module 408 may create a short-range communication network using a short-range communications protocol to connect to other devices. For example, the communication circuitry module 408 may be operative to create a local communication network using the Bluetooth, RTM protocol to couple the communication device 400 with a 30 Bluetooth, RTM headset.
Page 35 of 40
 It may be noted that the computing device is shown to have only one communication operation; however, those skilled in the art would appreciate that the communication device 400 may include one more instances of the communication circuitry module 408 for simultaneously performing several communication operations using different communication networks. For 5 example, the communication device 400 may include a first instance of the communication circuitry module 408 for communicating over a cellular network, and a second instance of the communication circuitry module 408 for communicating over Wi-Fi or using Bluetooth RTM.
10
 In an embodiment of the present disclosure, the same instance of the communications circuitry module 408 may be operative to provide for communications over several communication networks. In another embodiment of the present disclosure, the communication device 400 may be coupled to a host device for data transfers and sync of the communication device 400. In addition, 15 the communication device 400 may be coupled to software or firmware updates to provide performance information to a remote source (e.g., to providing riding characteristics to a remote server) or performing any other suitable operation that may require the communication device 400 to be coupled to the host device. Several computing devices may be coupled to a single host device using the host 20 device as a server. Alternatively or additionally, the communication device 400 may be coupled to the several host devices (e.g., for each of the plurality of the host devices to serve as a backup for data stored in the communication device 400).
25
 The present disclosure has numerous disadvantages over the prior art. The present disclosure provides a novel method to detect any new advertisement running for the first time on any television channel. The advertisements are detected robustly and dedicated supervised and unsupervised central processing unit (hereinafter “CPU”) are installed. Further, the present disclosure provides a 30 method and system that is economic and provides high return of investment. The detection of each repeated advertisement on supervised CPU and each new
Page 36 of 40
advertisement on unsupervised CPU significantly saves processing power and saves significant time. The disclosure provides a cost efficient solution to a scaled mapping and database for advertisement broadcast.
 The foregoing descriptions of specific embodiments of the present 5 technology have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the present technology to the precise forms disclosed, and obviously many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the present technology and its practical 10 application, to thereby enable others skilled in the art to best utilize the present technology and various embodiments with various modifications as are suited to the particular use contemplated. It is understood that various omissions and substitutions of equivalents are contemplated as circumstance may suggest or render expedient, but such are intended to cover the application or implementation 15 without departing from the spirit or scope of the claims of the present technology.
 While several possible embodiments of the invention have been described above and illustrated in some cases, it should be interpreted and understood as to have been presented only by way of illustration and example, but 20 not by limitation. Thus, the breadth and scope of a preferred embodiment should not be limited by any of the above-described exemplary embodiments.
Page 37 of 40
CLAIMS
We claim:
1. A computer-implemented method for an automated tagging of one or more advertisements broadcasted on a channel in real time, the computer-implemented method comprising: 5
fetching, with a processor, a set of prominent frames and a pre-defined section of an audio clip corresponding to a detected advertisement;
retrieving, with the processor, a plurality of features corresponding to the set of prominent frames and the pre-defined section of the audio clip; 10
comparing, with the processor, each of the plurality of features with corresponding pre-defined set of features; and
tagging, with the processor, the detected advertisement with a tag.
2. The computer-implemented method as recited in claim 1, wherein the 15 plurality of features comprises a brand logo displayed in one or more prominent frames of the set of prominent frames, a brand tagline displayed in the one or more prominent frames of the set of prominent frames and a brand tagline recited corresponding to the pre-defined section of the audio clip. 20
3. The computer-implemented method as recited in claim 1, wherein the pre-defined set of features being stored in a reference database.
4. The computer-implemented method as recited in claim 1, wherein the tag 25 being a brand name corresponding to a detected advertisement.
5. The computer-implemented method as recited in claim 1, further comprising extracting, with the processor, a first set of audio fingerprints and a first set of video fingerprints corresponding to a media content 30 broadcasting on the channel, wherein the first set of audio fingerprints and the first set of video fingerprints being extracted sequentially in the real
Page 38 of 40
time, wherein the extraction of the first set of video fingerprints being done by sequentially extracting one or more prominent fingerprints corresponding to one or more prominent frames of a pre-defined number of frames present in the media content for a pre-defined interval of broadcast. 5
6. The computer-implemented method as recited in claim 1, further comprising generating, with the processor, a set of digital signature values corresponding to the extracted set of video fingerprints, wherein the generation of each digital signature value of the set of digital signature 10 values being done by: dividing each prominent frame of the one or more prominent frames into a pre-defined number of blocks, wherein each block of the pre-defined number of blocks having a pre-defined number of pixels; grayscaling each block of each prominent frame of the one or more 15 prominent frames; calculating a first bit value and a second bit value for each block of the prominent frame, wherein the first bit value and the second bit value being calculated from comparing a mean and a variance for the pre-defined number of pixels in each block of the prominent frame with a 20 corresponding mean and variance for a master frame in a master database; and obtaining a 32 bit digital signature value corresponding to each prominent frame, wherein the 32 bit digital signature value being obtained by sequentially arranging the first bit value and the second bit value for 25 each block of the pre-defined number of blocks of the prominent frame. 7. The computer-implemented method as recited in claim 6, wherein the first bit value and the second bit value being assigned a binary 0 when the mean and the variance for each block of the prominent frame being less 30 the corresponding mean and variance of each master frame.
Page 39 of 40
8. The computer-implemented method as recited in claim 6, wherein the first bit value and the second bit value being assigned a binary 1 when the mean and the variance for each block of the prominent frame being greater than the corresponding mean and variance of each master frame. 5
9. The computer-implemented method as recited in claim 1, further comprising detecting, with the processor, the one or more advertisements broadcasted on the channel, wherein the detection of the one or more advertisements being a supervised detection and an unsupervised 10 detection.
10. The computer-implemented method as recited in claim 1, further comprising storing, with the processor, the generated set of digital signature values, the first set of audio fingerprints and the first set of video 15 fingerprints in a first database and a second database.
11. The computer-implemented method as recited in claim 1, further comprising, updating, with the processor, a first metadata comprising the set of digital signature values and the first set of video fingerprints 20 corresponding to a detected advertisement in the master database for the unsupervised detection.

Documents

Application Documents

# Name Date
1 201611008282-REQUEST FOR CERTIFIED COPY [10-08-2018(online)].pdf 2018-08-10
1 Form 5 [09-03-2016(online)].pdf 2016-03-09
2 Form 3 [09-03-2016(online)].pdf 2016-03-09
2 201611008282-Correspondence-161117.pdf 2017-11-24
3 Drawing [09-03-2016(online)].pdf 2016-03-09
3 201611008282-OTHERS-161117.pdf 2017-11-24
4 201611008282-Proof of Right (MANDATORY) [15-11-2017(online)].pdf 2017-11-15
4 Description(Complete) [09-03-2016(online)].pdf 2016-03-09
5 Form 26 [11-04-2017(online)].pdf 2017-04-11
5 abstract.jpg 2016-07-14
6 201611008282-GPA-(19-07-2016).pdf 2016-07-19
6 201611008282-Correspondence Others-(19-07-2016).pdf 2016-07-19
7 201611008282-Form-1-(19-07-2016).pdf 2016-07-19
8 201611008282-GPA-(19-07-2016).pdf 2016-07-19
8 201611008282-Correspondence Others-(19-07-2016).pdf 2016-07-19
9 Form 26 [11-04-2017(online)].pdf 2017-04-11
9 abstract.jpg 2016-07-14
10 201611008282-Proof of Right (MANDATORY) [15-11-2017(online)].pdf 2017-11-15
10 Description(Complete) [09-03-2016(online)].pdf 2016-03-09
11 201611008282-OTHERS-161117.pdf 2017-11-24
11 Drawing [09-03-2016(online)].pdf 2016-03-09
12 Form 3 [09-03-2016(online)].pdf 2016-03-09
12 201611008282-Correspondence-161117.pdf 2017-11-24
13 Form 5 [09-03-2016(online)].pdf 2016-03-09
13 201611008282-REQUEST FOR CERTIFIED COPY [10-08-2018(online)].pdf 2018-08-10