Sign In to Follow Application
View All Documents & Correspondence

Real Time Traffic Detection

Abstract: Systems and methods for real-time traffic detection are described. In one embodiment, the method comprises capturing ambient sounds as an audio sample in a user device, and segmenting the audio sample into a plurality of audio frames. Further, the method comprises identifying periodic frames amongst the plurality of audio frames. Spectral features of the identified periodic frames are extracted, and horn sounds are identified based on the spectral features. The identified horn sounds are then used for real-time traffic detection.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
12 October 2012
Publication Number
24/2014
Publication Type
INA
Invention Field
COMMUNICATION
Status
Email
Parent Application
Patent Number
Legal Status
Grant Date
2023-12-29
Renewal Date

Applicants

TATA CONSULTANCY SERVICES LIMITED
Nirmal Building  9th Floor  Nariman Point  Mumbai-400021  Maharashtra

Inventors

1. BANERJEE  Rohan
Tata Consultancy Services Plot A2  M2 & N2  Sector V  Block GP  Salt Lake Electronics Complex  Kolkata-700091 West Bengal
2. SINHA  Aniruddha
Tata Consultancy Services Plot A2  M2 & N2  Sector V  Block GP  Salt Lake Electronics Complex  Kolkata-700091 West Bengal

Specification

FORM 2
THE PATENTS ACT, 1970
(39 of 1970)
&
THE PATENTS RULES, 2003
COMPLETE SPECIFICATION
(See section 10, rule 13)
1. Title of the invention: REAL-TIME TRAFFIC DETECTION
2. Applicant(s)
NAME NATIONALITY ADDRESS
TATA CONSULTANCY
SERVICES LIMITED
Indian Nirmal Building, 9th Floor,
Nariman Point,
Mumbai-400021, Maharashtra
India
3. Preamble to the description
COMPLETE SPECIFICATION
The following specification particularly describes the invention and the manner in which it
is to be performed.
1
2
TECHNICAL FIELD
[0001] The present subject matter relates, in general, to traffic detection and, in
particular, to systems and methods for real-time traffic detection.
BACKGROUND
[0002] Traffic congestion is an ever increasing problem, particularly, in urban areas.
Since the urban areas are usually populated, it has become difficult to travel without incurring
delays due to traffic congestion, accidents, and other problems. It has become necessary to
monitor the traffic congestion in order to provide travelers with accurate and real-time traffic
information to avoid problems.
[0003] Several traffic detection systems have been developed in the past few years for
detecting the traffic congestion. Such traffic detection systems include a system comprising a
plurality of user devices, such as mobile phones and smart phones communicating with a
central server, such as a backend server, through a network for detecting the traffic congestion
at various geographical locations. The user devices capture ambient sounds, i.e., the sounds
present in an environment surrounding the user devices, which is processed for traffic
detection. In some of the traffic detection systems, processing is entirely carried out at the
user devices, and the processed data is sent to the central server for traffic detection. While in
other traffic detection systems, the processing is entirely carried out by the central server for
traffic detection. Thus, the processing overhead increases on a single entity, i.e., either on the
user device or the central server, thereby leading to slow response time, and delay in
providing the traffic information to the users.
SUMMARY
[0004] This summary is provided to introduce concepts related to real-time traffic
detection. These concepts are further described below in the detailed description. This
summary is not intended to identify essential features of the claimed subject matter nor is it
intended for use in determining or limiting the scope of the claimed subject matter.
3
[0005] Systems and methods for real-time traffic detection are described. In one
embodiment, the method comprises capturing ambient sounds as an audio sample, and
segmenting the audio sample into a plurality of audio frames. Further, the method comprises
identifying periodic frames amongst the plurality of audio frames. Spectral features of the
identified periodic frames are extracted, and horn sounds are identified based on the spectral
features. The identified horn sounds are then used for real-time traffic detection.
BRIEF DESCRIPTION OF THE DRAWINGS
[0006] The detailed description is provided with reference to the accompanying
figures. In the figures, the left-most digit(s) of a reference number identifies the figure in
which the reference number first appears. The same numbers are used throughout the
drawings to reference like features and components.
[0007] Fig. 1 illustrates a traffic detection system, in accordance with an embodiment
of the present subject matter.
[0008] Fig. 2 illustrates details of the traffic detection system, according to an
embodiment of the present subject matter.
[0009] Fig. 3 illustrates an exemplary tabular representations depicting comparison of
total time taken for detecting the traffic congestion by the present traffic detection system and
a conventional traffic detection system.
[0010] Figs. 4a and 4b illustrate a method for real-time traffic detection, in
accordance to another embodiment of the present subject matter.
DETAILED DESCRIPTION
[0011] Conventionally, various sound based traffic detection systems are available for
detecting traffic congestion at various geographical locations, and providing traffic
information to users in order avoid problems due to the traffic congestion. Such sound based
traffic detection systems capture ambient sounds, which is processed for traffic detection. The
processing of the ambient sounds typically involves extracting spectral features of the ambient
sounds, determining level, i.e., pitch or volume, of the ambient sounds based on the spectral
features, and comparing the detected level with a predefined threshold to detect the traffic
congestion. For example, when the comparison indicates that the detected levels of the
4
ambient sounds are above the predefined threshold, the traffic congestion at the geographical
location of the user device is detected and traffic information is provided to the users, such as
travelers.
[0012] Such conventional traffic detection systems, however, suffers from numerous
drawbacks. The processing of the ambient sounds in the conventional traffic detection
systems is typically carried out either by the user devices or the central server. In both the
cases, the processing overhead increases on a single entity, i.e., the user device or the central
server, thereby leading to slow response time. Because of the slow response time, there is a
time delay in providing the traffic information to the users. The conventional systems,
therefore, fail to provide real-time traffic information to the users. Moreover, when the entire
processing is carried out at the user devices, battery consumption of the user devices increases
tremendously, posing difficulties to the users.
[0013] Further, the conventional traffic detection systems rely on the pitch or volume,
of the ambient sounds for detecting the traffic congestion. However, the ambient sounds are
usually a mixture of different types of sounds including human speech, environmental noise,
vehicle’s engine noise, music being played in vehicles, horn sounds, etc. Taking a scenario,
where a pitch of the human speech and music being played in the vehicles is too high, and the
user devices placed in the vehicles captures these ambient sounds containing high volume of
human speech and music along with the other sounds. In such a scenario, if the level of these
ambient sounds is identified as higher than the predefined threshold, traffic congestion is
detected falsely and the false traffic information is provided to the users. Thus, these
conventional traffic detection systems fail to provide reliable traffic information.
[0014] In accordance with the present subject matter, systems and methods for
detecting real time traffic congestion are described. In one embodiment, the traffic detection
system comprises a plurality of user devices and a central server (hereinafter referred to as
server). The user devices communicate with the server through a network for real-time traffic
detection. The user devices referred herein may include, but are not restricted to,
communication devices, such as mobile phones and smart phones, or computing devices, such
as Personal Digital Assistants (PDA) and laptops.
[0015] In one implementation, the user devices capture ambient sounds, i.e., the
sounds present in an environment surrounding the user devices. The ambient sounds may
5
include, for example, tire noise, music being played in vehicle(s), human speech, horn sound,
and engine noise. Additionally, the ambient sounds may contain background noise including
environmental noise and background traffic noise. The ambient sounds are captured as an
audio sample of short time duration, say, few minutes. The audio sample, thus, captured by
the user devices can be stored within a local memory of the user devices.
[0016] The audio sample is then processed partly by the user devices and partly by the
server to detect the traffic congestion. At the user device end, the audio sample is segmented
into a plurality of audio frames. Subsequent to the segmentation, background noise is filtered
from the plurality of audio frames. The background noise may affect the sound which
produces peaks of high frequency. Therefore, the background noise is filtered from the
plurality of audio frames to generate a plurality of filtered audio frames. The plurality of
filtered audio frames may be stored in the local memory of the user devices.
[0017] Once the plurality of audio frames is filtered, the audio frames are separated
into three types of frames, i.e., periodic frames, non-periodic frames, and silenced frames.
The periodic frames may include a mixture of horn sound and human speech, and the nonperiodic
frames may include a mixture of tire noise, music played in the vehicle(s), and
engine noise. The silenced frames, does not include any kind of sound.
[0018] Out of the above mentioned three types of frames, the periodic frames are then
picked up for further processing. To pick up or identify the periodic frames, the non-periodic
frames and the silenced frames are rejected based on the Power Spectral Density (PSD) and
short term energy level (En) of the audio frames respectively.
[0019] In one implementation, spectral features of the identified periodic frames are
extracted by the user device. The spectral features used in this application are disclosed in copending
Indian Patent Application No. 462/MUM/2012, which is incorporated herein by
reference. The spectral features referred herein may include, but not limited to, one or more of
Mel-Frequency Cepstral Coefficients (MFCC), inverse Mel-Frequency Cepstral Coefficients
(inverse MFCC), and modified Mel-Frequency Cepstral Coefficients (modified MFCC).
Since, the periodic frames include mixture of the horn sound and the human speech, the
extracted spectral features corresponds to the features of both the horn sound and the human
speech. The extracted spectral features are then transmitted to the server, via the network, for
traffic detection.
6
[0020] At the server end, the spectral features are received from the plurality of user
devices at a particular geographical location. Based on the spectral features, the horn sound
and the human speech is segregated using one or more known sound models. In one
implementation, the sound models include a horn sound model and a traffic sound model. The
horn sound model is configured to detect only the horn sound, while the traffic sound model is
configured to detect different type of traffic sounds other than the horn sounds. Based on the
segregation, level or rate of the horn sounds is compared with a predefined threshold, to detect
the traffic congestion at the geographical location, and real-time traffic information is
subsequently provided to the users, via, the network.
[0021] In one implementation, the user devices are capable of operating in an online
mode as well as an offline mode. For example, in the online mode, the user devices can be
connected to the server, via, the network during the complete processing. While, in the offline
mode, the user devices are capable of performing the in-part processing, without being
connected to the server. In order to communicate with the server for further processing, the
user devices can be switched to the online mode, and the server will carry out rest of the
processing to detect traffic.
[0022] According to the systems and the methods of the present subject matter,
processing load on the user devices and the server is segregated. Thus, real-time traffic
detection is achieved. Moreover, only the required audio frames, i.e., the periodic frames, are
taken up for processing, unlike the prior art where the entire audio frames are processed
containing additional noises that may lead to erroneous traffic detection, and circulation of
false traffic information to the users. Thus, the systems and the methods of the present subject
matter provide reliable traffic information to the users. Also, processing of only required
audio frames by the user devices further reduces processing load and processing time, thereby
reducing battery consumption.
[0023] The following disclosure describes system and method of real-time traffic
detection. While aspects of the described system and method may be implemented in any
number of different computing systems, environments, and/or configurations, embodiments
are described in the context of the following exemplary system architecture(s).
[0024] Fig. 1 illustrates a traffic detection system 100, in accordance with an
embodiment of the present subject matter. In one implementation, the traffic detection system
7
100 (hereinafter referred to as system 100) comprises a plurality of user devices 102-1, 102-2,
102-3,...102-N are connected, through a network 104, to a server 106. The user devices 102-1,
102-2, 102-3,...102-N are collectively referred to as the user devices 102 and individually
referred to as a user device 102. The user devices 102 may be implemented as any of a variety
of conventional communication devices, including, for example, mobile phones and smart
phones, and/or conventional computing devices, such as Personal Digital Assistants (PDAs)
and laptops.
[0025] The user devices 102 are connected to the server 106 over the network 104
through one or more communication links. The communication links between the user devices
102 and the server 106 are enabled through a desired form of communication, for example,
via dial-up modem connections, cable links, digital subscriber lines (DSL), wireless or
satellite links, or any other suitable form of communication.
[0026] The network 104 may be a wireless network. In one implementation, the
network 104 can be an individual network, or a collection of many such individual networks,
interconnected with each other and functioning as a single large network, e.g., the Internet or
an intranet. Examples of the individual networks include, but are not limited to, Global
System for Mobile Communication (GSM) network, Universal Mobile Telecommunications
System (UMTS) network, Personal Communications Service (PCS) network, Time Division
Multiple Access (TDMA) network, Code Division Multiple Access (CDMA) network, Next
Generation Network (NGN), and Integrated Services Digital Network (ISDN). Depending on
the technology, the network 104 may include various network entities, such as gateways,
routers, network switches, and hubs, however, such details have been omitted for ease of
understanding.
[0027] In an implementation, each of the user devices 102 includes a frame separation
module 108 and an extraction module 110. For example, the user device 102-1 includes a
frame separation module 108-1 and the extraction module 110-1, and the user device 102-2
includes a frame separation module 108-2 and the extraction module 110-2, and so on. The
server 106 includes a traffic detection module 112.
[0028] In one implementation, the user devices 102 capture ambient sounds. The
ambient sounds may include tire noise, music played in vehicles, human speech, horn sound,
and engine noise. The ambient sounds may also contain background noise including
8
environmental noise and background traffic noise. The ambient sounds are captured as an
audio sample, for example, an audio sample of short time duration, say, few minutes. The
audio sample may be stored within a local memory of the user device 102.
[0029] The user device 102 segments the audio sample into a plurality of audio frames
and then filters the background noise from the plurality of audio frames. In one
implementation, the filtered audio frames may be stored within the local memory of the user
device 102.
[0030] Subsequent to the filtration, the frame separation module 108 separates the
filtered audio frames into periodic frames, non-periodic, and silenced frames. The periodic
frames may include a mixture of horn sound and human speech, and the non-periodic frames
may include a mixture of tire noise, music played in the vehicle(s), and engine noise. The
silenced frames, does not include any kind of sound. Based on the separation, the frame
separation module 108 identifies the periodic frames.
[0031] The extraction module 110 within the user device 102 then extracts spectral
features of the periodic frames, such as one or more of Mel-Frequency Cepstral Coefficients
(MFCC), inverse Mel-Frequency Cepstral Coefficients (inverse MFCC), and modified Mel-
Frequency Cepstral Coefficients (modified MFCC), and transmits the extracted spectral
features to the server 106. As indicated previously, the periodic frames include mixture of the
horn sound and the human speech, the extracted spectral features, thus, corresponds to the
features of both the horn sound and the human speech. In one implementation, the extracted
spectral features can be stored within the local memory of the user device 102. Upon
receiving the extracted spectral features from a plurality of user devices 102 at a geographical
location, the server 106 segregates the horn sound and human speech based on known sound
models. Based on the horn sound, the traffic detection module 112 within the server 106
detects the real-time traffic at the geographical location.
[0032] Fig. 2 illustrates details of traffic detection system 100, according to an
embodiment of the present subject matter.
[0033] In said embodiment, the traffic detection system 100 may include a user device
102 and a server 106. The user device 102 includes one or more device processor(s) 202, a
device memory 204 coupled to the device processor 202, and device interface(s) 206. The
9
server 106 includes one or more server processor(s) 230, a server memory 232 coupled to the
server processor 230, and server interface(s) 234.
[0034] The device processor 202 and the server processor 230 can be a single
processing unit or a number of units, all of which could include multiple computing units. The
device processor 202 and the server processor 230 may be implemented as one or more
microprocessors, microcomputers, microcontrollers, digital signal processors, central
processing units, state machines, logic circuitries, and/or any devices that manipulate signals
based on operational instructions. Among other capabilities, the device processor 202 and the
server processor 230 are configured to fetch and execute computer-readable instructions and
data stored in the device memory 204 and the server memory 232 respectively.
[0035] The device interfaces 206 and the server interfaces 234 may include a variety
of software and hardware interfaces, for example, interface for peripheral device(s), such as a
keyboard, a mouse, an external memory, a printer, etc. Further, the device interfaces 206 and
the server interfaces 234 may enable the user device 102 and the server 106 to communicate
with other computing devices, such as web servers and external databases. The device
interfaces 206 and the server interfaces 234 may facilitate multiple communications within a
wide variety of protocols and networks, such as a network including wireless networks, e.g.,
WLAN, cellular, satellite, etc. The device interfaces 206 and the server interfaces 234 may
include one or more ports to allow communication between the user device 102 and the server
106.
[0036] The device memory 204 and the server memory 232 may include any
computer-readable medium known in the art including, for example, volatile memory such as
static random access memory (SRAM) and dynamic random access memory (DRAM), and/or
non-volatile memory, such as read only memory (ROM), erasable programmable ROM, flash
memories, hard disks, optical disks, and magnetic tapes. The device memory 204 further
includes device module(s) 208 and device data 210, and the server memory 232 further
includes server module(s) 236 and server data 238.
[0037] The device modules 208 and the server modules 236 include routines,
programs, objects, components, data structures, etc., which perform particular tasks or
implement particular abstract data types. In one implementation, the device module(s) 208
include an audio capturing module 212, a segmentation module 214, a filtration module 216,
10
the frame separation module 108, the extraction module 110, and device other module(s) 218.
In said implementation, the server module(s) 236 include a sound detection module 240, the
traffic detection module 112, and the server other module(s) 242. The device other module(s)
218 and the server other module(s) 242 may include programs or coded instructions that
supplement applications and functions, for example, programs in the operating system of the
user device 102 and the server 106 respectively.
[0038] The device data 210 and the server data 238, amongst other things, serves as
repositories for storing data processed, received, and generated by one or more of the device
module(s) 208 and the server module(s) 236. The device data 210 includes audio data 220,
frame data 222, feature data 224, and device other data 226. The server data 238 includes
sound data 244 and server other data 248. The device other data 226 and the server other data
248 includes data generated as a result of the execution of one or more modules in the device
other module(s) 218 and the server other modules 242.
[0039] In operation, the audio capturing module 212 of the user device 102 captures
ambient sounds, i.e., the sounds present in an environment surrounding the user device 102.
Such ambient sounds may include tire noise, music played in vehicles, human speech, horn
sound, engine noise. Additionally, the ambient noise includes background noise containing
environmental noise, and background traffic noise. The ambient sounds may be captured as an
audio sample either continuously or at predefined time intervals, say, after every 10 minutes.
Time duration of the audio sample captured by the user device 102 may be short, say, few
minutes. In one implementation, the captured audio sample may be stored in a local memory
of the user device 102, as the audio data 220, which can be retrieved when required.
[0040] In one implementation, the segmentation module 214 of the user device 102
retrieves the audio sample, and segments the audio sample into a plurality of audio frames. In
one example, the segmentation module 214 segments the audio sample using a conventionally
known hamming window segmentation technique. In the hamming window segmentation
technique, a hamming window of a predefined duration, for example, 100ms is defined. As an
instance, if the audio sample of about 12 minutes of time duration is segmented with a
hamming window of 100ms, then the audio sample is segmented into about 7315 audio
frames.
11
[0041] In one implementation, the segmented audio frames, thus, obtained are
provided as an input to the filtration module 216, which is configured to filter the background
noise from the plurality of audio frames, as the background noise may affect that sound which
produces peaks of high frequency. For example, the horn sounds that are considered to
produce peaks of high frequency are susceptible to the background noise. Therefore, the
filtration module 216 filters the background noise, to boost up such kind of sounds. The audio
frames, thus, generated as a result of the filtration is hereinafter referred to as filtered audio
frames. In one implementation, the filtration module 216 may store the filtered audio frames
as the frame data 222 with the local memory of the user device 102.
[0042] The frame separation module 108 of the user device 102 is configured to
segregate the audio frames or the filtered audio frames into periodic frames, non-periodic
frames, and silenced frames. The periodic frames may be a mixture of horn sound and human
speech, and the non-periodic frames may be a mixture of tire noise, music played in the
vehicles, and the engine noise. The silenced frames are the frames without any sound, i.e.,
soundless frames. For segregation, the frame separation module 108 computes short term
energy level (En) of each of the audio frames or the filtered audio frames, and compares the
computed short term energy level (En) to a predefined energy threshold (EnTh). The audio
frames having the short term energy level (En) less than the energy threshold (EnTh) are
rejected as the silenced frames and the remaining audio frames are further examined to
identify the periodic frames amongst them. For example, if the total number of filtered audio
frames is about 7315, the energy threshold (EnTh) is 1.2 and the number of filtered audio
frames with short term energy level (En) less than 1.2 is 700. In said example, the 700 filtered
audio frames are rejected as silenced frames and the remaining 6615 filtered audio frames are
further examined to identify the periodic frames amongst them.
[0043] The frame separation module 108 calculates total power spectral density (PSD)
of the remaining audio frames, and maximum PSD of a filtered audio frame. The total PSD of
remaining filtered audio frames taken together is denoted as PSDTotal and the maximum PSD
of the filtered audio frame is denoted as PSDMax to identify the periodic frames amongst the
plurality of filtered audio frames. According to one implementation, the frame separation
module 108 identifies the periodic frames using the equation (1) provided below:
12
r _
PSD__
PSD___
…. _1_
wherein, PSD__ represents the maximum PSD of a filtered audio frame,
PSD___ represents the total PSD of the filtered audio frames, and
r represents the ratio of the PSDMax to the PSDTotal.
[0044] The ratio as obtained by the above equation is then compared with the
predefined density threshold (PSDTh) by the frame separation module 108 to identify the
periodic frames. For example, an audio frame is identified to be periodic, if the ratio is greater
than the density threshold (PSDTh). While, the audio frame is rejected if the ratio is lesser than
the density threshold (PSDTh). Such a comparison is carried out separately for each of the
filtered frames to identify all the periodic frames.
[0045] Once the periodic frames are identified, the extraction module 110 of the user
device 102 is configured to extract spectral features of the identified periodic frames. The
extracted spectral features may include one or more of Mel-Frequency Cepstral Coefficients
(MFCC), inverse Mel-Frequency Cepstral Coefficients (inverse MFCC), and modified Mel-
Frequency Cepstral Coefficients (modified MFCC). In one implementation, the extraction
module 110 extracts the spectral features based on conventionally known feature extraction
techniques. As indicated earlier, the periodic frames include a mixture of horn sound and the
human speech, the extracted spectral features therefore corresponds to the horn sound and the
human speech.
[0046] Subsequent to extraction of the spectral features, the extraction module 110
transmits the extracted spectral features to the server 106 for further processing. The
extraction module 110 may store the extracted spectral features of the periodic frames as the
feature data 244 in the local memory of the user device 102.
[0047] At the server end, the sound detection module 240 of the server 106 receives
the extracted spectral features from multiple user devices 102 falling under a common
geographical location, and segregates the collated spectral features into horn sounds and
human speech. The sound detection module 240 performs the segregation based on
conventionally available sound models including a horn sound model and a traffic sound
model. The horn sound model is configured to identify the horn sounds, and the traffic sound
model is configured to identify traffic sounds other than the horn sounds, for example, human
13
speech, tire noise, and music played in the vehicles. The horn sound and the human speech
have different spectral properties. For example, the human speech produces peaks in the range
of 500-1500 KHz (Kilo Hertz) and the horn sound produce peaks above 2000 KHz (Kilo
Hertz). When the spectral features are fed as an input to these sound models, the horn sounds
are identified. The sound detection module 240 may store the identified horn sounds as sound
data 224 in the server 106.
[0048] The traffic detection module 112 of the server 106 is then configured to detect
the real-time traffic based on the identification of the horn sound. As the horn sounds
represents rate of honking on the road, which is more when there is traffic congestion. The
identified horn sounds are compared with predefined threshold by the traffic detection module
112 to detect traffic at the geographical location.
[0049] Thus, according to present subject matter for detecting the real-time traffic
congestion, the periodic frames are separated from the audio sample and spectral features are
extracted only for the periodic frames, thereby reducing the overall processing time and the
battery consumption by the user devices 102. Also, since the extracted features of only the
periodic frames are transmitted by the user devices 102 to the server 106, the load on the
server is also reduced and thus, time taken by the server 106 to detect traffic is significantly
reduced.
[0050] Fig. 3 illustrates an exemplary tabular representations depicting comparison of
total time taken for detecting the traffic congestion by the present traffic detection system and
a conventional traffic detection system.
[0051] As shown in the Fig 3, the table 300 corresponds to the conventional traffic
detection system and the table 302 corresponds to the present traffic detection system 100. As
shown in the table 300, three audio samples, namely, a first audio sample, a second audio
sample, and a third audio sample, are processed by the conventional traffic detection system
for detecting the traffic congestion. Such audio samples are segmented into a plurality of
audio frames, such that each audio frame is of a time duration 100ms. For example, the first
audio sample is segmented into 7315 audio frames of duration 100ms. Likewise, the second
audio sample is segmented into 7927 audio frames, and the third audio sample is segmented
into 24515 audio frames. Further, spectral features are extracted for all the three audio frames.
The total processing time taken by the conventional traffic detection system for the
14
processing, especially, the spectral feature extraction of three audio samples are 710 sec, 793
sec, and 2431 sec respectively and corresponding size of extracted spectral features is 1141
KB, 1236 KB, and 3824 KB respectively.
[0052] On the other hand, the present traffic detection system 100 also processed the
same three audio samples as shown in the table 302. The audio samples are segmented into a
plurality of audio frames, such as periodic frames, non-periodic frames and silenced frames.
However, the present traffic detection system 100 picks up only the periodic frames for
processing. The time taken to identify the periodic frames from the first audio sample, the
second audio sample, and the third audio sample is 27 sec, 29 sec, and 62 sec respectively.
The spectral features are then extracted for the identified periodic frames. Time taken by the
present traffic detection system 100 to extract the spectral features of the periodic frames is
351 sec, 362 sec, and 1829 sec, for the first audio sample, the second audio sample, and the
third audio sample respectively, and the corresponding size of extracted spectral features is
544 KB, 548 KB, and 2776 KB. Therefore, total processing time taken by the present traffic
detection system 100 for processing the first audio sample, the second audio sample, and the
third audio sample is 378 sec, 391 sec, and 1891 sec.
[0053] It is clear from the table 300 and the table 302 that the total time taken by the
present traffic detection system 100 for processing of the audio samples is significantly less
than the total processing time taken by the conventional traffic detection system. Such a
reduction in the processing time is achieved due to separation of frames into periodic, nonperiodic,
and silenced frames, and processing only the periodic frames for spectral features
extraction unlike the conventional traffic detection systems where all the frames were taken
into consideration.
[0054] Figs. 4a and 4b illustrate a method 400 for real-time traffic detection, in
accordance with an embodiment of the present subject matter. Specifically, the Fig. 4a
illustrates a method 400-1 for extracting the spectral features from an audio sample, and the
Fig. 4b illustrates a method 400-2 for detection of real-time traffic congestion based on the
spectral features. The methods 400-1 and 400-2 are collectively referred to as the methods
400.
[0055] The methods 400 may be described in the general context of computer
executable instructions. Generally, computer executable instructions can include routines,
15
programs, objects, components, data structures, procedures, modules, functions, etc., that
perform particular functions or implement particular abstract data types. The methods 400
may also be practiced in a distributed computing environment where functions are performed
by remote processing devices that are linked through a communications network. In a
distributed computing environment, computer executable instructions may be located in both
local and remote computer storage media, including memory storage devices.
[0056] The order in which the methods 400 are described is not intended to be
construed as a limitation, and any number of the described method blocks can be combined in
any order to implement the methods 400, or alternative methods. Additionally, individual
blocks may be deleted from the methods without departing from the spirit and scope of the
subject matter described herein. Furthermore, the methods 400 can be implemented in any
suitable hardware, software, firmware, or combination thereof.
[0057] Referring to Fig. 4a, at block 402, the method 400-1 includes capturing
ambient sounds. The ambient sounds include tire noise, music played in vehicle(s), human
speech, horn sound, and engine noise. Further, the ambient sounds may include background
noise containing environmental noise and background traffic noise. In one implementation,
the audio capturing module 212 of the user device 102 captures ambient sounds as an audio
sample.
[0058] At block 404, the method 400-1 includes segmenting the audio sample into
plurality of audio frames. The audio sample is segmented into the plurality of audio frames
using a hamming window segmentation technique. The hamming window is a predefined
duration window. In one implementation, the segmentation module 214 of the user device 102
segments the audio sample into a plurality of audio frames.
[0059] At block 406, the method 400-1 includes filtering background noise from the
plurality of audio frames. Since the background noise affects the sounds producing peaks of
high frequency, the background noise is filtered from the audio frames. In one
implementation, the filtration module 216 filters the background noise from the plurality of
audio frames. The audio frames obtained as a result of filtration are referred to as filtered
audio frames.
[0060] At block 408, the method 400-1 includes identifying the periodic frames
amongst the plurality of filtered audio frames. In one implementation, the frame separation
16
module 108 of the user device 102 is configured to segregate the plurality of audio frames
into periodic frames, non-periodic frames, and silenced frames. The periodic frames may
include a mixture of horn sound and human speech, and the non-periodic frames may include
a mixture of tire noise, music played in the vehicle(s), and engine noise. The silenced frames,
does not include any kind of sound. Based on the segregation, the frame separation module
108 identifies the periodic frames for further processing.
[0061] At block 410, the method 400-1 includes extracting the spectral features of the
periodic frames. The extracted spectral features may include one or more of Mel-Frequency
Cepstral Coefficients (MFCC), inverse Mel-Frequency Cepstral Coefficients (inverse MFCC),
and modified Mel-Frequency Cepstral Coefficients (modified MFCC). As indicated earlier,
the periodic frames include a mixture of horn sound and human speech, thus, the extracted
spectral features corresponds to the horn sound and the human speech. In one implementation,
the extraction module 110 is configured to extract spectral features of the identified periodic
frames.
[0062] At block 412, the method 400-1 includes transmitting the extracted spectral
features to the server 106 for detecting real-time traffic congestion. In one implementation,
the extraction module 110 transmits the extracted spectral features to the server 106.
[0063] Referring to Fig. 4b, at block 414, the method 400-2 includes receiving the
spectral features from a plurality of user devices 102 in a geographical location, via, the
network 104. In one implementation, the sound detection module 240 of the server 106
receives the spectral features.
[0064] At block 416, the method 400-2 includes identifying the horn sound from the
received spectral features. The horn sound is identified, for example, based on conventionally
available sound models including the horn sound model and the traffic sound model. Based on
these sound models, distinction between the horn sound and the human speech is made and
the horn sound is therefore identified. In one implementation, the sound detection module 240
of the server 106 identifies the horn sound.
[0065] At block 418, the method 400-2 includes detecting real-time traffic congestion
based on the horn sound identified at the previous block. The horn sound is indicative of rate
of honking on the road, which is considered as a parameter for accurately detecting the traffic
congestion in the present description. Based on comparing the rate of honking or the level of
17
horn sounds with a predefined threshold value, the traffic detection module 112 detects the
traffic congestion at the geographical location.
[0066] Although embodiments for the traffic detection system have been described in
language specific to structural features and/or methods, it is to be understood that the
invention is not necessarily limited to the specific features or methods described. Rather, the
specific features and methods are disclosed as exemplary implementations for the traffic
detection system.
18
I/we claim:
1. A method for real-time traffic detection, wherein the method comprising:
capturing ambient sounds as an audio sample in a user device (102);
segmenting the audio sample into a plurality of audio frames;
identifying periodic frames amongst the plurality of audio frames; and
extracting spectral features of the periodic frames for real-time traffic detection.
2. The method as claimed in claim 1, wherein the ambient sounds include one or more of
tire noise, horn sound, engine noise, human speech, and background noise.
3. The method as claimed in claim 1, wherein the identifying comprises separating the
plurality of audio frames into the periodic frames, non-periodic frames, and silenced
frames.
4. The method as claimed in claim 3, wherein the separating comprises
computing a short term energy level for the plurality of audio frames; and
comparing the short term energy level of each of the plurality of audio frames
with a predefined energy threshold to identify the silenced frames amongst the
plurality of audio frames;
calculating a ratio of a maximum power spectral density and a total power
spectral density of remaining audio frames, wherein the remaining audio frames
exclude the silenced frames; and
identifying the periodic frames amongst the remaining audio frames based on
comparing the ratio of the maximum power spectral density and the total power
spectral density with a predefined density threshold.
5. The method as claimed in claim 1 further comprising filtering background noise from
the plurality of audio frames.
6. The method as claimed in claim 1, wherein the spectral features include one or more
of Mel-Frequency Cepstral Coefficients (MFCC), inverse MFCC, and modified
MFCC.
7. A method for real-time traffic detection, wherein the method comprising:
receiving spectral features of periodic frames from a plurality of user devices
(102) in a geographical location;
identifying horn sounds based on the spectral features; and
19
detecting real-time traffic congestion at the geographical location based on the
horn sounds.
8. The method as claimed in claim 7, wherein the spectral features include one or more
of Mel-Frequency Cepstral Coefficients (MFCC), inverse MFCC, and modified
MFCC.
9. The method as claimed in claim 7, wherein the identifying is based on at least one
sound model, wherein the at least one sound model is any one of a horn sound model
and a traffic sound model.
10. A user device (102) for real-time traffic detection comprising:
a device processor (202); and
a device memory (204) coupled to the device processor (202), the device
memory (204) comprising:
a segmentation module (214) configured to segment an audio sample
captured in the user device (102) into a plurality of audio frames;
a frame separation module (108) configured to separate the plurality of
audio frames into at least periodic frames and non-periodic frames; and
an extraction module (110) configured to extract spectral features of the
periodic frames, wherein the spectral features are transmitted to a server (106)
for real-time traffic detection.
11. The user device (102) as claimed in claim 10, wherein the user device (102) further
comprising a filtration module (216) configured to filter background noise from the
plurality of audio frames.
12. The user device (102) as claimed in claim 10, wherein the frame separation module
(108) is configured to separate the plurality of audio frames based on a short term
energy level (En) and a Power Spectral Density (PSD) of the plurality of audio frames.
13. A server (106) for real-time traffic detection comprising:
a server processor (230); and
a server memory (232) coupled to the server processor (230), the server memory
(232) comprising:
a sound detection module (240) configured to:
20
receive spectral features of periodic frames from a plurality of user
devices (102) in a geographical location; and
identify horn sounds based on the spectral features; and
a traffic detection module (242) configured to detect real-time traffic
congestion at the geographical location based on the horn sounds.
14. The server (106) as claimed in claim 13, wherein the sound detection module (240) is
configured to identify the horn sounds based on at least one of a horn sound model and
a traffic sound model.
15. A computer-readable medium having embodied thereon a computer program for
executing a method comprising:
capturing ambient sounds as an audio sample;
segmenting the audio sample into a plurality of audio frames;
identifying periodic frames amongst the plurality of audio frames;
extracting spectral features of the periodic frames;
identifying horn sounds based on the spectral features; and
detecting real-time traffic congestion based on the horn sounds.

Documents

Application Documents

# Name Date
1 3005-MUM-2012-FORM 18(17-10-2012).pdf 2012-10-17
2 3005-MUM-2012-CORRESPONDENCE(17-10-2012).pdf 2012-10-17
3 3005-MUM-2012-FORM 1(18-10-2012).pdf 2012-10-18
4 3005-MUM-2012-CORRESPONDENCE(18-10-2012).pdf 2012-10-18
5 ABSTRACT1.jpg 2018-08-11
6 3005-MUM-2012-FORM 3(11-3-2014).pdf 2018-08-11
7 3005-MUM-2012-FORM 26(4-12-2012).pdf 2018-08-11
8 3005-MUM-2012-CORRESPONDENCE(4-12-2012).pdf 2018-08-11
9 3005-MUM-2012-CORRESPONDENCE(11-3-2014).pdf 2018-08-11
10 3005-MUM-2012-FER.pdf 2018-09-20
11 3005-MUM-2012-OTHERS [15-03-2019(online)].pdf 2019-03-15
12 3005-MUM-2012-Information under section 8(2) (MANDATORY) [15-03-2019(online)].pdf 2019-03-15
13 3005-MUM-2012-FORM 3 [15-03-2019(online)].pdf 2019-03-15
14 3005-MUM-2012-FER_SER_REPLY [15-03-2019(online)].pdf 2019-03-15
15 3005-MUM-2012-DRAWING [15-03-2019(online)].pdf 2019-03-15
16 3005-MUM-2012-CORRESPONDENCE [15-03-2019(online)].pdf 2019-03-15
17 3005-MUM-2012-COMPLETE SPECIFICATION [15-03-2019(online)].pdf 2019-03-15
18 3005-MUM-2012-CLAIMS [15-03-2019(online)].pdf 2019-03-15
19 3005-MUM-2012-ABSTRACT [15-03-2019(online)].pdf 2019-03-15
20 3005-MUM-2012-US(14)-HearingNotice-(HearingDate-17-10-2023).pdf 2023-09-13
21 3005-MUM-2012-Correspondence to notify the Controller [27-09-2023(online)].pdf 2023-09-27
22 3005-MUM-2012-FORM-26 [13-10-2023(online)].pdf 2023-10-13
23 3005-MUM-2012-Written submissions and relevant documents [30-10-2023(online)].pdf 2023-10-30
24 3005-MUM-2012-PETITION UNDER RULE 137 [30-10-2023(online)].pdf 2023-10-30
25 3005-MUM-2012-Response to office action [02-11-2023(online)].pdf 2023-11-02
26 3005-MUM-2012-PatentCertificate29-12-2023.pdf 2023-12-29
27 3005-MUM-2012-IntimationOfGrant29-12-2023.pdf 2023-12-29

Search Strategy

1 searchstrategy_20-09-2018.pdf

ERegister / Renewals

3rd: 18 Jan 2024

From 12/10/2014 - To 12/10/2015

4th: 18 Jan 2024

From 12/10/2015 - To 12/10/2016

5th: 18 Jan 2024

From 12/10/2016 - To 12/10/2017

6th: 18 Jan 2024

From 12/10/2017 - To 12/10/2018

7th: 18 Jan 2024

From 12/10/2018 - To 12/10/2019

8th: 18 Jan 2024

From 12/10/2019 - To 12/10/2020

9th: 18 Jan 2024

From 12/10/2020 - To 12/10/2021

10th: 18 Jan 2024

From 12/10/2021 - To 12/10/2022

11th: 18 Jan 2024

From 12/10/2022 - To 12/10/2023

12th: 18 Jan 2024

From 12/10/2023 - To 12/10/2024

13th: 08 Oct 2024

From 12/10/2024 - To 12/10/2025

14th: 09 Oct 2025

From 12/10/2025 - To 12/10/2026