Sign In to Follow Application
View All Documents & Correspondence

Lane Detection

Abstract: A system and method for lane detection are provided. An edge image including inbound edges of lane-markings is generated corresponding to a road image. Based on an adaptive thresholding of intensity of lane-markings from image strips selected from the image, a plurality of lane masks are extracted. A plurality of valid lane masks are selected from the plurality of lane masks based on a structure and intensity information of the plurality of valid lane masks in the image strips. An overlap between the inbound edges and the valid lane masks is determined to identify a plurality of edge segments of the pair of inbound edges belonging to the lane-markings. A set of valid edge segments of the pair of inbound edges are obtained based on the curvature information of the edge segments.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
28 May 2015
Publication Number
50/2016
Publication Type
INA
Invention Field
COMPUTER SCIENCE
Status
Email
ip@legasis.in
Parent Application
Patent Number
Legal Status
Grant Date
2023-03-14
Renewal Date

Applicants

Tata Consultancy Services Limited
Nirmal Building, 9th Floor, Nariman Point, Mumbai 400021, Maharashtra, India

Inventors

1. SUDDAMALLA, Upendra
Tata Consultancy Services Limited, Salarpuria G. R. Tech Park, Dhara Block, Mahadevapura, K R Puram, Bangalore 560 066, Karnataka, India
2. DAS, Apurba
Tata Consultancy Services Limited, Salarpuria G. R. Tech Park, Dhara Block, Mahadevapura, K R Puram, Bangalore 560 066, Karnataka, India
3. KUNDU, Saikat
Tata Consultancy Services Limited, Salarpuria G. R. Tech Park, Dhara Block, Mahadevapura, K R Puram, Bangalore 560 066, Karnataka, India
4. FARKADE, Sameer Narendra
Tata Consultancy Services Limited, Salarpuria G. R. Tech Park, Dhara Block, Mahadevapura, K R Puram, Bangalore 560 066, Karnataka, India

Specification

DESC:FORM 2

THE PATENTS ACT, 1970
(39 of 1970)
&
THE PATENTS RULES, 2003

COMPLETE SPECIFICATION
(See section 10 and rule 13)

Title of invention:
LANE DETECTION

Applicant:
Tata Consultancy Services Limited
A company Incorporated in India under the Companies Act, 1956
Having address:
Nirmal Building, 9th floor,
Nariman point, Mumbai 400021,
Maharashtra, India

The following specification particularly describes the invention and the manner in which it is to be performed.

CROSS REFERENCE TO RELATED APPLICATIONS AND PRIORITY
[001] The present invention claims priority to Indian provisional specification (Title: Method and system for lane detection) No. 2079/MUM/2015, filed in India on May 28, 2015.

TECHNICAL FIELD
[002] The embodiments herein generally relate to lane detection, and, more particularly, to lane detection under varying illumination and road conditions.

BACKGROUND
[003] Detection of lane marking using visual information is a primary requirement for supporting driver assistance and safety features such as Lane Departure Warning (LDW), Lane Keep Assistance (LKA) and Forward Collision Alert (FCA). For detecting lane markings, cameras with clear sensors or color filtered sensors are used. For driver assistance appli-cations, a vision sensor, i.e., camera is mounted in front of the vehicle to capture the road im-ages. The camera is an assembly of a CMOS / CCD sensor of size VGA or HD used along with a lens of horizontal field of view (FOV) around 45 degree and vertical field of view of around 30 degree. The lane markings may vary in color (white and yellow for India, USA, Europe; Blue for South Korea), width and shape (solid and dashed).

SUMMARY
[004] The following presents a simplified summary of some embodiments of the disclosure in order to provide a basic understanding of the embodiments. This summary is not an extensive overview of the embodiments. It is not intended to identify key/critical elements of the embodiments or to delineate the scope of the embodiments. Its sole purpose is to present some embodiments in a simplified form as a prelude to the more detailed description that is presented below.
[005] In view of the foregoing, an embodiment herein provides methods and systems for lane detection. In one aspect, a computer implemented method for lane detection is pro-vided, where the method includes generating an edge image corresponding to at least one im-age of a road having lane markings, where the edge image includes a pair of inbound edges corresponding to the lane markings. The method further includes extracting a plurality of lane masks from a plurality of horizontal image strips of the at least one image. A lane mask of the plurality of lane masks in a horizontal image strip is indicative of a subset of the lane markings in the horizontal image strip. The plurality of lane masks are extracted based on an adaptive thresholding of intensity of the subset of the lane markings in the plurality of horizontal image strips. Furthermore, the method includes selecting a plurality of valid lane masks from amongst the plurality of lane masks based at least on structure information and intensity in-formation of the plurality of lane masks in the plurality of horizontal image strips. Moreover, the method includes determining an overlap between the pair of inbound edges and the plu-rality of valid lane masks to identify a plurality of edge segments of the pair of inbound edges corresponding to the lane markings. The plurality of edge segments of the pair of inbound edges includes a curvature information, where the curvature information of an edge segment being indicative of curvature of the edge segment with respect to the lane markings. A set of valid edge segments of the pair of inbound edges is obtained based on the curvature informa-tion. The set of valid edge segments facilitates in detecting the lane markings.
[006] In another aspect, a system for lane detection is provided. The system includes at least one memory; and at least one processor, the at least one memory coupled to the at least one processor and the at least one media sensor wherein the at least one processor is ca-pable of executing programmed instructions stored in the at least one memory to: generate an edge image corresponding to at least one image of a road having lane markings where the edge image comprising a pair of inbound edges corresponding to the lane markings. The sys-tem is further caused to extract a plurality of lane masks from a plurality of horizontal image strips of the at least one image, where a lane mask of the plurality of lane masks in a horizon-tal image strip is indicative of a subset of the lane markings in the horizontal image strip. The plurality of lane masks are extracted based on an adaptive thresholding of intensity of the subset of the lane markings in the plurality of horizontal image strips. The system is further caused to select a plurality of valid lane masks from amongst the plurality of lane masks based at least on structure information and intensity information of the plurality of lane masks in the plurality of horizontal image strips. An overlap is determined between the pair of inbound edges and the plurality of valid lane masks to identify a plurality of edge segments of the pair of inbound edges corresponding to the lane markings. The plurality of edge segments of the pair of inbound edges includes a curvature information. The curvature information of an edge segment being indicative of curvature of the edge segment with respect to the lane markings. The system is further caused to obtain a set of valid edge segments of the pair of inbound edges based on the curvature information, where the set of valid edge segments facilitates in detecting the lane markings.
[007] In yet another aspect, a non-transitory computer-readable medium having em-bodied thereon a computer program for executing a method for lane detection is provided. In one aspect, a computer implemented method for lane detection is provided, where the method includes generating an edge image corresponding to at least one image of a road having lane markings, where the edge image includes a pair of inbound edges corresponding to the lane markings. The method further includes extracting a plurality of lane masks from a plurality of horizontal image strips of the at least one image. A lane mask of the plurality of lane masks in a horizontal image strip is indicative of a subset of the lane markings in the horizontal image strip. The plurality of lane masks are extracted based on an adaptive thresholding of intensity of the subset of the lane markings in the plurality of horizontal image strips. Furthermore, the method includes selecting a plurality of valid lane masks from amongst the plurality of lane masks based at least on structure information and intensity information of the plurality of lane masks in the plurality of horizontal image strips. Moreover, the method includes determining an overlap between the pair of inbound edges and the plurality of valid lane masks to identify a plurality of edge segments of the pair of inbound edges corresponding to the lane markings. The plurality of edge segments of the pair of inbound edges includes a curvature information, where the curvature information of an edge segment being indicative of curvature of the edge segment with respect to the lane markings. A set of valid edge segments of the pair of inbound edges is obtained based on the curvature information. The set of valid edge segments facilitates in detecting the lane markings.

BRIEF DESCRIPTION OF THE FIGURES
[008] The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The same numbers are used throughout the drawings to reference like features and modules.
[009] FIG. 1 illustrates an example environment for lane detection, in accordance with an example embodiment;
[0010] FIG. 2 illustrates a block diagram of a system for lane detection, in accordance with an embodiment;
[0011] FIG. 3A illustrates an image of a road, FIG. 3B illustrates an edge image corresponding to the image of FIG 3A, and FIG. 3C illustrates image including inbound edges of lane markings of the edge image of FIG. 3B, in accordance with an example embodiment;
[0012] FIG. 4 illustrates an image of a road, an example horizontal image strip and a lane mask for the road, and a reconstructed image corresponding to the image, in accordance with an example embodiment;
[0013] FIG. 5 illustrates an example representation for extracting a plurality of horizontal image strips corresponding to an image of a road, in accordance with an example embodiment;
[0014] FIGS. 6A and 6B illustrate an example histogram for a horizontal image strip associated with an image, and a processed histogram corresponding to the horizontal image strip, in accordance with an example embodiment;
[0015] FIG. 7A, 7B and 7C illustrates an example scenario for removing unintended edge portions from lane edges, in accordance with an example embodiment; and
[0016] FIG. 8 illustrates a flow diagram of a method for lane detection, in accordance with the present disclosure.
[0017] It should be appreciated by those skilled in the art that any block diagrams herein represent conceptual views of illustrative systems and devices embodying the principles of the present subject matter. Similarly, it will be appreciated that any flow charts, flow diagrams, and the like represent various processes which may be substantially represented in computer readable medium and so executed by a computer or processor, whether or not such computer or processor is explicitly shown.

DETAILED DESCRIPTION
[0018] The embodiments herein and the various features and advantageous details thereof are explained more fully with reference to the non-limiting embodiments that are illustrated in the accompanying drawings and detailed in the following description. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiments herein may be practiced and to further enable those of skill in the art to practice the embodiments herein. Accordingly, the examples should not be construed as limiting the scope of the embodiments herein.
[0019] The methods and systems are not limited to the specific embodiments described herein. In addition, the method and system can be practiced independently and separately from other modules and methods described herein. Each device element/module and method can be used in combination with other elements/modules and other methods.
[0020] The manner, in which the system and method for lane detection shall be implemented, has been explained in details with respect to the FIGS. 1 through 8. While aspects of described methods and systems for lane detection can be implemented in any number of different systems, utility environments, and/or configurations, the embodiments are described in the context of the following exemplary system(s).
[0021] FIG. 1 illustrates an example environment 100 for lane detection, in accordance with an example embodiment. Lane detection refers to detecting the margins or bounds of a road, for example a road 102 so as to define a safe zone for driving. Such margins are defined by forming lane markings, such as markings 104 and 106. Lane markings on the road indicate the road traffic rules for the vehicle drivers and pedestrians. The markings used are in white, yellow or blue (in some geographies) in color. As part of advanced driver assistance systems and autonomous vehicles, accurate lane detection is an important task for different safety functionalities such as lane departure warning, lane keeping assistance, forward collision warning, and so on.
[0022] Various embodiments disclose methods and systems for lane detection by using a lane intensity (brightness) information and a structure information of lane markings and edges to identify accurate lane bounds. In addition, said embodiments target to detect solid, dashed, straight and curvy lane markings up to long distances under medium to low level noisy scenarios. Moreover, said embodiments facilitates in handling noises in a scene for detection of lane markings. For example, proposed methods and system enables lane detection in noisy scenarios such as light shadows, other vehicles in the scene. The proposed embodiments use the structure and color information of the lane markings, so as to extracts accurate inbounds of the lane marks which improves the accuracy in further components such a lane departure warning.
[0023] In an embodiment, the methods and systems detect the lane markings at farther distances by detecting the curved, straight lanes which are solid or dashed, using the pixel intensity and edge information. Herein, an edge may be defined region associated with a sharp change in the intensity with respect to neighboring regions (or pixels). For instance, the edges of lane markings 104 are illustrated as 104a, 104b, and the edges of the lane markings 106 are illustrated as 106a, 106b. Typically, lane markings are brighter compared to the road, hence the edges of the lane marking may be associated with the sharp change in intensity. An example system implemented for lane detection is described further with reference to FIG. 2.
[0024] FIG. 2 illustrates a block diagram of a system 200 for lane detection, in accor-dance with an embodiment of the present disclosure. In an embodiment, the system 200 faci-litates in lane detection by detecting distance and detection of shapes like curvy and dashed lane markings on the road. For example, the system 200 is caused to detect the lane markings at farther distances by detecting the curved, straight lanes which are solid or dashed, using the pixel intensity and edge information of the edges of lane markings to identify the lane markings reliably. In order to achieve accurate inbounds of the lane, masks of the lane marking from spatial image strips and edges from the entire arbitrary frame are determined along with their mutual correspondence.
[0025] The system 200 includes or is otherwise in communication with at least one processor such as a processor 202, at least one memory such as a memory 204, and at least one media sensor such as media sensor 206. In an embodiment, the processor 202, memory 204, and the memory 204, and the media sensor 206 may be coupled by a system bus such as a system bus 208 or a similar mechanism.
[0026] The processor 202 may include circuitry implementing, among others, audio and logic functions associated with the communication. For example, the processor 202 may include, but are not limited to, one or more digital signal processors (DSPs), one or more mi-croprocessor, one or more special-purpose computer chips, one or more field-programmable gate arrays (FPGAs), one or more application-specific integrated circuits (ASICs), one or more computer(s), various analog to digital converters, digital to analog converters, and/or other support circuits. The processor 202 thus may also include the functionality to encode messages and/or data or information. The processor 202 may include, among other things, a clock, an arithmetic logic unit (ALU) and logic gates configured to support operation of the processor 202. Further, the processor 202 may include functionality to execute one or more software programs, which may be stored in the memory 204 or otherwise accessible to the processor 202.
[0027] The at least one memory such as a memory 204, may store any number of pieces of information, and data, used by the system to implement the functions of the system. The memory 204 may include for example, volatile memory and/or non-volatile memory. Examples of volatile memory may include, but are not limited to volatile random access memory (RAM). The non-volatile memory may additionally or alternatively comprise an electrically erasable programmable read only memory (EEPROM), flash memory, hard drive, or the like. Some examples of the volatile memory includes, but are not limited to, random access memory, dynamic random access memory, static random access memory, and the like. Some example of the non-volatile memory includes, but are not limited to, hard disks, mag-netic tapes, optical disks, programmable read only memory, erasable programmable read only memory, electrically erasable programmable read only memory, flash memory, and the like. The memory 204 may be configured to store information, data, applications, instructions or the like for enabling the system 200 to carry out various functions in accordance with various example embodiments. Additionally or alternatively, the memory 204 may be configured to store instructions which when executed by the processor 202 causes the system to behave in a manner as described in various embodiments.
[0028] In an embodiment, the media sensor 206 may include an image capturing module, such as a camera, video and/or audio module, in communication with the processor 202. The media sensor 206 may be any means for facilitating capturing images, video and/or audio for storage, display, or transmission. For example, in an exemplary embodiment in which the media sensor 206 may be embodied in a camera, such that the camera may be con-figured to form and save a digital image file from an image of the marker captured by the camera. The media sensor 206 may include hardware such as a CMOS/CCD (complementary metal-oxide semiconductor/charged coupled device) sensors configured for capturing images. In some embodiments, the media content such as images and/or videos of the road are cap-tured and stored in a memory device, such as, memory 204. The media sensor 206 can include all hardware, such as circuitry, a lens or other optical component(s), and software for creating a digital image file from a captured image.
[0029] In some example embodiments, the media sensor may include only the hard-ware needed to view an image, while a memory device, such as the memory device of the system 200 stores instructions for execution by the processor 202 in the form of software to create a digital image file from a captured image. In an exemplary embodiment, the media sensor 206 may further include a processor or co-processor which assists the processor 202 in processing image data and an encoder and/or decoder for compressing and/or decompressing image data. The encoder and/or decoder may encode and/or decode according to, for example, a joint photographic experts group (JPEG) standard or other format.
[0030] In an example embodiment, the system 200 may be embodied in a vehicle. In an embodiment, the system 200 or at least the media sensors of the system 200 may be mounted on the vehicle, for example, behind the rear view mirror of vehicle, so that the sys-tem 200 may facilitate in capturing the images of the road in real-time. In an embodiment, the media sensors 206 may be embodied in a forward facing camera installed behind the rear view mirror. In an embodiment, the system 200 may facilitate in capturing images of the road when the vehicle is in motion or is stationary. The system 200 may utilize the captured images for estimating/detecting lane markings in real-time.
[0031] The system 200 is caused to detect lane markings on the road so as to deter-mine a safe zone for driving. Herein, a width of the road lying within the pair of lane mark-ings is determined to a safe zone for driving. As a lane is margined by lane markings on left and right and the lane markings on the road are of measurable thickness, the inbounds of the lane markings represent a safe region of the road. For this purpose, the system 200 is caused to identify the right edge of left lane marking and left edge of the right lane marking that may facilitate in determining the safe region.
[0032] In an embodiment, the system 200 is caused to convert the captured images into respective gray scale images, and perform further processing on the gray scale images for lane detection. Hereinafter, the gray scale image may be referred to as an image. In an embodiment, the system 200 is caused to generate an edge image corresponding to one or more images (or gray scale images) of a road having lane markings captured by the cam-era/media sensor. The edge image distinctly illustrates pair of inbound edges corresponding to the lane markings. Herein, the edges of the lane markings represent the left edge and the right edge of each of the lane markings. An example of an image (gray scale image corresponding to a captured image) and a corresponding edge image are illustrated and explained further with reference to FIGS. 3A and 3B.
[0033] In an example embodiment, the system 200 is caused to identify the edges in the image by using an edge detection filter. An example of an edge detection filter may in-clude a Canny edge detector. The detection of inbound edges corresponding to a single lane may include detecting rising and trailing edges of lane markings of said lane. Herein, the ris-ing edge of the lane marking may correspond to the edge at which a transition from low pixel intensity to high pixel intensity is observed. Also, the trailing edge of the lane marking in-cludes that edge of the lane marking at which a transition from high pixel intensity to low pixel intensity is observed. In an embodiment, the system 200 is caused to identify the rising and/or trailing edges of the lane markings based on the sign of the gradient along with the magnitude at each pixel. In an embodiment, the following pseudo code may be performed for extraction of edges from lane markings:
program ExtractEdges(In: InputImage,Output: EdgeDetails)
begin
DetectGradientImages()
DetectTrailingEdges()
DetectRisingEdges()
CombineTrailingRisingEdges()
end.

[0034] The system 200 is caused to select the pair of inbound edges of the lane mark-ings upon selecting the rising and trailing edges from the edge image. For example, the system 200 may be caused to select the right edge of the left lane marking and the left edge of the right lane marking as the pair of inbound edges from the edge image. Herein, it will be noted that due to selection of only one edge from each of the lane markings, the number of edges to be considered for further processing are reduced, and hence the complexity in further processes to select the accurate inbounds of the lane markings is reduced.
[0035] In an embodiment, the system 200 is caused to extract a plurality of horizontal image strips from the image of the road. Herein, the image strips from amongst the plurality of horizontal image strips includes a subset of the lane markings. An example of an image of a road and a corresponding image strip are illustrated and described further with reference to FIG. 4. In an example embodiment, the image strips are selected along the height of the image. It will be noted herein that the image is a gray scaled image of the image of the road that is captured by the camera/media sensor. In an example embodiment, selection of the strips along the height of the image facilitates in supporting dashed lane markings and curved lane markings (for instance, marking on a curved road as illustrated in FIG. 1). In an example em-bodiment, the selection of the strips and height of each strip are configured in such a way that lane marking at different distances ranging from nearer to farther distance from the cam-era/system 200 are covered. Also, such selection of the image strips facilitates in minimizing the larger variations in the views farther from the image (i.e. larger strips in nearer view and smaller strips in farther view). An example of variation in sizes of image strips along the height of the image is illustrated further with reference to FIG. 5, where the left lane marking is a solid line and right lane marking is a dashed line.
[0036] In an example embodiment, the plurality of image strips includes spatial sub-sets of the lane markings. In an example embodiment, the spatial subset of the lane marking may be present in most of the image strips and missing in few of the image strips. Herein, the spatial subset of the lane marking present in an image strip may be referred to as a ‘lane mask’. In particular, a lane mask of the plurality of lane masks in a horizontal image strip is indicative of a subset of the lane markings in the horizontal image strip. In an embodiment, the system 200 is caused to extract the plurality of lane masks from the plurality of image strips. Herein, it will be noted that the images are captured by a forward facing camera mounted on a vehicle, and during the motion of the vehicle both the lane intensity profile and environmental illumination profile may be influenced. In addition, shadows of other objects such as vehicles, trees and so on influence the detection accuracy. Accordingly, the system 200 is caused to utilize adaptive thresholding of intensity of the subset of the lane markings in the plurality of horizontal image strips, to separate the lane masks from background in the plurality of horizontal image strips. Additionally, the system 200 is caused to validate the extracted lane masks based on a structure information of the extracted lane markings, as discussed below.
[0037] In an embodiment, applying the adaptive thresholding includes adaptive thre-sholding of intensity of lane markings in the at least one image. Since the road intensities vary across the width and length of the road, considering small strips across the length of the road may facilitate in determining threshold values (of intensity) in each of the horizontal image strips to segregate the bright regions from the road background in said image strips. Said de-termination of the threshold value to segregate the bright regions from the road background may include adaptive binarization which may utilize an intensity profile or a histogram of the gray values in the horizontal image strips, and further determination of the best suitable thre-shold value to identify potential candidate image masks for lane marks with minimal noise. The highest peak in the histogram represents the intensity value of the lane region and the position of the peak varies according to the intensity variation of the road. The next highest peak of the histogram is observed to be around the intensity of the lane marking. In an embo-diment, the system 200 is caused to eliminate the local peaks in the histogram that may occur due to image noises. In an embodiment, the system 200 may be caused to eliminate the local peaks by performing the smoothing operation on the histogram to produce the low pass fil-tered envelop of the intensity profile. Further, a minima between the two maxima are deter-mined from the smooth histogram. An example histogram of an image strip is further illu-strated and described with reference to FIG. 6A, and a smoothed histogram is described with reference to FIG. 6B. The system 200 is caused to adaptively determine an intensity threshold to be taken for binarization looking for the lane-specific signature threshold by deriving mi-nima between peak (first maxima) and following peak (second maxima) of the intensity pro-file from the intensity profile. Said adaptively determination of the intensity threshold is termed as Min Between Max Threhsolding (MBMT). The method of MBMT for adaptively thresholding the intensity of lane markings in an image is further described in the Indian pa-tent application number 1594/MUM/2015 titled “Method and System of Ground Truth De-termination and Validation of Lane Detection Warning” and is incorporated herein by refer-ence.
[0038] In an embodiment, the lane masks determined based on the adaptive thresholding of the plurality of image strips may include one or more non-lane marks also in addition to valid lane masks. Such non-lane masks may be caused due to shadows, local intensity variations of the road, vehicles in the scene, and other such factors. In an embodiment, the system 200 is caused to remove the non-lane masks from the potential lane masks (or the plurality of lane masks) based at least on the structure information and intensity information of lane masks in the respective horizontal image strips. The structure information of the lane masks may include shape and size of the lane masks in respective image strips. The lane markings in a smaller region or lane masks in an image strip may exhibit characteristics of a square or rectangle independent of curvy shape thereof in longer distance along the road. An example of a lane mask in an image strip is illustrated and described further with reference to FIG. 4. Also, the intensity information may include intensity values associated with the pixels of the lane markings.
[0039] In an embodiment, the system 200 is caused to select a plurality of valid lane masks from amongst the plurality of lane masks based at least on shape, size and intensity variation of the lane masks in the plurality of image strips. For example, the system 200 may be caused to extract potential lane masks from based on the intensity of the pixels associated with the lane masks, and further eliminate those masks from the potential lane masks that do not belong to lane markings. In an embodiment, the system 200 is caused to select a set of image strips having the valid lane masks therein. The system 200 may further be caused to combine/join such image strips to reconstruct the image of the road having lane markings such that the reconstructed image includes the plurality valid lane masks. An example of a reconstructed image is described herein with reference to FIG. 4.
[0040] In an embodiment, the system 200 is caused to determine determining an over-lap between the pair of inbound edges and the plurality of valid lane masks to identify a plu-rality of edge segments of the pair of inbound edges corresponding to the lane markings. In an embodiment, the overlap is determined by determining for each pixel associated with the pair of inbound edges, a presence of a corresponding pixel on one of the plurality of horizon-tal image strips. Herein, it will be noted that a plurality of portions or edge segments of the pair of inbound edges includes a curvature information. The curvature information of an edge segment of the pair of inbound edges at any point may include indicative of curvature of the edge segment with respect to the lane markings. In other words the edge information may include an angle by which said point or patch may deviate from being linear or in a straight line. Since, the lane markings exhibit spatial collinearity across different regions of the road, the system 200 may be caused to group selected lane edge portions/segments based on their collinearity/ contiguity and a distance criteria. On the grouped edges, a smoothing is performed to minimize the local noises. The system 200 is further caused to compare each edge portion with other short listed edges to determine whether said edge portion belong to same lane marking (left lane marking or right lane marking). The system 200 is further caused to select or obtain a set of valid edge segments from amongst the pair of inbound edges (for example, the left lane marking and the right lane markings) that represents the LEFT and RIGHT lane marking based on location thereof in the image. An example pseudo code for edge grouping for LEFT lane is shown below. Same steps can be followed for identifying edge group for RIGHT lane.
program GroupEdges(In: Shortlisted Edges, Output: LaneDetails)
begin
Create LEFT_Group
Add first LEFT Edge to the group
For each LEFT edge from the edge list
FitLine()
FindContiguityWithGroup()
FindDistanceWithGroup()
If(Contiguity==TRUE &&
Distance<=DIST_THD)
MergeCurEdgeToGrp
end
end
Smooth LEFT_Group
end

[0041] In an embodiment, the edge segments that are determined to be overlapping with the lane masks are considered as lane bounds or valid edge segments. In an example embodiment, the system 200 is caused to accommodate the boundary cases where the mask is enlarged by performing the morphological dilation operation which enhances the boundaries of the mask. The morphological dilation operation may include expanding boundaries of the lane masks by a small amount (for example, one pixel on all sides) so as to accommodate the variations in rate of intensity changes at boundaries of lane mark and road under practical en-vironmental conditions.
[0042] In certain scenarios, presence of vehicles and other objects in the scene may cause unintended edges that are connected to the edges of lane marking, thereby impacting the accuracy of the detection of the lane marking. The system 200 may be caused to separate the unintended edges from the lane edges so as to improve the quality of the edges passed for further processing. The unintended edges may connect with the lane edges by forming cor-ners. A corner on the edge is the point which exhibits a high curvature. An example of such corners being formed due to unintended edge is illustrated and described further with refer-ence to FIG. 7.
[0043] In an embodiment, the system 200 is caused to identify the unintended edges based on the curvature information so as to determine presence of corners. The system 200 may further be caused to separate the unintended edges at the points where corner is detected, and obtain a set of valid edge segments from the pair of inbound edges. Herein, the curvature information may include a geometrical angle between multiple points on the lane edge and the lane markings. In an embodiment, the system 200 is cased to determine whether a variation between the angles of curvature for the set of points located on each of the edge segment of the plurality of edge segments is greater than or equal to a threshold value of angle. This is done by considering three equally spaced points, for instance, Pi-n (xi-n, yi-n), Pi (xi, yi), and Pi+n(xi+i, yi+1) on the contour of an edge. Assuming y-values are equally spaced, then (xi-n + xi-n -2xi) is computed for different set of three consecutive points on the contour of the lane edge. The system 200 may be caused to detect the set of points forming a potential corner by applying the condition as described below in equation (1)

Here, ‘n’ represents a distance between consecutive points i, and
CThd represents a threshold value of curvature.
[0044] In an embodiment, the values of ’n’ and ’CThd’ are optimally selected based on the image resolution and to avoid the local curvature variations due to image digitization. In an example embodiment, the value of ’CThd’ may be selected in such a way that any cur-vature above 30degree may be considered as corner point. In an embodiment, the local max-ima is detected from the set of potential corner points to detect a corner point. An example pseudo code for detecting corner is as shown below:
program SeparateUnintendedEdges(In: Edge, Output: SeparatedEdges)
begin
ExtractContourOfTheEdge
At each point Pi(xi,yi)
FindAngleMeasureUsingThreePts()
FindPotentialConrners()
end
DetectTheLocalMaxima()
SplitTheContourAtCornerPts
end

[0045] In an example embodiment, the system is caused to mark one or more edge segments of the plurality of edge segments as a noisy edge segments on determination of said variation being greater than or equal to the threshold value of angle. The system may further be caused to eliminate the noisy edge segments from the plurality of edge segments to obtain the set of valid edge segments.
[0046] FIG. 3A illustrates an image 310 of a road, FIG. 3B illustrates an edge image 350 corresponding to the image 310 of FIG 3A, and FIG. 3C illustrates image 360 including inbound edges of lane markings of the edge image 350 of FIG. 3B, in accordance with an ex-ample embodiment.
[0047] The visual representation of a road by means of the lane detection facilitates in performing various functions such as supporting driver assistance and safety features such as Lane Departure Warning (LDW), Lane Keep Assistance (LKA) and Forward Collision Alert (FCA). As discussed with reference to FIG. 2, an image of a road, for example a road 312 may be captured by a forward facing camera installed on or inside a vehicle. For instance, the camera may be installed behind the rear view mirror of the vehicle so as to enable the camera to capture image of the road. In an embodiment, such cameras may embody clear sensors or color filtered sensors. For Driver Assistance applications, a vision sensor, i.e., camera is mounted in front of the vehicle to capture the road images. The camera is an assembly of a CMOS / CCD sensor of size VGA or HD used along with a lens of horizontal field of view (FOV) around 45 degree and vertical field of view of around 30 degree. In an embodiment, the image captured by the camera may be a colored image, and may be converted to a gray scale image prior to performing further processing thereon for lane detection. For instance, the image 310 illustrated in FIG. 3A represents a gray scale image of the road 312.
[0048] As illustrated herein, the image 310 of the road 312 may include lane markings 314 and 316. The lane markings may vary in color for different countries. For example, the lane marking are in white and yellow color for India, USA, and Europe; and blue colored for South Korea. Moreover, the lane marking may be different width and shapes such as solid lane markings and dashed lane markings. In an embodiment, the disclosed system (for exam-ple, the system 200 of FIG. 2) is caused to detect the edges, and more specifically, the in-bound edges of the lane markings in the image.
[0049] Herein the lane markings 314, 315 may be left and right side lane markings on the road. Herein, the term ‘left’ and ‘right’ may be used with reference to the vehicle, where the lane marking towards the left side of vehicle may be referred to as left lane marking, and the lane marking towards the right side of vehicle may be referred to as right lane marking. For the sake of clarity of description and distinguishing the pair of lane markings from each other, the pair of lane markings may be referred to as the first lane marking and a second lane marking.
[0050] Referring to FIG. 3B, the edge image 350 corresponding to the image 310 is illustrated herein. The edge image 350 may include the edges respectively of lane markings, for example the lane markings 314 and 316. For example, referring collectively to FIGS. 3B and 3C, the lane markings 314 include edges 314a and 314b, and the lane marking 316 in-cludes edges 316a and 316b. As described with reference to FIG. 2, the system 200 is caused to detect the inbound edges of the lane markings for example, lane markings. The inbound edges refer to the edges of the lane markings that may define a safe zone for driving. Herein, the inbound edges are assumed to be left edge of the right lane marking and right edge of the left lane marking. For example, the inbound edges of the lane edges 314b and 316a may form the inbound edges of the lane markings 314 and 316. As is already described with reference to FIG. 2, the edges of lane markings may be identified by using edge detection filters.
[0051] FIG. 4 illustrates an image 410 of a road, an example horizontal image strip 420 and a lane mask 430 for the road, and a reconstructed image 450 corresponding to the image 410, in accordance with an example embodiment. The image 410 may be a gray scale image of the image captured by a camera. The image 410 of the road is shown to include a pair of lane markings, for example a lane marking 412 and 414. As described with reference to FIG. 2, the system 200 is caused to extract a plurality of horizontal image strips from the image, for example the image 410 of the road. An example horizontal image strip 420 is illustrated herein. An example of extracting horizontal image strips from the image is described further with reference to FIG. 5.
[0052] Herein, the horizontal image strip 420 includes a lane masks or spatial subset of the lane markings. For instance, the horizontal image strip 420 is shown to include lane masks 422, 424 of lane markings 412 and 414, respectively. Such spatial subsets 422, 424 of lane markings 412, 414, are referred to as lane masks. An example of lane masks corresponding to the lane marking portions of lane subsets 422, 424 is illustrated in image portion 430.
[0053] The image portion 430 is shown to include lane masks 432, 434 corresponding to the portion of lane markings 422, 424 in the horizontal image strip 420. In an embodiment, the lane masks may be detected or generated by applying adaptive thresholding on the image strip, for example the image strip 420. An example of applying adaptive thresholding is de-scribed with reference to FIGS. 6A and 6B.
[0054] In an embodiment, the lane masks generated by applying adaptive thresholding to the plurality of lane masks may be combined together to reconstruct an image, for instance, the image 450. In an example embodiment, the plurality of lane masks may be validated to obtain or select a plurality of valid lane masks from amongst the plurality of lane masks, and the plurality of valid lane masks may be utilized for reconstructing the image. As is unders-tood, the reconstructed image 450 represents a region of interest (ROI) image with multiple thresholded strips, for instance the horizontal strip or image portion 430.
[0055] FIG. 5 illustrates an example representation 500 for extracting a plurality of ho-rizontal image strips corresponding to an image of a road, in accordance with an example embodiment. As illustrated herein, the image represents a road region having lane markings 502, 504. The region of the road captured in an image is partitioned into a plurality of hori-zontal strips along the height of the image. For example, the image herein is partitioned by a plurality of horizontal strips such as strips 506, 508, and 510 along the height of the image. In an embodiment, the selection of the strips and the height of each strip are configured in such a way that lane marking at different distances ranging from nearer to farther distance from the camera are covered. In addition, such a configuration of the strips minimizes larger variations in the farther view, i.e. larger strips are configured in nearer views and smaller stripes in farther view. For example, the strip 506 which is closest to the camera is larger as compared to the strip 508 which is at a farther distance from the camera. In an embodiment, the described configuration facilitates in suitably capturing the strips from the image in which portions of the lane marking are dashed. As illustrated in FIG. 5, the lane marking 502 is a sol-id line and the lane marking 504 is a dashed line.
[0056] In an example embodiment, for each of the strips that are extracted from the image, potential candidates for lane masks may be extracted. In an embodiment, the potential candidates for lane masks can be extracted by applying adaptive thresholding to the plurality of horizontal image strips. As discussed with reference to FIG. 2, the system 200 implements adaptive binarization which uses histogram of the gray values in the horizontal image stripe and determines most suitable threshold value to identify potential candidate masks for lane marks with minimal noise. An example describing use of histogram (or an intensity profiles) of the horizontal image stripe is described further with reference to FIGS. 6A and 6B.
[0057] FIGS. 6A and 6B illustrate an example an intensity profiles or a histogram 610 for a horizontal image strip associated with an image, and a processed histogram 650 cor-responding to the horizontal image strip, in accordance with an example embodiment. A his-togram of an image is a graphical representation of distribution of pixel intensity across an image, for instance image of a road. In practice, the pixel intensities may vary across the width and length of the road. For example, the pixel intensity in the regions of lane markings may be higher that the pixel intensities in the other regions of the road. Thus, the histogram 610 corresponding to a horizontal image strip may indicate variation in pixel intensities of image pixels in the image strip. For example, the histogram 610 corresponding to an image strip, for instance, image strip 420 (FIG. 4) is shown to include a variation in the pixel intensi-ties across the road. In an embodiment, the variation in the pixel intensities across the road in the histogram can facilitate in determining a threshold value to segregate the bright regions (indicating lane markings) from the road background.
[0058] The highest peak 612 (or maxima) in the histogram 610 represents an intensity value of the lane region and the position of the peak varies according to the intensity variation of the road. The next highest peak 614 (subsequent maxima) of the histogram 610 is observed to be around the intensity of the lane marking. The local peaks such in the histogram due to image noises are eliminated by performing the smoothing operation on the histogram 610 to produce the low pass filtered envelop of the intensity profile or a processed histogram 650, as illustrated in FIG. 6B. Further, minima 616, for example the minima 616 between the two maxima 612, 614 is determined from the smooth histogram 650. The minima 616 between the two maxima 612, 614 may be indicative of an intensity threshold for the respective horizontal image strip.
[0059] FIG. 7A, 7B and 7C illustrates an example scenario for removing unintended edge portions from lane edges, in accordance with an example embodiment. For example, FIG. 7A illustrates an image 710 of a road having lanes such as lanes 712, 714. The image 710 may be a gray scale image. The image 710 also shows a vehicle 712 driving on the road such that shadow 714 of the vehicle 712 is overlapping with one of the lanes 712. Presence of unin-tended objects such as other vehicles, tress and so on in a scene where lane detection is to be performed can cause unintended edges that are connected to the edges of lane marking. For instance, presence of the vehicle 716 and shadow 718 of the vehicle 716 may cause unin-tended edges. Such unintended edges can connect with the lane edges by forming corners. An example corner being formed due to unintended edges is illustrated in FIG. 7B.
[0060] Referring to FIG. 7B, the corner being formed due to the unintended edge 722 formed by the vehicle shadow is marked as 724. In particular, edges of the car shadow are connected to the lane marking edges. In an embodiment, the system 200 (FIG. 2) is caused to identify and separate the edges at the points where such corner is detected. In an embodiment, the system 200 may utilize the curvature information or the angular spread at the edge to identify the corner. For example, the system may detect segmentation point in the edge based on the curvature information. In an embodiment, the system may consider those segment points as corner for which the curvature or angular spread is more than a threshold value of curvature, as described with reference to FIG.2. For example, herein the curvature for point 724 on the edge is determined to be more than the threshold value of curvature.
[0061] FIG. 8 illustrates a flow diagram of a method 800 for lane detection, in accor-dance with the present disclosure. In an embodiment, for lane detection a system (or a lane detection system) may be embodied in a vehicle. Example of such a system may include the system 200 (FIG. 2). In an embodiment, the system may include an image sensor (or a camera) for capturing the image of road on which vehicle may drive. Alternatively, the system 200 may be in communication with the camera for receiving input images from the camera, and process the input images so as to detect a safe zone within the lane markings for driving on the road. In an embodiment, the system may be caused to capture images of the road in real-time, and process the captured images. In an embodiment, prior to processing, the images may be preprocessed by for instance, converting the captured images into gray scale image.
[0062] At 802, an edge image corresponding to the image of the road is generated. The edge is representative of edges of lane markings on the road. In an embodiment, the edge image may include a pair of inbound edges corresponding to the lane markings. Examples of an edge image including the lane markings and inbound edges are illustrated with reference to FIG. 3B
[0063] At 804, a plurality of horizontal image strips are extracted from the image such that the image strips include subset of the lane markings. An example of multiple image strips being extracted from the image is described and illustrated with reference to FIG. 4. In an embodiment, the a plurality of image strips may be selected from the image, and based on adaptive thresholding of intensity of lane markings in each of image strips, lane marking are extracted.
[0064] At 806, a plurality of valid lane markings are selected from amongst the plural-ity of potential lane markings. In an embodiment, the plurality of valid lane markings are se-lected based at least on shape, size and intensity variation of the subset of the lane markings in the respective image strips.
[0065] At 808, an overlapping between the inbound edges in the edge image and the subsets of the valid lane markings in the plurality of image strips is determined. Said overlap-ping facilitates in determining a plurality of portions of the pair of inbound edges belonging to the lane markings. In an embodiment, the plurality of portions of the pair of inbound edges are associated with a curvature information.
[0066] At 810, valid edge segments of the pair of inbound edges are obtained by ex-cluding noise from the plurality of portions of the pair of inbound edges based on the curva-ture information. In an embodiment, the noise in the plurality of portions of the inbound edges unintended edge portions that may be formed due to shadows or other objects such as vehicles, trees, and so on.
[0067] Various embodiments of the disclosure provide method and system for lane de-tection in varied road conditions including unconstrained environment, curvature, shape, and noise deviation. The proposed method and system utilizes adaptive figure extraction, shape-based outlier rejection and empirical noise removal for performing lane detection.
[0068] The written description describes the subject matter herein to enable any person skilled in the art to make and use the embodiments. The scope of the subject matter embodiments is defined by the claims and may include other modifications that occur to those skilled in the art. Such other modifications are intended to be within the scope of the claims if they have similar elements that do not differ from the literal language of the claims or if they include equivalent elements with insubstantial differences from the literal language of the claims.
[0069] It is, however to be understood that the scope of the protection is extended to such a program and in addition to a computer-readable means having a message therein; such computer-readable storage means contain program-code means for implementation of one or more steps of the method, when the program runs on a server or mobile device or any suitable programmable device. The hardware device can be any kind of device which can be programmed including e.g. any kind of computer like a server or a personal computer, or the like, or any combination thereof. The device may also include means which could be e.g. hardware means like e.g. an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), or a combination of hardware and software means, e.g. an ASIC and an FPGA, or at least one microprocessor and at least one memory with software modules located therein. Thus, the means can include both hardware means and software means. The method embodiments described herein could be implemented in hardware and software. The device may also include software means. Alternatively, the embodiments may be implemented on different hardware devices, e.g. using a plurality of CPUs.
[0070] The embodiments herein can comprise hardware and software elements. The embodiments that are implemented in software include but are not limited to, firmware, resident software, microcode, etc. The functions performed by various modules described herein may be implemented in other modules or combinations of other modules. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can comprise, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
[0071] A data processing system suitable for storing and/or executing program code will include at least one processor coupled directly or indirectly to memory elements through a system bus. The memory elements can include local memory employed during actual execution of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution.
[0072] Input/output (I/O) devices (including but not limited to keyboards, displays, pointing devices, etc.) can be coupled to the system either directly or through intervening I/O controllers. Network adapters may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
[0073] A representative hardware environment for practicing the embodiments may include a hardware configuration of an information handling/computer system in accordance with the embodiments herein. The system herein comprises at least one processor or central processing unit (CPU). The CPUs are interconnected via system bus to various devices such as a random access memory (RAM), read-only memory (ROM), and an input/output (I/O) adapter. The I/O adapter can connect to peripheral devices, such as disk units and tape drives, or other program storage devices that are readable by the system. The system can read the inventive instructions on the program storage devices and follow these instructions to execute the methodology of the embodiments herein.
[0074] The system further includes a user interface adapter that connects a keyboard, mouse, speaker, microphone, and/or other user interface devices such as a touch screen device (not shown) to the bus to gather user input. Additionally, a communication adapter connects the bus to a data processing network, and a display adapter connects the bus to a display device which may be embodied as an output device such as a monitor, printer, or transmitter, for example.
[0075] The foregoing description of the specific implementations and embodiments will so fully reveal the general nature of the implementations and embodiments herein that others can, by applying current knowledge, readily modify and/or adapt for various applications such specific embodiments without departing from the generic concept, and, therefore, such adaptations and modifications should and are intended to be comprehended within the meaning and range of equivalents of the disclosed embodiments. It is to be understood that the phraseology or terminology employed herein is for the purpose of description and not of limitation. Therefore, while the embodiments herein have been described in terms of preferred embodiments, those skilled in the art will recognize that the embodiments herein can be practiced with modification within the spirit and scope of the embodiments as described herein.
[0076] The preceding description has been presented with reference to various embodiments. Persons having ordinary skill in the art and technology to which this application pertains will appreciate that alterations and changes in the described structures and methods of operation can be practiced without meaningfully departing from the principle, spirit and scope. ,CLAIMS:WE CLAIM:

1. A computer implemented method for lane detection comprising:
generating an edge image corresponding to at least one image of a road having lane markings, the edge image comprising a pair of inbound edges corresponding to the lane markings;
extracting a plurality of lane masks from a plurality of horizontal image strips of the at least one image, a lane mask of the plurality of lane masks in a horizontal image strip being indicative of a subset of the lane markings in the horizontal image strip, wherein the plurality of lane masks are extracted based on an adaptive thresholding of intensity of the subset of the lane markings in the plurality of horizontal image strips;
selecting a plurality of valid lane masks from amongst the plurality of lane masks based at least on structure information and intensity information of the plurality of lane masks in the plurality of horizontal image strips;
determining an overlap between the pair of inbound edges and the plurality of valid lane masks to identify a plurality of edge segments of the pair of inbound edges corresponding to the lane markings, wherein the plurality of edge segments of the pair of inbound edges comprises a curvature information, the curvature information of an edge segment being indicative of curvature of the edge segment with respect to the lane markings; and
obtaining a set of valid edge segments of the pair of inbound edges based on the curvature information of the plurality of edge segments, wherein the set of valid edge segments facilitates in detecting the lane markings.

2. The method as claimed in claim 1, wherein the at least one image comprises a gray scale image of the road with the lane markings.

3. The method as claimed in claim 1, wherein the adaptive thresholding of the intensity of a subset of the lane markings in a horizontal image strip of the plurality of horizontal image strips comprises:
determining an intensity profile of the horizontal image;
deriving, from the intensity profile, a minima between a maxima and a subsequent maxima; and
adaptively determining an intensity threshold for the horizontal image strip based on the minima between the maxima and the subsequent maxima.

4. The method as claimed in claim 1, wherein determining the overlap between the pair of inbound edges and the plurality of valid lane masks comprises determining for each pixel associated with the pair of inbound edges, a presence of a corresponding pixel on one of the plurality of horizontal image strips.

5. The method as claimed in claim 1, wherein the curvature information of the edge segment comprises angles of curvature for a set of points located on the edge segment.

6. The method as claimed in claim 5, wherein obtaining the set of valid edge segments comprises:
determining whether a variation between the angles of curvature for the set of points located on each of the edge segment of the plurality of edge segments is greater than or equal to a threshold value of angle;
marking one or more edge segments of the plurality of edge segments as a noisy edge segments on determination of said variation being greater than or equal to the threshold value of angle; and
eliminating the noisy edge segments from the plurality of edge segments to obtain the set of valid edge segments.

7. The method as claimed in claim 1, wherein the plurality of horizontal image strips are of different heights.

8. A computer implemented system for lane detection, the system comprising:
at least one memory; and
at least one processor, the at least one memory coupled to the at least one processor wherein the at least one processor is capable of executing programmed instructions stored in the at least one memory to:
generate an edge image corresponding to at least one image of a road having lane markings, the edge image comprising a pair of inbound edges corresponding to the lane markings,
extract a plurality of lane masks from a plurality of horizontal image strips of the at least one image, a lane mask of the plurality of lane masks in a horizontal image strip being indicative of a subset of the lane markings in the horizontal image strip, wherein the plurality of lane masks are extracted based on an adaptive thresholding of intensity of the subset of the lane markings in the plurality of horizontal image strips,
select a plurality of valid lane masks from amongst the plurality of lane masks based at least on structure information and intensity information of the plurality of lane masks in the plurality of horizontal image strips,
determine an overlap between the pair of inbound edges and the plurality of valid lane masks to identify a plurality of edge segments of the pair of inbound edges corresponding to the lane markings, wherein the plurality of edge segments of the pair of inbound edges comprises a curvature information, the curvature information of an edge segment being indicative of curvature of the edge segment with respect to the lane markings, and
obtain a set of valid edge segments of the pair of inbound edges based on the curvature information of the plurality of edge segments, wherein the set of valid edge segments facilitates in detecting the lane markings.

9. The system as claimed in claim 8, wherein to perform the adaptive thresholding of the intensity of a subset of the lane markings in a horizontal image strip of the plurality of horizontal image strips, the at least one processor is further configured by the instructions to:
determine an intensity profile of the horizontal image strip;
derive, from the intensity profile, a minima between a maxima and a subsequent maxima; and
adaptively determine an intensity threshold for the horizontal image strip based on the minima between the maxima and the subsequent maxima.

10. The system as claimed in claim 8, wherein to determine the overlap between the pair of inbound edges and the plurality of valid lane masks, the at least one processor is further configured by the instructions to determine for each pixel associated with the pair of inbound edges, a presence of a corresponding pixel on one of the plurality of horizontal image strips.

11. The system as claimed in claim 8, wherein the curvature information of the edge segment comprises angles of curvature for a set of points located on the edge segment.

12. The system as claimed in claim 11, wherein to obtain the set of valid edge segments, the at least one processor is further configured by the instructions to:
determine whether a variation between the angles of curvature for the set of points located on each of the edge segment of the plurality of edge segments is greater than or equal to a threshold value of angle;
mark one or more edge segments of the plurality of edge segments as a noisy edge segments on determination of said variation being greater than or equal to the threshold value of angle; and
eliminate the noisy edge segments from the plurality of edge segments to obtain the set of valid edge segments.

13. The system as claimed in claim 8, wherein the plurality of horizontal image strips are of different heights.

14. The system as claimed in claim 8, further comprising at least one image sensor coupled to the at least one processor and the at least one memory, wherein the at least one image sensor facilitates in capturing the at least one image.

Documents

Application Documents

# Name Date
1 Drawing [09-10-2015(online)].pdf 2015-10-09
2 Description(Complete) [09-10-2015(online)].pdf 2015-10-09
3 Assignment [09-10-2015(online)].pdf 2015-10-09
4 REQUEST FOR CERTIFIED COPY [19-01-2016(online)].pdf 2016-01-19
5 Form 3 [15-02-2017(online)].pdf 2017-02-15
6 Request For Certified Copy-Online.pdf 2018-08-11
7 Form 2.pdf 2018-08-11
8 ABSTRACT1.jpg 2018-08-11
9 2079-MUM-2015-Power of Attorney-270715.pdf 2018-08-11
11 2079-MUM-2015-Correspondence-270715.pdf 2018-08-11
12 2079-MUM-2015-CORREPONDENCE (2-7-2015).pdf 2018-08-11
13 1998-MUM-2015-FORM 1(2-7-2015).pdf 2018-08-11
14 1998-MUM-2015-CORREPONDENCE(2-7-2015).pdf 2018-08-11
15 2079-MUM-2015-FER.pdf 2020-01-29
16 2079-MUM-2015-OTHERS [28-07-2020(online)].pdf 2020-07-28
17 2079-MUM-2015-FER_SER_REPLY [28-07-2020(online)].pdf 2020-07-28
18 2079-MUM-2015-COMPLETE SPECIFICATION [28-07-2020(online)].pdf 2020-07-28
19 2079-MUM-2015-CLAIMS [28-07-2020(online)].pdf 2020-07-28
20 2079-MUM-2015-US(14)-HearingNotice-(HearingDate-23-11-2022).pdf 2022-11-03
21 2079-MUM-2015-FORM-26 [17-11-2022(online)].pdf 2022-11-17
22 2079-MUM-2015-FORM-26 [17-11-2022(online)]-1.pdf 2022-11-17
23 2079-MUM-2015-Correspondence to notify the Controller [17-11-2022(online)].pdf 2022-11-17
24 2079-MUM-2015-Written submissions and relevant documents [05-12-2022(online)].pdf 2022-12-05
25 2079-MUM-2015-PatentCertificate14-03-2023.pdf 2023-03-14
26 2079-MUM-2015-IntimationOfGrant14-03-2023.pdf 2023-03-14

Search Strategy

1 search1AE_07-01-2021.pdf
1 search_17-01-2020.pdf
2 search1AE_07-01-2021.pdf
2 search_17-01-2020.pdf

ERegister / Renewals

3rd: 25 May 2023

From 28/05/2017 - To 28/05/2018

4th: 25 May 2023

From 28/05/2018 - To 28/05/2019

5th: 25 May 2023

From 28/05/2019 - To 28/05/2020

6th: 25 May 2023

From 28/05/2020 - To 28/05/2021

7th: 25 May 2023

From 28/05/2021 - To 28/05/2022

8th: 25 May 2023

From 28/05/2022 - To 28/05/2023

9th: 25 May 2023

From 28/05/2023 - To 28/05/2024

10th: 28 May 2024

From 28/05/2024 - To 28/05/2025

11th: 23 May 2025

From 28/05/2025 - To 28/05/2026