Abstract: A display device displays an image of the area around a subject vehicle including an image of the subject vehicle and threshold lines defining ranges of distance from the subject vehicle. The image of the area around the subject vehicle is changed in accordance with the positional relationship between the ranges of distance and other vehicles.
DESCRIPTION TITLE OF INVENTION
DRIVING SUPPORT APPARATUS AND DRIVING SUPPORT METHOD 5 TECHNICAL FIELD [0001]
The present invention relates to a driving support apparatus and driving support method for informing a driver of the approach of a moving object to a user's vehicle. 10 BACKGROUND ART [0002]
A technique in which a display device arranged on a user's
vehicle displays information about a moving object by using a
mark or letters has been proposed. The technique is used for
15 informing a driver of the approach of the moving object to the
user's vehicle.
For example, an information providing device disclosed in Patent Literature 1 identifies a vehicle that is likely to the approach to a user's vehicle. The device displays the 20 identified vehicle with an icon, changes a size of the icon in accordance with a distance between the user's vehicle and the identified vehicle, and changes a stripe pattern which is a background image of the icon, in accordance with a speed of the identified vehicle. Thereby, the driver can intuitively 25 recognize the approach of the identified vehicle to the user's vehicle. [0003]
Patent Literature 2 discloses a driving support apparatus
that displays a mark having an arrow shape representing either
30 degree of approaching of another target vehicle to a user's
2
vehicle or degree of leaving of another target vehicle from the user's vehicle when the target vehicle is located obliguely behind the user's vehicle.
In this apparatus, a guide line is displayed in addition 5 to the mark having the arrow shape. The guide line indicates the risk to the user's vehicle due to the target vehicle traveling on a lane adjacent to the user's vehicle when the user's vehicle tries to move to the adj a cent lane . For example, when the target vehicle is approaching to the user's vehicle
10 at a relatively high speed, the mark representing a movement of the target vehicle, which is a relatively long arrow toward the user's vehicle, is displayed such that the relatively long arrow is across the guide line. With such a display, the driver can recognize that lane-change of the user's vehicle is risky
15 because the target vehicle is approaching to the user ' s vehicle . CITATION LIST PATENT LITERATURES [0004] Patent Literature 1: Japanese Patent Application Publication
20 No. 2007-102577.
Patent Literature 2: Japanese Patent Application Publication No. 2008-15758. TECHNICAL PROBLEM [0005]
25 In the prior art, a displayed pattern of a moving object is changed in accordance with degree of approaching of the moving object to a user's vehicle. However, in the prior art, the user's vehicle is not displayed and a positional relationship between the user's vehicle and the moving object
30 is not indicated. Thus, there exists a problem that a driver
3
cannot intuitively recognize a situation around the user's vehicle including the positional relationship between the user's vehicle and the moving object. [0006] 5 For example, the device disclosed in Patent Literature 1 does not display the user's vehicle, but displays the icon indicating the identified vehicle that is likely to approach the user's vehicle. The device indicates the degree of approaching of the identified vehicle to the user's vehicle by
10 changing the size of the icon and changing the pattern of the background image. The driver can recognize the direction in which the identified vehicle is present, based on a location of the icon appearing on a display screen. However, the driver cannot intuitively recognize the situation around the user's
15 vehicle, because the driver needs to guess the positional relationship between the user's vehicle and the identified vehicle based on the change of the icon size and the change of the background image pattern. [0007]
20 The apparatus disclosed in Patent Literature 2 does not display the user's vehicle, similarly to the case of Patent Literature 1, and instead displays the arrow indicating the target vehicle and the guide line indicating the risk of the target vehicle to the user's vehicle. The driver can roughly
25 recognize the positional relationship between the user's vehicle and the target vehicle by using the guide line as a reference. However, the driver needs to guess the positional relationship between the user's vehicle and the target vehicle based on both the arrow indicating the target vehicle and the
30 guide line.
■1
[0008]
The present invention is made in order to solve the above-mentioned problem, and it is therefore an object of the present invention to provide a driving support apparatus and 5 driving support method with which a driver can intuitively recognize a situation around a user's vehicle including a positional relationship between the user' s vehicle and a moving object around the user's vehicle. SOLUTION TO PROBLEM
10 [0009]
In accordance with of the present invention, there is provided a driving support apparatus which includes: an acquiring unit configured to acquire a positional relationship between at least one moving object and one or more ranges of
15 distance from a user's vehicle, each of the one or more ranges of distance being a range in which a notification of approach of the moving object to the user's vehicle is to be provided; and an informing processor configured to cause a display device to display an image of nearby surroundings of the user's vehicle
20 including images of the user's vehicle and threshold lines that define the one or more ranges of distance, and configured to change the image of nearby surroundings of the user's vehicle in accordance with the positional relationship acquired by the acquiring unit.
25 ADVANTAGEOUS EFFECTS OF INVENTION [0010]
According to the present invention, the image of the nearby surroundings of a user's vehicle is displayed, including the images of the user's vehicle and the threshold lines
30 defining the one or more ranges of distance, each range of
5
distance being a range in which a notification of the approach of the moving object to the user's vehicle is to be provided. Thereby, the driver can intuitively recognize the situation around the user ' s vehicle including the positional relationship 5 between the user's vehicle and the moving object around the user's vehicle.
BRIEF DESCRIPTION OF DRAWINGS [0011]
Figs. 1A and IB are diagrams illustrating a configuration 30 of a driving support apparatus in accordance with Embodiment 1;
Fig. 2 is a flowchart illustrating operations of the driving support apparatus in accordance with Embodiment 1; Fig. 3 is a schematic diagram illustrating ranges of 15 distance in each of which a notification of the approach of another vehicle is to be provided;
Fig. 4 is a schematic diagram illustrating the ranges of distance when relative velocities between a user's vehicle and the target vehicles are smaller as compared to the case shown 20 in Fig. 3;
Fig. 5 is a schematic diagram illustrating the ranges of
distance when the relative velocities between the user's
vehicle and the target vehicles are larger as compared to the
case shown in Fig. 3;
25 Fig. 6 illustrates an image of the nearby surroundings
of a user's vehicle including images of the user's vehicle and
threshold lines arranged at the sides of the user's vehicle;
Fig. 7 illustrates an image of the nearby surroundings
of a user's vehicle including images of the user's vehicle and
30 threshold lines arranged in front of and at the back of the user's
6
vehicle;
Fig. 8 illustrates an image of the nearby surroundings of a user's vehicle including images of the user's vehicle and threshold lines arranged at the sides of, in front of and at 5 the back of the user's vehicle;
Fig. 9 is a diagram illustrating examples of icons indicating target vehicles, each icon being set corresponding to one vehicle type;
Fig. 10 is a diagram illustrating an exemplary image 10 displayed when the driving of another target vehicle to the user's vehicle is in a safe state;
Fig. 11 is a diagram illustrating an exemplary image displayed when the driving of another target vehicle to the user's vehicle is in an attention-requiring state; 15 Fig. 12 is a diagram illustrating an exemplary image displayed when the driving of another target vehicle to the user's vehicle is in an alert state;
Fig. 13 is a diagram illustrating an exemplary image displayed when the target vehicle further approaches the user's 20 vehicle from a position shown in Fig. 12;
Fig. 14 is a diagram illustrating a layer structure of a display screen in accordance with Embodiment 1;
Fig. 15 is a diagram illustrating an exemplary display of ranges of distance set at both sides of, in front of and at 25 the back of the user's vehicle;
Fig. 16 is a diagram illustrating an exemplary image
displayed when another target vehicle moves into an area A;
Fig. 17 is a diagram illustrating an exemplary image
displayed when another target vehicle moves into a zone D;
30 Fig. 18 is a diagram illustrating an exemplary image
7
displayed when another target vehicle moves into a zone E from the area A;
Fig. 19 is a diagram illustrating an exemplary image displayed when another target vehicle moves into the zone E from 5 the zone D;
Fig. 20 is a diagram illustrating an exemplary image displayed when another target vehicle moves into the area A and the zone D;
Fig. 21 is a diagram illustrating an exemplary image 10 displayed when another target vehicle moves into the area D from the area A after a state shown in Fig. 20;
Fig. 22 is a diagram illustrating an exemplary image displayed when two other target vehicles move into the area A and the zone D, respectively; 15 Fig. 23 is a diagram illustrating an exemplary image displayed when one target vehicle moves into the zone E from the area A after the state shown in Fig.22;
Figs . 24A to 24C illustrate exemplary images displayed when the user's vehicle is viewed from various viewpoints at 20 different vertical locations in a plan view;
Figs. 25A to 25C illustrate exemplary images displayed when the user's vehicle is viewed from various viewpoints at different horizontal locations in a plan view;
Fig. 26 is a block diagram illustrating a configuration 25 of a driving support apparatus in accordance with Embodiment 2;
Fig. 2 7 is a flowchart illustrating operations of the driving support apparatus in accordance with Embodiment 2; and Fig. 2 8 is a diagram illustrating an exemplary image 30 displayed in accordance with Embodiment 2.
s
DESCRIPTION OF EMBODIMENTS [0012]
Hereinafter, for detailed descriptions of the invention, various embodiments for carrying out the invention will be 5 explained with reference to the drawings.
Embodiment 1 Fig. 1A and Fig. IB are diagrams illustrating a configuration of a driving support apparatus 1 in accordance with Embodiment 1. Fig. 1A is a functional block diagram and 10 Fig. IB is a diagram of a hardware configuration. In Fig. 1A, the driving support apparatus 1 is an apparatus that informs a driver of the approach of a moving object such as another vehicle to a user's vehicle. The driving support apparatus 1 includes an information collector 10, a vehicle detector 11, 15 a distance determiner 12, a distance state calculator 13, a vehicle type identifier 14, an informing processor 15, and an informing device 16. [0013]
Among components of the driving support apparatus 1, the 20 vehicle detector 11, the distance determiner 12, the distance state calculator 13, the vehicle type identifier 14, and the informing processor 15 can be constituted, for example, by a processor 100 such as a Central Processing Unit (CPU) or a system LSI (Large-scale Integrated circuit). The processor 100 25 executes a program that is stored in a memory 101 shown in Fig. IB. Processing circuits may work together to function as above-mentioned components. [0014]
The information collector 10 collects information 30 indicating a situation around the user's vehicle, and transmits
9
the collected information to the vehicle detector 11.
The information collector 10 can be implemented, for example, using any of devices: imaging devices such as a camera and an infrared camera, and various sensors such as a ranging 5 sensor, a velocity sensor, a color sensor, and a millimeter wave radar.
Each of the camera and the infrared camera has an area for imaging that includes part of the surroundings of the user's vehicle and nearby areas at both sides of the user's vehicle, 10 and obtains image information indicating the nearby
surroundings of a user's vehicle. The ranging sensor and the millimeter wave radar obtain information on a distance between an object around the user's vehicle and the user's vehicle. The velocity sensor obtains information on a velocity of a moving 15 object around the user's vehicle. The color sensor detects a color of the object around the user's vehicle. [0015]
The vehicle detector 11 detects at least one moving object moving in an area around the user's vehicle. The vehicle 20 detector 11 identifies the moving object and a stationary object which exist around the user's vehicle, on the basis of information received from the information collector 10. Thus, the vehicle detector 11 obtains information on the moving object except for the stationary object. The vehicle detector 11 25 transmits the information on the moving object to the distance determiner 12 and the vehicle type identifier 14. Note that the information on the moving obj ect includes the position, size and shape of the moving object as well as detected information indicating white lines at both sides of a lane on which the user' s 30 vehicle is traveling, and image information indicating part of
10
areas at both sides of the user's vehicle. [0016]
For example, the vehicle detector 11 performs image analysis on a captured image of the nearby surroundings of a user's vehicle thereby to identify or detect an object as a moving object such as another vehicle traveling in front of, at the back of, or at one of the sides of the user's vehicle or a pedestrian walking on a shoulder of the road, and/or to identify or detect an object as a stationary object such as a building, a road sign, or a guard fence around the road on which the user's vehicle is traveling. Additionally, the vehicle detector 11 may determine whether the detected object .is the moving object in accordance with temporal change in distance between the detected object and the user's vehicle, which is measured by the ranging sensor.
Note that the vehicle detector 11 may use information collected by an information collector provided by another apparatus other than the driving support apparatus 1. In this case, the driving support apparatus 1 may not have the information collector 10. [0017]
The distance determiner 12 determines the distance between the user's vehicle and the moving object detected by the vehicle detector 11. For example, the distance determiner 12 determines the distance between the user's vehicle and the moving object in accordance with location information of the moving object received from the vehicle detector 11, and transmits the determined distance, as measurement information of the distance associated with the moving object, to the distance state calculator 13.
11
Additionally, the distance determiner 12 may extract distance information on the moving object detected by the vehicle detector 11 from among pieces of distance information on objects around the user's vehicle obtained by the ranging 5 sensor. [0018]
The distance determiner 12 calculates a lane width of a traffic route on which the user's vehicle is traveling, in accordance with the detected information indicating the white 10 lines at both sides of the lane, which is included in the information on the moving object.
In addition, the distance determiner 12 calculates a width of the user's vehicle by performing image analysis on the image information indicating part of areas at both sides of the 15 user's vehicle, which is included in the information on the moving object.
The measurement information of the distance associated with the moving object includes the measurement result of the distance between the moving object and the user's vehicle, as 20 well as measurement information of the lane width of the traffic route on which the user's vehicle is traveling and the width of the user's vehicle.
The width of the user's vehicle can be measured by the distance determiner 12 as initial setting when the driving 25 support apparatus 1 is installed on the user's vehicle.
Alternatively, the width of the user's vehicle may be stored in advance in the distance state calculator 13 instead of being measured by the distance determiner 12. [0019] 30 The distance state calculator 13 is a component that
12
corresponds to an acquiring unit in the invention, and performs a calculation on the basis of the measurement information obtained by the distance determiner 12 to thereby acquire the positional relationship between the moving object and the one 5 or more ranges of distance from the user's vehicle in which a notification of the approach of the moving object is to be provided.
For example, as setting information for the distance state calculator 13, distance thresholds, which define the
10 range of distance from the user's vehicle in which a
notification of approach of the moving object is to be provided, are set in advance for both sides of, in front of and at the back of the user's vehicle.
The distance state calculator 13 calculates the
15 positional relationship indicating whether the moving object around the user's vehicle moves into the range of distance, on the basis of the distance between the user's vehicle and the moving object, which is measured by the distance determiner 12, the lane width of the traffic route on which the user's vehicle
20 is traveling, the width of the user's vehicle, and the distance thresholds set in advance. Information indicating the positional relationship is transmitted to the informing processor 15 from the distance state calculator 13. [0020]
25 The vehicle type identifier 14 is a component
corresponding to a determiner in the invention, to identify a type of the moving object detected by the vehicle detector 11. For example, the vehicle type identifier 14 matches image information indicating the moving object received from the
30 vehicle detector 11, against pre-provided shape patterns of
13
moving objects, thus identifying the type of the moving object, and transmits an object type identification result to the informing processor 15. As types of moving object, a normal-sized vehicle such as a sedan and a Sport Utility Vehicle 5 (SUV), a large-sized vehicle such as a truck and a bus, a small-sized vehicle such as a motorcycle and a bicycle, a human being, and an animal are exemplified. [0021]
The informing processor 15 is a component that 10 corresponds to an informing processor in the invention. For example, the informing processor 15 causes a display device of the informing device 16 to display the image of the nearby surroundings of a user's vehicle.
The image of the nearby surroundings of a user's vehicle 15 includes images of the user's vehicle and threshold lines that define the one or more ranges of distance from the user' s vehicle (e.g. an attention zone and an alert zone explained later with Fig. 3) in each of which a notification of approach of the moving object is to be provided. Each threshold line is displayed as 20 a line image corresponding to the distance threshold that
defines a range of distance from the user's vehicle in which a notification of approach of the moving object is to be provided.
In addition, the informing processor 15 changes the image 25 of the nearby surroundings of the user's vehicle in accordance with the positional relationship acguired by the distance state calculator 13. For example, in accordance with the positional relationship acquired by the distance state calculator 13, the informing processor 15 may change colors of the attention zone 30 and the alert zone in the image of the nearby surroundings of
14
the user' s vehicle, and may causes an image of the moving object to appear in the image of the nearby surroundings of the user's vehicle, in accordance with the positional relationship acquired by the distance state calculator 13. [0022]
Moreover, the informing processor 15 selects, as an image to be displayed, the image of the moving object corresponding to the type of the moving object identified by the vehicle type identifier 14 from among pre-provided images of moving objects for each type.
Alternatively, the informing processor 15 may use a pre-provided image without relying on a type of moving object. In this case, the driving support apparatus 1 may not have the vehicle type identifier 14. [0023]
The informing device 16 is a unit to notification information for the driver. The information unit 16 includes a display 102 that can at least visually inform the driver of the information as shown in Fig. IB.
The display 102 is a display device that is arranged at a place on which the driver in the user's vehicle can visually focus with ease, and displays the image of the nearby surroundings of the user's vehicle generated by the informing processor 15. As the display 102, a display device disposed on a middle portion of a dashboard, a display device as an instrument panel, or a head-up display is exemplified. Note that the informing device 16 may be a display device mounted in a different apparatus other than the driving support apparatus 1. In this case, the driving support apparatus 1 may not have the informing device 16.
15
[0024]
Next, operations will be described.
Fig. 2 is a flowchart illustrating operations of the driving support apparatus in accordance with Embodiment 1, and indicating operations in which the driving support apparatus informs the driver of the approach of another vehicle to the user's vehicle.
In Fig. 2, processes in Steps ST1 to ST4 are consecutively and repeatedly executed until a driver of the user's vehicle finish driving. [0025]
First, the information collector 10 monitors the situation around the user's vehicle, and obtains the information indicating the situation around the user's vehicle (Step ST1) . The information indicating the situation around the user's vehicle includes information about an object existing around the user's vehicle and information about the road on which the user's vehicle is traveling. When the object such as another vehicle around the user's vehicle is a moving object, the information collector 10 obtains a relative location of the moving object with respect to the user's vehicle in addition to appearance information such as the size and shape of the moving object. The information about the road on which the user's vehicle is traveling includes image data of separate white lines defining the lane. [0026]
Note that a plurality of information collectors 10 can be disposed, for example, at both sides of, in front of and at the back of the user's vehicle. Each information collector 10 independently and continuously obtains the information
16
indicating the situation around the user's vehicle.
Information collection scope of the information collector 10 is a disc-like range spreading in a horizontal direction around each information collector. The radius of the disc-like range is, for example, about 10 meters. Note that the radius of the information collection scope depends on the types or capabilities of a sensor and a camera included in the information collector 10. [0027]
In Step ST2, the vehicle detector 11 receives the information indicating the situation around the user's vehicle obtained in real-time by the information collector 10, and, among the received information, identifies information on the moving object and information on the stationary object. The moving object is an object such as another vehicle, a human being or an animal around the user's vehicle, which has a probability of colliding with the user's vehicle when moving. The stationary object is an object such as a road side tree, a guard fence or a road sign, which does not move.
Since the stationary object does not actively approach the user's vehicle, the stationary object is excluded from targets that should trigger notifications of approaches to the user's vehicle. On the other hand, the moving object has a probability of coming in contact with or colliding with the user's vehicle when both the user's vehicle and the moving object move, and thus is a target that should trigger a notification of approach to the user's vehicle. [0028]
After identifying the information on the moving object from the information indicating the situation around the user ' s
17
vehicle, the vehicle detector 11 outputs the information on the moving object to the distance determiner 12 and the vehicle type identifier 14. The information on the moving object includes the relative location of the moving object with respect to the 5 user's vehicle in addition to the appearance information of the moving object. [0029]
The distance determiner 12 extracts information corresponding to the relative location of the moving object with JO respect to the user's vehicle from among the information on the moving object. Then, the distance determiner 12 determines the distance between the moving object and the user's vehicle in accordance with the information corresponding to the relative location. 15 Note that the determination of the distance between the moving object and the user's vehicle is continuously and repeatedly executed for the moving object identified by the vehicle detector 11. The distance determiner 12 outputs the distance between the moving object and the user's vehicle to 20 the distance state calculator 13 one after another. [0030]
The vehicle type identifier 14 extracts the appearance
information of the moving object from among the information on
the moving object that is identified by the vehicle detector
25 11. Then, the vehicle type identifier 14 identifies the type
of the moving object based on the appearance information.
For example, the vehicle type identifier 14 obtains a captured image of the moving object as the appearance information of the moving object, and performs image analysis 30 on the captured image thereby to match an image of the moving
18
object against the pre-provided shape pattern of the moving object, thus identifying the type of the moving object. As the types of a moving object, a normal-sized vehicle such as a sedan or a Sport Utility Vehicle (SUV), a large-sized vehicle such 5 as a truck or a bus, a small-sized vehicle such as a motorcycle or a bicycle, a human being, and an animal are exemplified. [0031]
In Step ST3, the distance state calculator 13 performs a calculation on the basis of the measurement information 10 obtained by the distance determiner 12 to thereby acquire the positional relationship between the moving object and the one or more ranges of distance from the user's vehicle in each of which a notification of the approach of the moving object to the user's vehicle is to be provided. 15 Note that the range of distance from the user's vehicle, in which a notification of the approach of the moving object is to be provided, is a range that is defined in accordance with predetermined distance thresholds from the user's vehicle. When the moving object is another target vehicle other than the 20 user's vehicle, the distance state calculator 13 determines, as a traveling state of the target vehicle, a positional relationship as to whether the target vehicle moves into the range of distance. This determination process will be explained later in detail with reference to Figs. 3 to 8. 25 [0032]
In Step ST4, the informing processor 15 generates the
image of the nearby surroundings of the user's vehicle including
the images of the user's vehicle and the threshold lines
defining at least one of the ranges of distance, and causes the
30 display of the informing device 16 to display the image of the
19
nearby surroundings of the user's vehicle.
When the informing processor 15 receives the traveling state of the target vehicle acquired by the distance state calculator 13 and the type of the target vehicle identified by the vehicle type identifier 14, the informing processor 15 superimposes the image of the target vehicle on the image of the nearby surroundings of the user's vehicle for display. In this way, the notification information based on the determination results generated by the distance state calculator 13 is provided by the informing device 16. [0033]
The display, which is disposed at the place on which the driver can visually focus with ease, displays the image of the nearby surroundings of the user's vehicle in a plan-view or in a bird's-eye-view. In the bird's-eye-view, the display displays a user's vehicle, a moving object, and a road on which the user's vehicle is traveling, which are much more three-dimensional as compared to the plan-view image.
As the display disposed at the place on which the driver can visually focus with ease, the display device disposed on the middle portion of the dashboard, the display device as the instrument panel, and the head-up display are exemplified.
Further, when the display displays, as the image of the target vehicle, an icon corresponding to the type of the target vehicle identified by the vehicle type identifier 14 , the driver can intuitively recognize the target vehicle. [0034]
The processes in Steps ST1 to ST4 are consecutively and repeatedly executed.
That is, when the driver drives the user's vehicle (NO
20
in Step ST5), the process returns to Step ST1 and a series of processes in Step ST1 to Step ST4 is repeatedly executed.
On the other hand, when the driver stops driving the user ' s vehicle (YES in Step ST5), a series of processes in Step ST1 to Step ST4 is terminated. Note that the stop of driving corresponds to, for example, a state in which a drive portion such as an engine of the user's vehicle is stopped and power supply to electrical components of the user's vehicle is stopped. [0035]
Next, a determination process in Step ST3 shown in Fig. 2 will be described in detail.
An example of a case in which target vehicles as moving objects are present around the user's vehicle will be explained. Fig. 3 is a schematic diagram illustrating the ranges of distance in each of which a notification of the approach of another vehicle is to be provided. In Fig. 3, it is assumed that the user's vehicle is traveling on local streets in an ordinary state. Note that the traveling in an ordinary state means, for example, a traveling state in which the user's vehicle is traveling on the road at a legal speed.
Fig. 4 is a schematic diagram illustrating the ranges of distance when relative velocities between the user's vehicle and the target vehicles are smaller as compared to the case shown in Fig. 3. Fig. 5 is a schematic diagram illustrating the ranges of distance when the relative velocities between the user's vehicle and the target vehicles are larger as compared to the case shown in Fig. 3. [0036]
In this invention, with regard to the positional
21
relationship between a moving object and the range of distance from the user's vehicle, that is, the traveling state of the moving object, determined by the distance state calculator 13, there are three types of the traveling state of the target vehicle: a safe state, an attention-requiring state, and an alert state.
The safe state is a state in which the user's vehicle and a target vehicle are separated from each other such that the target vehicle is outside the attention zone set in advance. When this positional relationship is maintained, the user's vehicle is able to safely travel without coming in contact with the target vehicle.
The attention-requiring state is a state in which the distance between the user's vehicle and a target vehicle is in a predetermined attention zone but is outside an alert zone. The attention zone corresponds to a first range of distance in this invention, and has the probability that the target vehicle comes in contact with the user's vehicle in a case where the user's vehicle approaches the target vehicle through an operation of the user's vehicle, such as a lane-change or a left/right turn.
The alert state is a state in which the distance between the user's vehicle and a target vehicle is in a predetermined alert zone. The alert zone is the nearest zone away from the user's vehicle, and corresponds to a second range of distance in this invention . When the target vehicle moves into the alert zone, the driver is required to immediately maneuver the user' s vehicle to avoid contact with the target vehicle. [0037]
In Fig. 3, a lane width A is a width of the lane on which
22
the user's vehicle is traveling. An inter-vehicle distance B is a distance between the user's vehicle and another vehicle behind the user's vehicle. A user's vehicle width C is a width of the user's vehicle. An inter-vehicle distance D is a 5 distance between the user's vehicle and another vehicle at the left side of the user's vehicle.
A distance threshold H is a predetermined threshold value of the distance set at the left side of the user's vehicle. The distance threshold H defines the alert zone arranged at the left
10 side of the user's vehicle. The range of distance LI and the range of distance L2 correspond predetermined threshold values of the distances set at the back of the user's vehicle. LI defines the alert zone arranged at the back of the user ' s vehicle L1+L2 defines the attention zone arranged at the back of the
15 user's vehicle.
Note that both the attention zone and the alert zone are set at the left side of the user's vehicle and at the back of the user's vehicle in Fig. 3. Alternatively, these zones may be set at the right side of the user's vehicle and in front of
20 the user's vehicle. [0038]
Conditions for the positional relationship indicating the safe state are represented by the following expression (1) .
D > (A-C)/2 or B > (L1+L2) ••■ (1)
25 When the positional relationship between the user's vehicle and the target vehicle satisfies the expression (1), the target vehicle is outside the attention zone. Thus, the traveling state of the target vehicle is judged as being the safe state. 30 Conditions for the positional relationship indicating
23
the attention-requiring state are represented by the following expression (2).
(A-C)/2 > D > H or (L1+L2) > B > LI •■• (2) When the positional relationship between the user's 5 vehicle and the target vehicle satisfies the expression (2), the target vehicle is in the attention zone. Thus, the traveling state of the target vehicle is judged as the attention-requiring state.
Conditions for the positional relationship indicating 10 the alert state are represented by the following expression (3} .
H > D or LI > B •■• (3)
When the positional relationship between the user's vehicle and the target vehicle satisfies the expression (3), the target vehicle is in the alert zone. Thus, the traveling 15 state of the target vehicle is judged as the alert state.
The distance thresholds H, LI, L2 are predetermined by the distance state calculator 13 using the lane width A, and may be set to any values by the driver. [0039] 20 Alternatively, the distance state calculator 13 may dynamically change the distance thresholds H, LI, L2 in accordance with a traveling scene or a traveling situation of the user's vehicle. In this case, the above-mentioned conditions for the positional relationship are automatically 25 changed in accordance with the traveling scene or the traveling situation of the user's vehicle.
For example, the distance threshold H set at the left side
of the user ' s vehicle can be classified into the following three
threshold values: a value Hn for a close distance from the user's
30 vehicle, a value Hm for an intermediate distance from the user's
2vehicle, and a value Hf for a far distance from the user' s vehicle . The distance threshold H can be changed into one of the threshold values Hn, Hm, and Hf in accordance with the traveling scene or the traveling situation of the user's vehicle. 5 Similarly, the distance thresholds LI, L2 set at the back of the user's vehicle can be classified into the following three pairs of values: a pair of values Lin, L2n for close distances from the user's vehicle, a pair of values Llm, L2m for intermediate distances from the user's vehicle, and a pair of
10 values Llf, L2f for far distances from the user's vehicle. The distance thresholds LI, L2 can be changed into one of the pair of Llf and L2f, the pair of Llm and L2m, and the pair of Llf and L2f in accordance with the traveling scene or the traveling situation of the user's vehicle.
15 [0040]
Let us assume that the user's vehicle is stuck in a traffic jam or the user's vehicle is waiting for traffic signal change. In such a traveling scene, the relative velocity between the user's vehicle and the target vehicle is smaller as compared
20 to the case in which the user's vehicle is traveling at the legal speed set for the local road.
Also, in a traveling situation where the relative velocity between the user's vehicle and the target vehicle, which is measured by the vehicle detector 11, is less than a
25 predetermined threshold value of relative velocity, the
relative velocity between the user's vehicle and the target vehicle is smaller as compared to the case in which the user's vehicle is traveling at the legal speed set for the local road. Such a traveling scene or a traveling situation is
30 illustrated in Fig. 4. In this case, a probability of contact
5
between the user's vehicle and the target vehicle is lower as compared to the case in which the user's vehicle is traveling at the legal speed set for the local road. Thus, the distance state calculator 13 changes distance thresholds H, LI, L2 into the threshold values Hn, Lin, L2n, respectively.
Thereby, the attention zone and the alert zone become narrower as shown in Fig. 4. [0041]
On the other hand, let us assume that the user's vehicle is traveling on a highway. In such a traveling scene, the relative velocity between the user's vehicle and the target vehicle is larger as compared to the case in which the user's vehicle is traveling at the legal speed set for the local road. Let us assume that the relative velocity between the user's vehicle and the target vehicle, which is measured by the vehicle detector 11, is more than a predetermined threshold value of relative velocity. In such a traveling situation, the relative velocity between the user's vehicle and the target vehicle is larger as compared to the case in which the user's vehicle is traveling at the legal speed set for the local road.
Such a traveling scene or a traveling situation is illustrated in Fig. 5. In this case, a probability of contact between the user's vehicle and a following target vehicle is higher as compared to the case in which the user's vehicle is traveling at the legal speed set for the local road. Thus, the distance state calculator 13 changes distance thresholds Ll, L2 of the distances into the threshold values Llf, L2f, respectively. Thereby, the attention zone and the alert zone at the back of the user's vehicle become broader as shown in Fig. 5.
26
[0042]
In the case shown in Fig. 5, the relative velocity between the user's vehicle and another vehicle that is present at the left side of the user's vehicle is not substantially changed. Thus, the distance state calculator 13 does not change the distance threshold H shown in Fig. 3.
As mentioned above, the distance threshold H set at the left side of the user's vehicle and the distance thresholds Ll, L2 set at the back of the user's vehicle may not be necessarily changed at the same time, and are appropriately changed in accordance with the traveling scene or the traveling situation. [0043]
Next, a displaying process for displaying the image of the nearby surroundings of the user's vehicle in Step ST4 shown in Fig. 2 will be described in detail.
The informing processor 15 displays, on the display, the icon corresponding to the type of a moving object identified by the vehicle type identifier 14, in accordance with the positional relationship between the moving object and the range of the distance from the user's vehicle, which is successively calculated and acquired by the distance state calculator 13.
Fig. 6 illustrates the image of the nearby surroundings of a user's vehicle including the images of the user's vehicle and threshold lines arranged at the sides of the user's vehicle. In Fig. 6, the nearby surroundings of the user's vehicle are viewed in a plan view. In the image 16A of the nearby surroundings of the user's vehicle as shown in Fig. 6, an icon 17 indicating the user's vehicle is located at the center of the image. At both sides of the icon 17 indicating the user's vehicle, a plurality of the threshold lines 18a, 19a, 18b, 19b
27
is displayed.
Moreover, in the image 16A of the nearby surroundings of the user's vehicle, icons 20, 21 indicating target vehicles as moving objects around the user's vehicle are displayed. [0044]
The threshold lines 19a, 19b disposed in the lateral direction, which are arranged at close distances from the user' s vehicle and at both sides of the user's vehicle, correspond to the distance thresholds that define an alert zones arranged at the lateral sides of the user's vehicle. The threshold lines 19a, 19b can be obtained by drawing lines along the road, each line passing through a point located the distance H away from one lateral side of the user's vehicle as shown in Figs. 3 to 5.
The alert zones arranged at the lateral sides of the user ' s vehicle correspond to the range of distance that is a region sandwiched in between the threshold lines 19a and 19b.
For example, the alert zone is a surrounded zone defined by margin distances set in front of and at the back of the user ' s vehicle and the distance thresholds H set at the lateral sides of the user's vehicle. Even when another target vehicle beyond the margin distances is traveling ahead of or behind the user's vehicle and passes across the threshold line 19a or 19b, the target vehicle is not judged to have moved into the alert zone. The margin distances may be set based on the distance thresholds Ll and L2 set in front of and at the back of the user's vehicle. Alternatively, without using the margin distances, the alert zone may be a surrounded zone defined by a distance substantially equal to the longitudinal length of the user's vehicle and the distance thresholds H set at the lateral sides
28
of the user's vehicle. [0045]
The threshold lines 18a, 18b disposed in the lateral direction, which are arranged at both sides of the user's vehicle, correspond to the distance thresholds that define attention zones arranged at both sides of the user's vehicle. The threshold lines 18a, 18b can be obtained by drawing lines along the road, each line passing through a point located the threshold distance of (A-C)/2 away from one side of the user's vehicle.
The attention zones arranged at the lateral sides of the user's vehicle correspond to the ranges of distance that are regions sandwiched in between the threshold lines 18a and 18b and positioned beyond the threshold lines 19a and 19b.
For example, the attention zones are surrounded zones defined by margin distances set in front of and at the back of the user's vehicle and the distances (A-C)/2-H set at the lateral sides of the user's vehicle. Even when the target vehicle traveling ahead of or behind the user's vehicle and beyond the margin distances from the user's vehicle passes across the threshold line 18a or 18b, the target vehicle is not judged to have moved into the attention zone. The margin distances may be set based on the distance thresholds LI and L2 set in front of and at the back of the user's vehicle. Alternatively, without using the margin distances, the attention zone may be a surrounded zone defined by a distance substantially equal to the longitudinal length of the user's vehicle and the distances (A-C)/2-H set at the lateral sides of the user's vehicle.
29
Fig. 7 illustrates an image of the nearby surroundings of a user's vehicle including the images of the user's vehicle and threshold lines arranged in front of and at the back of the user's vehicle. In Fig. 7, the nearby surroundings of the user's vehicle are viewed in a plan view like Fig. 6. In the image 16B of the nearby surroundings of the user's vehicle as shown in Fig. 7, the icon 17 indicating the user's vehicle is located at the center of the image. In front of the icon 17 of the user's vehicle, threshold lines 22b, 23b are displayed. At the back of the icon 17 of the user's vehicle, threshold lines 22a, 23a are displayed.
Moreover, in the image 16B of the nearby surroundings of the user's vehicle, the icons 20, 21 indicating the target vehicles as moving objects around the user's vehicle are displayed. [0047]
The threshold lines 23a, 23b, which are arranged at close distances from the user's vehicle and in front of and at the back of the user's vehicle, correspond to the distance thresholds that define an alert zone. Each threshold line is, for example, a displayed line that passes through a point the threshold distance of LI away from the front or rear of the user ' s vehicle shown in Figs. 3 to 5 . The alert zone arranged at the front and rear of the user's vehicle is the range of distance that is a region sandwiched in between the threshold lines 23a and 23b.
The threshold lines 22a, 22b, which are arranged in front of or at the back of the user's vehicle, correspond to the distance thresholds that define attention zones in front of and at the back of the user's vehicle. Each threshold line is, for
30
example, a displayed line that passes through a point the threshold distance of (L1+L2) away from the front or rear of the user's vehicle shown in Figs. 3 to 5.
The attention zones in front of and at the back of the 5 user's vehicle are the ranges of distance that are disposed beyond the threshold lines 23a, 23b and sandwiched in between the threshold lines 22a and 22b. [0048]
Fig. 8 illustrates an image of the nearby surroundings 10 of a user's vehicle including the images of the user's vehicle and threshold lines arranged at the sides of, in front of and at the back of the user's vehicle. In Fig. 8, the nearby surroundings of the user's vehicle are viewed in a plan view. As shown in Fig. 8, the image 16C of the nearby 15 surroundings of the user's vehicle shows a combination of the attention zone and alert zone shown in Fig. 6 and the attention zone and alert zone shown in Fig. 7.
The threshold lines shown in Figs. 6 to 8 may be displayed
at a prescribed small scale corresponding to actual distances
20 of the real world. Alternatively, the threshold lines may be
displayed at a larger scale to enhance visibility of the driver.
[0049]
The moving object around the user's vehicle is continuously displayed by the icon corresponding to the type 25 of the moving object.
Fig. 9 is a diagram illustrating examples of icons indicating target vehicles. Each icon is set corresponding to a vehicle type. In the informing processor 15, the icons corresponding to the vehicle types are registered in advance. 30 The vehicle types represent a normal-sized vehicle, a
31
small-sized vehicle, and a large-sized vehicle as shown in Fig. 9. The informing processor 15 matches the type of a target vehicle identified by the vehicle type identifier 14, against the registered vehicle types associated with the icons, thus 5 identifying its corresponding icon.
Correspondences between the types of moving objects and the icons may be set by a user.
In this case, since the moving object is displayed by the icon set by the user himself/herself, he/she can more easily 10 recognize the type of the moving object that is approaching to the user's vehicle. [0050]
Next, the change of the image of the nearby surroundings of the user's vehicle in accordance with the traveling state 15 of another target vehicle will be explained.
Fig. 10 is a diagram illustrating an exemplary image displayed when the driving of the target vehicle to the user's vehicle is in a safe state. Fig. 11 is a diagram illustrating an exemplary image displayed when the driving of the target 20 vehicleto the user ' s vehicle is in an attention-requiring state Fig. 12 is a diagram illustrating an exemplary image displayed when the driving of the target vehicle to the user's vehicle is in an alert state. Fig. 13 is a diagram illustrating an exemplary image displayed when the target vehicle further 25 approaches the user's vehicle from a position shown in Fig. 12. [0051]
Figs. 10 to 13 illustrate representations of exemplary images of the nearby surroundings of a user's vehicle displayed on a display screen 16D of an instrument panel in the user's 30 vehicle. Specifically, the display screen 16D displays the
32
image 16A of the nearby surroundings of a user's vehicle, in addition to display items in the instrument panel, such as a residual fuel meter 24, a tachometer 25, and a current value 26 of a vehicle's engine speed. [0052]
When a moving object with a safe state is outside the attention zone set for the user's vehicle, the icon 17 indicating the user's vehicle and the threshold lines 18a, 19a, 18b, 19b that define the above-described zones are appearing in the image 16A of the nearby surroundings of the user's vehicle.
In this regard, as shown in Fig. 10, even in the case of the safe state, the icons 20, 21 indicating other target vehicles detected around the user's vehicle may be displayed to appear in the image 16A of the nearby surroundings of the user's vehicle .
Note that the positional relationship between each zone set for the user's vehicle and the target vehicle is updated in real-time by using consecutively calculated positional relationship acquired by the distance state calculator 13.
Moreover, in the case of the safe state, the icon 17 indicating the user's vehicle and the icons 20, 21 indicating the target vehicles are displayed with a default color. This default color may be the same color as a background color of the image 16A of the nearby surroundings of the user's vehicle so as not to draw attention of the driver. [0053]
When the target vehicle moves into an attention zone from the position shown in Fig. 10 while being kept outside an alert zone, if the distance state calculator 13 determines that the
33
an attention-requiring state occurs, the informing processor 15 changes the image 16A of the nearby surroundings of the user ' s vehicle into a display pattern corresponding to the attention-requiring state.
For example, as shown in Fig. 11, when the target vehicle moves into the attention zone arranged at the right side of the user's vehicle, the informing processor 15 changes both the color of the icon 20 indicating the target vehicle that has moved into the attention zone and the color of the right-hand region including this attention zone, into an attention-drawing color . The attention-drawing color corresponds to a first color in this invention. The attention-drawing color is, for example, yellow that is capable of drawing the attention of the driver. However, since attitudes toward colors depend on countries and their cultures, the attention-drawing color is not limited to yellow in this invention. [0054]
The informing processor 15 causes a display item drawing the attention of the driver to appear near the icon 17 indicating the user's vehicle in the image 16A of the nearby surroundings of the user's vehicle.
The display item drawing the attention can be freely selected under the condition that the display item can draw the attention of the driver about the approach of the target vehicle to the user's vehicle. For example, the display items are an attention mark 27 and letters 28 of "Caution". In Fig. 11, the display items are arranged in front of and in the traveling direction of the user's vehicle. Alternatively, the display items can be arranged at another place on which the driver can visually focus with ease.
34
With such a display, the driver can intuitively recognize which direction the driver should pay attention to. [0055]
In addition, in the image 16A of the nearby surroundings
5 of the user's vehicle, the threshold lines 18a, 19a, 18b, 19b
that define attention zones at both sides of the user's vehicle
are displayed so as to extend along the road on which the user's
vehicle is traveling. When the target vehicle moves into the
attention zone arranged at one side of the user's vehicle, the
10 informing processor 15 changes the color of the attention zone
arranged at the corresponding side of the user's vehicle and
appearing in the image 16A of the nearby surroundings of the
user's vehicle, into an attention-drawing color . In this case,
the informing processor 15 may change the color of the whole
15 area partitioned by the threshold lines 18a, 19a, 18b, 19b, into
the attention-drawing color. For example, when another target
vehicle approaches the user's vehicle from the right side of
the user's vehicle, the informing processor 15 can change the
color of a band-shaped area arranged at the right side of the
20 user's vehicle including the attention zone, into the
attention-drawing color.
With such a display, the driver can intuitively recognize the direction in which the target vehicle approaches from an area at either one of the lateral sides of the user's vehicle. 25 Note that the change to the attention-drawing color may be applied to the attention zone or a region including the attention zone, and it is not necessary to change the color of the icon indicating the target vehicle which has moved into the attention zone. 30 [0056]
35
Next, when the target vehicle further moves into the alert zone from the position shown in Fig. 11, if the distance state calculator 13 determines that an alert state occurs, the informing processor 15 changes the contents of the image 16A of the nearby surroundings of the user's vehicle into a display pattern corresponding to the alert state. With regard to the alert state, for example, the informing processor 15 provides notification of an alert classified into two levels, to the driver. [0057]
As shown in Fig. 12, when the target vehicle passes across the threshold line 19b, the informing processor 15 changes both the color of the alert zone and the color the icon 20 indicating the target vehicle which has moved into the alert zone, from the default color into an alert color which is red. The informing processor 15 also changes the color of the attention zone through which the target vehicle moves into the alert zone, into a different alert color, such as pale red, which differs in tone from the alert color.
The alert color corresponds to a second color in this invention. As the alert color, red providing high visibility is exemplified. However, since the attitudes toward colors depend on countries and their cultures, the alert color is not limited to red in this invention. [0058]
As mentioned above, the ranges of distance from the user ' s vehicle, such as the attention zone and the alert zone, are set at different regions depending upon distances from the user's vehicle.
When the target vehicle moves into one or more ranges of
36
the ranges of distance, the informing processor 15 changes the
display colors of the one or more ranges of distance or the
display colors of both the image of the target vehicle and the
one or more ranges of distance, in a manner that the display
5 color of the alert zone which is a close range of distance from
the user's vehicle is different from the display color of the
attention zone which is a far range of distance from the user's
vehicle. Thereby, the driver can intuitively recognize that
the target vehicle approaches the user's vehicle at what range
10 of distance, on the basis of the change of the displayed color.
Moreover, the informing processor changes the display
color of the attention zone through which the target vehicle
has passed, into a color resembling the alert color, in addition
to the changing of the color of the alert zone. Thereby, the
15 driver can easily recognize the direction from which the target
vehicle approaches the user's vehicle, on the basis of a
stepwise color change.
[0059]
Moreover, the informing processor 15 causes an alert 20 display item to appear near the icon 17 indicating the user's vehicle in the image 16A of the nearby surroundings of the user's vehicle.
The alert display item can be freely selected under the condition that the alert display item can alert the driver that 25 the driving maneuver to avoid the approach of the target vehicle to the user's vehicle is required. For example, alert display items are the attention mark 27 and letters 29 of "Hit the Brakes" With such a display, the driver can intuitively recognize that the target vehicle approaches the user's vehicle by a positon 30 where the target vehicle is likely to come in contact with the
37
user's vehicle.
In Fig. 12, the alert display items are arranged at the
front side of the icon 17 in the traveling direction of the user' s
vehicle. Alternatively, the display items may be arranged at
5 another place on which the driver can visually focus with ease.
[0060]
As described above, when another target vehicle moves into the attention zone or the alert zone, the informing processor 15 changes the contents and sizes of the mark and
10 displayed letters which are associated with drawing of the attention of the driver, in a manner that the contents and sizes differ between the attention zone that is a far range of distance from the user's vehicle, and the alert zone that is a close range of distance from the user's vehicle.
15 With such a display, the driver can intuitively recognize how close the target vehicle is to the user's vehicle, on the basis of the contents of the mark and the displayed letters. [0061]
Moreover, when the target vehicle passes across the
20 threshold line 19b, as shown in Fig. 12, the informing processor 15 causes the display screen 16D of the display device to display a translucent blur 30 with an alert color in a region corresponding to the direction from which the target vehicle approaches the user's vehicle.
25 Note that the translucent blur 30 is semi-transparent and is displayed as a gradation pattern in which the alert color gradually becomes deeper toward the direction from which the target vehicle approaches the user's vehicle.
With such a display, the driver can recognize the
30 direction in which the target vehicle approaches the user's
38
vehicle. Thus, the driver can intuitively recognize that it is necessary to hit the brakes because the target vehicle approaches the user's vehicle from the right side.
Alternatively, a plurality of light emitting elements may 5 be disposed at places on which the driver can visually focus with ease. The informing processor 15 can cause the light emitting element disposed at the place close to the target vehicle in the direction in which the target vehicle approaches the user's vehicle, to emit light instead of using the 10 translucent blur 30. In this case, the plurality of light emitting elements can be used as the informing device 16. [0062]
Next, when the target vehicle further moves into the alert
zone from the position shown in Fig. 12, the informing processor
15 15 changes the image 16A of the nearby surroundings of the user's
vehicle into a display pattern corresponding to a higher alert
state.
For example, as shown in Fig. 13, the informing processor 15 enlarges the size of the alert display items as compared to 20 the case shown in Fig. 12, and arranges the enlarged alert display items at the center of the image 16A of the nearby surroundings of the user's vehicle. Thereby, the driver can strongly recognize that it is necessary to hit the brakes. [0063] 25 Moreover, the informing processor 15 changes the
translucent blur 30 into a blur 31 and enlarges the displayed area of the blur as the target vehicle approaches the user's vehicle. In the case where, as shown in Fig. 13, the target vehicle approaches the user's vehicle from the right side, the 30 blur 31 is displayed on the right-hand portion of the display
39
screen 16D.
Note that the blur 31 is also a translucent blur colored with the alert color like the translucent blur 30. The blur 31 is displayed as a gradation pattern in which the alert color gradually becomes deeper toward the direction from which the target vehicle approaches the user's vehicle.
With such a display, when the driver glances at the display screen 16D, from the large-sized blur 31, the driver can easily recognize that the right side of the user's vehicle will be in a dangerous state due to the approach of the target vehicle. [0064]
The display screen 16Dmay have a plurality of image layers LI to L6 as shown in Fig. 14 . In this case, it can be configured that, the closer the target vehicle approaches the user's vehicle, the higher is the level of a specific image layer on which the informing processor 15 causes the translucent blur to be displayed in a manner that the level of the specific image layer is lower than or equal to the level of an image layer L6 displaying a predetermined important display item such as the current value 2 6 of a vehicle's engine speed. [0065]
For example, in the traveling state of the target vehicle as shown in Fig. 12, the translucent blur 30 is displayed on the image layer L3 whose level is lower than the level of the image layer L4 displaying the tachometer 25 on the right-hand portion of the display screen 16D and higher than the level of the image layer L2 displaying the residual fuel meter 24. With such a configuration, the translucent blur 30 is displayed as if the translucent blur is projected from under the tachometer
40
In a state in which the target vehicle further approaches
the user's vehicle shown in Fig. 13, the blur 31 is displayed
on a higher image layer L5. Note that the level of the higher
image layer displaying the blur 31 is not higher than the level
5 of the image layer L6 displaying the current value 26 of the
vehicle ' s engine speed which is the indispensable display item.
Namely, the large-sized blur 31 is displayed on the image
screen 16D under the condition that the blur 31 is not displayed
on an image layer having a higher level than that of the image
10 layer displaying an important display item such as a vehicle's
engine speed or vehicle velocity that is legally required to
be displayed. With such an image layer structure, display items
that are hidden by the blur 31 can be the fewest possible items.
[0066]
15 In the above-mentioned examples, the positional
relationship among the target vehicle, the attention zones at both sides of the user's vehicle, and the alert zone has been explained. In another example in which the image 16C of the nearby surroundings of a user's vehicle shows a combination of 20 an alert zone and attention zones at both sides of, in front of and at the back of the user's vehicle, it is possible to determine the traveling state of the target vehicle like the above-mentioned examples. [0067] 25 Fig. 15 is a diagram illustrating an exemplary display of the ranges of distance set at both sides of, in front of and at the back of the user's vehicle. In Fig. 15, an area A is a region including the attention zone defined by the threshold lines IGa, 19a at the left side of the user's vehicle, and an 30 area B is a region including the attention zone defined by the
■11
threshold lines 18b, 19b at the right side of the user's vehicle. A zone C corresponds to the attention zone defined by the threshold lines 22b, 23b in front of the user's vehicle, and a zone D corresponds to the attention zone defined by the 5 threshold lines 22a, 23a at the back of the user's vehicle. A zone E corresponds to the alert zone defined by the threshold lines 19a, 19b, 23a, 23b. Note that the letters "A", "B", "C", "D", and "E" are not displayed in the actual image 16C of the nearby surroundings of the user's vehicle. 10 [0068]
When the target vehicle indicated by the icon 32 moves into an attention zone from the left side, the informing processor 15 changes the color of the whole area A in the image 16C of the nearby surroundings of the user's vehicle, into 15 yellow that is an attention-drawing color, as shown in Fig. 16. Similarly, when the target vehicle moves into the attention zone D from its rear side, the informing processor 15 changes the color of the attention zone D in the image 16C of the nearby surroundings of the user's vehicle, into yellow, as shown in 20 Fig. 17. [0069]
When the target vehicle passes across the threshold line 19a from the position indicated in shown in Fig. 16, the informing processor 15 changes the color of the alert zone E 25 in the image 16C of the nearby surroundings of the user's vehicle, into red that is an alert color, and changes the color of the area A including the attention zone from which the target vehicle has moved into the alert zone, into pale red, as shown in Fig. 18. 30 Similarly, when the target vehicle passes across the
42
threshold line 23a from a position indicated in shown in Fig. 17, the informing processor 15 changes the color of the alert zone E in the image 16C of the nearby surroundings of the user's vehicle, into red that is the alert color, and changes the color of a rear area including the attention zone D from which the target vehicle has moved into the alert zone, into pale red.
As mentioned above, when the target vehicle moves into an attention zone at the lateral side of, in front of, or at the back of the user's vehicle, the informing processor 15 may change the color of the whole area including the attention zone which is defined by the extending threshold lines. With such a display, the driver can easily recognize that the target vehicle approaching to the user's vehicle is present in front of or at the back of the user's vehicle. [0070]
When another target vehicle moves from an area at one side of the user's vehicle, into attention zones defined at the side of and at the back of the user's vehicle, or into attention zones defined at the side of and in front of the user's vehicle, the informing processor 15 changes the color of an area including the attention zone close to the area at the side of the user's vehicle from which the target vehicle has moved, into an attention color. [0071]
For example, as shown in Fig. 20, when the target vehicle moves, from an area at the left side of the user's vehicle, into attention zones defined at the left side of and at the back of the user's vehicle, the informing processor 15 changes the color of the left side area A including the attention zone at the left side of the user's vehicle into yellow, while the informing
43
processor 15 does not change the color of the attention zone D at the back of the user's vehicle.
When the target vehicle further moves into the attention zone D from the position indicated in Fig. 20, the informing processor 15 changes the color of the left side area A into the default color, and changes the color of the attention zone D into yellow, as shown in Fig. 21. With such a display, the driver can easily recognize from which side the target vehicle approaches the user's vehicle. [0072]
When two or more other target vehicles move into the attention zone and the alert zone, the informing processor 15 changes the display color of the zone into which each target vehicle moves, or the display colors of both an image of each target vehicle and the zone into which the target vehicle moves, into a display color corresponding to the zone into which each target vehicle moves.
For example, as shown in Fig. 22, there is a case where target vehicles indicated by the icons 33 and 34 move into an area A including an attention zone at the left side of the user's vehicle, and an alert zone D at the back of the user's vehicle, respectively. In this case, the informing processor 15 changes the colors of the area A the attention zone D into yellow that is a color corresponding to the attention zone. [0073]
When one target vehicle further moves into the alert zone E from the position indicated in Fig. 22, as shown in a display pattern corresponding to alert zones of Fig. 23, the informing processor 15 changes the color of the alert zone E into red, and changes the color of the area A into pale red.
■14
As explained above, even in the case where the two or more target vehicles exist, the display color of the nearby zone in the direction along which each target vehicle approaches is changed into a color depending on the degree of closeness in 5 distance between the user's vehicle and each target vehicle. Therefore, the driver can recognize the degree of closeness in distance for each of the target vehicles, on the basis of the change of the display color. [0074] 10 Moreover, the informing processor 15 displays, on the display, the image of the nearby surroundings of the user's vehicle including a plan-view image or bird's-eye-view image of both the user's vehicle and the target vehicle. Thereby, it is possible to clearly display a positional relationship 15 between the user's vehicle and the target vehicle.
In generating the image of the nearby surroundings of a
user's vehicle, as the target vehicle approaches the user's
vehicle, the informing processor 15 may move a location of a
viewpoint from which the user's vehicle is viewed in the
20 pi an-view or the bird's-eye-view, closer to the user's vehicle.
[0075]
For example, when the positional relationship between the user's vehicle and the target vehicle is in a safe state, the location of the viewpoint is set at the farthest position from 25 the user's vehicle in a vertical direction. Fig. 24A illustrates the image 16E of the nearby surroundings of a user's vehicle when viewed from the farthest viewpoint. [0076]
When the target vehicle indicated by an icon 35 approaches 30 the user's vehicle from the position indicated in Fig. 24A,
45
giving rise to an attention-requiring state, the location of the viewpoint is moved closer to the user's vehicle from its location in the safe state. Fig. 24B illustrates the image 16F of the nearby surroundings of a user's vehicle when viewed from that viewpoint. In the image 16F of the nearby surroundings of the user's vehicle, the icon 17 indicating the user's vehicle and the icon 35 indicating the target vehicle are enlarged as compared to those in the safe state. [0077]
When the target vehicle indicated by the icon 35 further approaches the user's vehicle, giving rise to an alert state, the location of the viewpoint is moved closer to the user's vehicle from its location in the attention-requiring state.
Fig. 24C illustrates the image 16G of the nearby surroundings of the user's vehicle when viewed from that viewpoint. In the image 16G of the nearby surroundings of the user's vehicle, the icon 17 indicating the user's vehicle and the icon 35 indicating the target vehicle are enlarged for display as compared to those in the attention-requiring state. [0078]
As mentioned above, as the target vehicle approaches the user's vehicle, the informing processor moves the location of the viewpoint from which the user's vehicle is viewed in the plan-view or the bird' s-eye-view, closer to the user's vehicle. Thereby, it is possible to display the positional relationship between the user ' s vehicle and the target vehicle in an expanded manner such that the driver can easily recognize the positional relationship. [0079]
In generating the image of the nearby surroundings of the
■16
user's vehicle, as the target vehicle approaches the user's vehicle, the informing processor 15 may move, in the horizontal direction, the location of the viewpoint from which the user's vehicle is viewed in the plan-view or the bird's-eye-view such 5 that the image of the nearby surroundings of the user's vehicle shows the target vehicle.
For example, when the positional relationship between the user's vehicle and the target vehicle is in the safe state or when the target vehicle is not present around the user ' s vehicle,
10 the location of the viewpoint is set just above the user's vehicle. Fig. 25B illustrates the image 161 of the nearby surroundings of a user's vehicle when viewed from that viewpoint. [0080]
15 When the target vehicle moves into the attention zone arranged in front of the user's vehicle from the position indicated in Fig. 25B, the informing processor 15 moves the location of the viewpoint forward in the horizontal direction. Fig. 25A illustrates the image 16H of the nearby surroundings
20 of the user's vehicle when viewed from such a viewpoint. By moving the viewpoint from which the user's vehicle is viewed, forward in the horizontal direction, the icon 17 indicating the user's vehicle is arranged at the center of the image 16H of the nearby surroundings of the user's vehicle, and the icon 35
25 indicating the target vehicle is arranged ahead of the icon 17. [0081]
When the target vehicle moves into the attention zone arranged at the back of the user's vehicle from the position indicated in Fig. 25B, the informing processor 15 moves the
30 viewpoint backward in the horizontal direction. Fig. 25C
17
illustrates the image 16J of the nearby surroundings of the
user's vehicle when viewed from such a viewpoint. By moving
the viewpoint from which the user's vehicle is viewed, backward
in the horizontal direction, the icon 17 indicating the user's
5 vehicle is arranged at the center of the image 16J of the nearby
surroundings of the user's vehicle, and the icon 35 indicating
the target vehicle is arranged at the rear side of the icon 17 .
In Fig. 25A and Fig. 25C, a case in which the viewpoint
is moved forward or backward is illustrated. Alternatively,
10 the viewpoint may be moved in a left or right direction when
the target vehicle approaches the user's vehicle from the left
or right side.
[0082]
As mentioned above, as the target vehicle approaches the 15 user's vehicle, the informing processor moves, in the horizontal direction, the location of the viewpoint from which the user's vehicle is viewed in the plan-view or the bird's-eye-view such that the image of the nearby surroundings of the user's vehicle shows the target vehicle. Thereby, it 20 is possible to appropriately display the target vehicle approaching to the user's vehicle in the image of the nearby surroundings of the user's vehicle. [0083]
In generating the image of the nearby surroundings of the 25 user's vehicle, the informing processor 15 may move the viewpoint in the vertical direction as well as in the horizontal direction. Thereby, it is possible to appropriately display the target vehicle approaching to the user's vehicle in the image of the nearby surroundings of the user's vehicle in an 30 expanded manner.
■18
[0084]
In Embodiment 1, a case has been described where the distance state calculator 13 performs a calculation on the basis of the measurement information obtained by the distance determiner 12 to thereby acquire the positional relationship between the moving object and the one or more ranges of distance from the user's vehicle in each of which a notification of approach of the moving object is to be provided. Nonetheless, the invention is not limited to the case in which the distance state calculator 13 itself calculates the above-mentioned positional relationship.
Namely, the distance state calculator 13 can function as an acquiring unit in the invention, and may be configured to acquire the positional relationship by receiving information on the positional relationship between the moving object and the one or more ranges of distance from the user's vehicle in each of which a notification of approach of the moving object, from an external device that has been calculated the positional relationship.
In this case, the driving support apparatus 1 may not have the vehicle detector 11 and the distance determiner 12. In addition, when the informing processor 15 generates a predetermined image as the icon of a moving object regardless of the type of the moving object, the driving support apparatus 1 may not have the vehicle type identifier 14. [0085]
As mentioned above, the driving support apparatus according to Embodiment 1 includes at least the distance state calculator 13 and the informing processor 15.
The distance state calculator 13 determines the
■49
positional relationship between another target vehicle and the attention and alert zones that are the ranges of distance from the user's vehicle in each of which a notification of approach of the moving object is to be provided. The informing processor 5 15 causes the display to display the image of the nearby surroundings of the user's vehicle including the images of the user's vehicle and threshold lines defining the range of distance, and changes the image of the nearby surroundings of the user's vehicle in accordance with the positional
10 relationship acquired by the distance state calculator 13. As mentioned above, in Embodiment 1, the image of the nearby surroundings of the user's vehicle including the images of the user's vehicle and threshold lines defining the range of distance is displayed on the display device, and the image of
15 the nearby surroundings of the user's vehicle is changed in accordance with the positional relationship between the moving object and the attention and alert zones.
Thereby, the driver can intuitively recognize the situation of the user's vehicle and the situation around the
20 user's vehicle. [0086]
According to Embodiment 1, the informing processor 15 causes the image of the moving object to appear in the image of the nearby surroundings of the user's vehicle, in accordance
25 with the above-mentioned positional relationship acquired by the distance state calculator 13.
Thereby, the driver can intuitively recognize the positional relationship between the user's vehicle and the moving object around the user's vehicle.
30 [0087]
50
According to Embodiment 1, the driving support apparatus includes the vehicle type identifier 14 that determines the type of the target vehicle detected by the vehicle detector 11, and the informing processor 15 causes the image of the target 5 vehicle corresponding to the type which is determined by the vehicle type identifier 14, to appear in the image of the nearby surroundings of the user's vehicle.
In this manner, since the informing processor automatically displays the image corresponding to the type 10 which is determined by the vehicle type identifier 14, as the image of the target vehicle, the driver can intuitively recognize the target vehicle. [0088]
According to Embodiment 1, the ranges of distance from 15 the user's vehicle, such as the attention zone and the alert zone, are set at different regions depending upon distances from the user's vehicle.
When the target vehicle moves into one or more ranges of
distance, the informing processor 15 changes the display colors
20 of the one or more ranges of distance or the display colors of
both the image of the target vehicle and the one or more ranges
of distance, into a display color in a manner that the display
color of the alert zone which is a close range of distance from
the user's vehicle is different from the display color of the
25 attention zone which is a far range of distance from the user's
vehicle-. Thereby, the driver can intuitively recognize that
the target vehicle approaches the user's vehicle at what range
of distance, on the basis of the change of the displayed color.
[0089]
30 According to Embodiment 1, when two or more target
51
vehicles move into their respective different ranges of distance at the same time, the informing processor 15 changes the display colors of the ranges of distance into which their respective target vehicles move, or the display colors of both 5 an image of their respective target vehicles and the corresponding ranges of distance into which their respective target vehicles move, into display colors corresponding to the ranges of distance into which their respective target vehicles move. In the case where such two or more target vehicles exit, 10 the display color of the range of distance in the direction along which each target vehicle approaches the user's vehicle is changed into the color corresponding to the degree of closeness in distance between those vehicles. Therefore, the driver can recognize the degree of closeness in distance for each of the 15 two or more target vehicles, on the basis of the change of the display color. [0090]
According to Embodiment 1, when the target vehicle moves
into the attention zone, the informing processor 15 changes the
20 color of the attention zone or the colors of both the image of
the target vehicle and the attention zone, into an
attention-drawing color. When the target vehicle moves from
the attention zone into the alert zone closer to the user's
vehicle, the informing processor 15 changes the color of the
25 alert zone or the colors of both the image of the target vehicle
and the alert zone, into an alert color, and also changes the
color of the attention zone through which the target vehicle
moves into the alert zone, into a different alert color that
differs in tone from the alert color.
30 In this case, the color of the alert zone is changed into
CO
the alert color, and the color of the attention zone through which the target vehicle moves into the alert zone is also changed into the different alert color resembling the alert color. Thereby, the driver can easily recognize the direction from which the target vehicle approaches the user's vehicle, on the basis of the stepwise color change. [0091]
According to Embodiment 1, when the target vehicle moves into the range of distance, the informing processor 15 causes the display screen 16D of the display for displaying the image of the nearby surroundings of the user's vehicle, to display the translucent blur 30 with a predetermined color on an area corresponding to the direction from which the target vehicle approaches the user's vehicle. The informing processor 15 also changes the translucent blur 30 into another blur 31, and enlarges the size of the area displaying the blur as the target vehicle approaches the user's vehicle. Thereby, the driver can easily recognize the direction from which the target vehicle approaches the user's vehicle, on the basis of the translucent blur 30 and the blur 31, when the driver glances at the display screen 16D. [0092]
According to Embodiment 1, the display screen 16D of the display has the plurality of image layers. In this case, the closer the target vehicle approaches the user's vehicle, the higher is the level of the image layer on which the informing processor 15 causes the blur to appear in a manner that the level of the image layer is lower than or equal to the level of another image layer which displays a predetermined important display item such as the current value 26 of a vehicle's engine speed.
53
Thereby, display items that are hidden by the blur can be the
fewest possible items.
[0093]
According to Embodiment 1, the image 16A of the nearby surroundings of the user's vehicle shows the threshold lines 18a, 19a, 18b, 19b, which define the attention zones at both sides of the user's vehicle as the ranges of distance as well as define the alert zone as another range of distance. The threshold lines 18a, 19a, 18b, 19b are displayed so as to extend along the road on which the user's vehicle is traveling.
When the target vehicle moves into the range of distance arranged in front of, at the back of or at one side of the user's vehicle, the informing processor 15 changes the color of the area including at least one of the ranges of distance that is defined by the extending threshold lines.
Thereby, the driver can easily recognize that the target vehicle approaching to the user's vehicle is in front of or at the rear side of the user's vehicle. [0094]
According to Embodiment 1, when the target vehicle moves into the range of distance, the informing processor 15 changes the contents and sizes of the mark and displayed letters which are associated with drawing of attention of the driver in a manner that the contents and sizes differ between the attention zone that is a far range of distance from the user's vehicle, and the alert zone that is a close range of distance from the user's vehicle. Thereby, the driver can intuitively recognize how close the target vehicle is to the user's vehicle, on the basis of the contents of the mark and displayed letters. [0095]
54
According to Embodiment 1, the informing processor 15 displays, on the display, the image of the nearby surroundings of the user's vehicle including one of the plan-view image of both the user's vehicle and the target vehicle and the 5 bird' s-eye-view image of both the user ' s vehicle and the target vehicle. Thereby, it is possible to clearly display the positional relationship between the user's vehicle and the target vehicle. [0096] 10 According to Embodiment 1, as the target vehicle approaches the user ' s vehicle, the informing processor 15 moves the location of the viewpoint from which the user's vehicle is viewed in the plan-view or the bird's-eye-view, closer to the user's vehicle.
15 Thereby, it is possible to display the positional relationship between the user's vehicle and the target vehicle in the expanded manner such that the driver can easily recognize the positional relationship. [0097]
20 According to Embodiment 1, the informing processor 15 moves the location of the viewpoint from which the user's vehicle is viewed in the plan-view or the bird's-eye-view in the horizontal direction such that the image of the nearby surroundings of the user's vehicle shows the target vehicle as
25 the target vehicle approaches the user's vehicle. Thereby, it is possible to appropriately display the target vehicle approaching to the user's vehicle in the image of the nearby surroundings of the user's vehicle. [0098]
30 Embodiment 2
55
Fig. 26 is a block diagram illustrating a configuration of a driving support apparatus 1A in accordance with Embodiment 2. In Fig. 26, the driving support apparatus 1A is an apparatus that informs a driver of the approach of a moving object such as another vehicle to a user's vehicle. The driving support apparatus 1A includes the information collector 10, the vehicle detector 11, the distance determiner 12, a distance state calculator 13a, the vehicle type identifier 14, the informing processor 15, the informing device 16 and a user's vehicle state acquiring unit 36. [0099]
Among components of the driving support apparatus 1A, the vehicle detector 11, the distance determiner 12, the distance state calculator 13a, the vehicle type identifier 14, the informing processor 15, and the user's vehicle state acquiring unit 36 are constituted by a processor 100 such as a system LSI (Large-Scale Integrated Circuit) or a Central Processing Unit (CPU) that executes a program stored in the memory 101 shown in Fig. IB. Processing circuits may work together to function as above-mentioned components. [0100]
The user's vehicle state acquiring unit 36 is a component to collect vehicle information on a driving maneuver of the user's vehicle. As the vehicle information on the driving maneuver of the user's vehicle, vehicle information corresponding to an operation to activate a blinker for the lane-change, left-turn, or right-turn, and vehicle information corresponding to an operation to accelerate or decelerate of the user's vehicle are exemplified.
The user's vehicle state acquiring unit 36 is connected
56
to an external device outside the driving support apparatus 1A
via wireless connection or wired connection, and receives
existing road data about the road on which the user's vehicle
is traveling, which is stored in the external device. As the
5 external device, a vehicle navigation device mounded on the
user's vehicle, or a cloud server with which the user's vehicle
state acquiring unit is capable of communicating via the
wireless connection, are exemplified.
[0101]
10 The distance state calculator 13a is a component that
corresponds to an acquiring unit in the invention. The distance
state calculator 13a determines the positional relationship
between the range of distance from the user's vehicle (e.g.,
the attention zone and the alert zone) and another target
15 vehicle by using information of the lane width included in the
existing road data received by the user's vehicle state
acquiring unit 36.
The distance state calculator 13a determines whether to
draw the attention of the driver about the driving maneuver of
20 the user's vehicle indicated by the vehicle information, based
on both the vehicle information on the driving maneuver of the
user's vehicle received by the user's vehicle state acquiring
unit 36 and the positional relationship between the range of
distance from the user's vehicle and the target vehicle.
25 The informing processor 15 causes both an image
indicating the driving maneuver that is determined as an act
to draw the attention by the distance state calculator 13a, and
display items associated with drawing of the attention, to
appear in the image of the nearby surroundings of the user's
30 vehicle.
57
[0102]
Next, operations will be described.
Fig. 27 is a flowchart illustrating operations of the
driving support apparatus in accordance with Embodiment 2, and
5 indicating operations in which the driving support apparatus
informs the driver of the approach of the target vehicle to the
user's vehicle.
In Fig. 27, processes in Step STla to Step ST A a are consecutively and repeatedly executed until a driver of the 10 user's vehicle finish driving. [0103]
In Step STla, the information collector 10 monitors the situation around the user's vehicle, and obtains the information indicating the situation around the user's vehicle . 15 When the information collector 10 cannot obtain the information about the road on which the user's vehicle is traveling, the user's vehicle state acquiring unit 36 acquires, from an external device, the information about the lane width of the traffic rout on which the user's vehicle is traveling. 20 At the same time, the user's vehicle state acquiring unit 36 also acquires the vehicle information on the driving maneuver of the user's vehicle. [0104]
In Step ST2, the vehicle detector 11 receives the
25 information indicating the situation around the user's vehicle
obtained in real-time by the information collector 10 as similar
with Step ST2 shown in Fig. 2, and, among the received
information, identifies information on the moving object and
information on the stationary object.
30 After identifying the information on the moving object
58
among the information indicating the situation around the user's vehicle, the vehicle detector 11 outputs the information on the moving object to the distance determiner 12 and the vehicle type identifier 14. [0105]
The distance determiner 12 extracts information corresponding to the relative location of the moving object with respect to the user's vehicle from among the information on the moving object. Then, the distance determiner 12 determines the distance between the moving object and the user's vehicle in accordance with the information corresponding to the relative location.
Note that the determination of the distance between the moving object and the user's vehicle is continuously and repeatedly executed for the moving object identified by the vehicle detector 11. The distance determiner 12 outputs the distance between the moving object and the user's vehicle to the distance state calculator 13a one after another.
The vehicle type identifier 14 extracts the appearance information of the moving object from among the information on the moving object identified by the vehicle detector 11. Then, the vehicle type identifier 14 identifies the type of the moving object based on the appearance information. [0106]
In Step ST3a, the distance state calculator 13a performs a calculation on the basis of both the measurement information obtained by the distance determiner 12 and the information about the lane width of the traffic route on which the user's vehicle is traveling, received by the user's vehicle state acquiring unit 36, to thereby acquire the positional relationship between
59
the moving object and the one or more ranges of distance from the user's vehicle in each of which a notification of approach of the moving object is to be provided.
Further, the distance state calculator 13a determines whether to draw the attention of the driver about the driving maneuver of the user's vehicle indicated by the vehicle information in accordance with both the vehicle information on the driving maneuver of the user's vehicle received by the user's vehicle state acquiring unit 36 and the positional relationship between the range of distance from the user's vehicle and the target vehicle. For example, when the driving maneuver of the user's vehicle corresponds to the operation to activate the blinker, the distance state calculator 13a determines whether to prompt the attention of the driver about the driving maneuver in accordance with both the positional relationship between the user's vehicle and the,target vehicle approaching to the user's vehicle and a direction indicated by the blinker. [0107]
In Step ST4a, the informing processor 15 causes both the image indicating the driving maneuver and the display item associated with drawing of the attention of the driver, to appear in the image of the nearby surroundings of the user's vehicle when the distance state calculator 13a determines that the driving maneuver is a maneuver requiring attention. For example, in a case where the operation to activate a right blinker for the right-turn is executed when the target vehicle approaches the user's vehicle from the right such that the driving state of the target vehicle becomes in the attention-requiring state, the informing processor 15 causes
60
an arrow 37 indicating the right-turn to appear in the image 16A of the nearby surroundings of the user's vehicle as shown in Fig. 28, and causes the display item prompting the attention of the driver to appear in the image 16A of the nearby surroundings of the user's vehicle. In this case, the items drawing the attention of the driver are the attention mark 27 and the letters 28 of "Caution".
Thereby, the driver can intuitively recognize that the driving maneuver corresponding to the right-turn of the user's vehicle is a driving maneuver requiring attention about the target vehicle indicated by the icon 20 . In the above-mentioned example, the driving maneuver corresponding to the right turn is explained. This embodiment is also applicable to the driving
maneuver corresponding to the-left turn. [0108]
The processes in Step STla to Step ST4a are continuously
and repeatedly executed.
When the driver drives the user' s vehicle (NO in Step ST5a) ,
the process returns to Step STla and a series of processes in
Step STla to Step ST 4 a is repeatedly executed.
On the other hand, when the driver stops driving the user ' s
vehicle (YES in Step ST5a) , a series of processes in Step STla
to Step ST4a is terminated.
Note that the stop of driving corresponds to, for example,
a state in which a drive portion such as an engine of the user's
vehicle is stopped and power supply to electrical components
of the user's vehicle is stopped. [0109]
Note that some processes explained in Embodiment 1 and
some processes explained in Embodiment 2 can be concurrently
61
executed. For example, even if the information collector 10 can obtain the information indicating the nearby surroundings of the user's vehicle, including the white lines at both sides of the lane, the user's vehicle state acquiring unit 36 also 5 obtains the existing road data about the road on which the user' s vehicle is traveling. Thereby, the distance state calculator 13a can obtain more accurate road information by using a combination of the information obtained by the information collector 10 and the information obtained by the user's vehicle 10 state acquiring unit 36.
Moreover, the distance state calculator is always able to determine the positional relationship between the range of distance from the user's vehicle and the target vehicle. Thus, when the driver tries to, for example, perform the lane change 15 of the user's vehicle or turn the user's vehicle to the right or left, the distance state calculator can always determine whether to prompt the attention of the driver such that the driver can recognize the approach of the target vehicle to the user's vehicle. 20 [0110]
In Embodiment 2, a case has been described where the distance state calculator 13a performs a calculation on the basis of the measurement information obtained by the distance determiner 12 to thereby acquire the positional relationship 25 between a moving object and the one or more ranges of distance from the user's vehicle in each of which a notification of the approach of the moving object is to be provided. Nonetheless, the invention is not limited to the case in which the distance state calculator 13a itself calculates the above-mentioned 30 positional relationship.
62
Namely, the distance state calculator 13a can function as an acquiring unit in the invention, similarly to Embodiment 1, and may be configured to acquire the positional relationship by receiving information on the positional relationship between 5 the moving object and the range of distance in which a notification of the approach of the moving object is to be provided, from an external device that has been calculated the positional relationship.
In this case, the driving support apparatus 1A may not 10 have the vehicle detector 11 and the distance determiner 12. In addition, when the informing processor generates a predetermined image as the icon of the moving object regardless of the type of the moving object, the driving support apparatus 1A may not have the vehicle type identifier 14. 15 [0111]
As described above, according to Embodiment 2, the
informing processor 15 causes the image indicating the driving
maneuver requiring attention and the display item prompting the
attention, to appear in the image of the nearby surroundings
20 of the user's vehicle.
Thereby, the driver can intuitively recognize whether the driving maneuver of the user's vehicle requires the attention because the target vehicle approaches the user's vehicle. [0112] 25 According to Embodiment 2, the distance state calculator 13a performs a calculation to acquire the positional relationship between another vehicle and the range of distance from the user's vehicle by using at least one of the detected information about the road on which the user's vehicle is 30 traveling and the existing road data about the road on which
63
the user's vehicle is traveling. Thereby, even if the information collector 10 is not able to obtain the information about the road on which the user's vehicle is traveling, the distance state calculator can calculate the positional relationship based on the existing road data. [0113]
In this disclosure, it is to be understood that an arbitrary combination of two or more of the above-mentioned embodiments can be made, various changes can be made in an arbitrary component in accordance with any one of the above-mentioned embodiments, and an arbitrary component in accordance with any one of the above-mentioned embodiments can be omitted within the scope of the invention. INDUSTRIAL APPLICABILITY [0114]
When a driving support apparatus according to this disclosure is used, a driver can intuitively recognize a positional relationship between a user's vehicle and a moving object around the user's vehicle. Thus, the driving support apparatus according to this disclosure is suitable for supporting a driving maneuver to avoid contact or collision between the user's vehicle and the moving object. REFERENCE SIGNS LIST [0115]
1A: driving support apparatus; 10: information collector; 11: vehicle detector; 12: distance determiner; 13, 13a: distance state calculator; 14: vehicle type identifier; 15: informing processor; 16: informing device; 16A to 16C, 16E to 16J: images of nearby surroundings of a user's vehicle; 16D: display screen; 17: user's vehicle; 18a, 18b, 19a, 19b, 22a,
64
22b, 23a, 23b: threshold line; 20, 21, 32, 33, 34, 35: other vehicles; 24: residual fuel meter; 25: tachometer; 26: a current value of speed at which engine of user's vehicle turns; 27: attention mark; 28, 29: letters; 30, 31: blur; 36: user's 5 vehicle state acquiring unit; and 37: arrow.
65
CLAIMS
1. A driving support apparatus comprising:
an acquiring unit configured to acquire a positional relationship between at least one moving object and one or more ranges of distance from a user's vehicle, each of the one or more ranges of distance being a range in which a notification of approach of the moving object to the user's vehicle is to be provided; and
an informing processor configured to cause a display device to display an image of nearby surroundings of the user's vehicle including images of the user's vehicle and threshold lines that define the one or more ranges of distance, and configured to change the image of nearby surroundings of the user's vehicle in accordance with the positional relationship acquired by the acquiring unit.
2. The driving support apparatus according to claim 1, wherein the informing processor causes an image of the moving object to appear in the image of nearby surroundings of the user's vehicle, in accordance with the positional relationship acquired by the acquiring unit.
3. The driving support apparatus according to claim 2, further comprising a determiner configured to determine a type of the moving object, wherein
the informing processer causes the image of the moving object corresponding to the type determined by the determiner, to appear in the image of nearby surroundings of the user's vehicle.
66
4. The driving support apparatus according to claim 1,
wherein:
the one or more ranges of distance are one or more of ranges set at different regions depending upon distances from the user's vehicle; and
when the moving object moves into the one or more ranges of distance/ the informing processor changes display colors of the one or more ranges of distance or display colors of both an image of the moving object and the one or more ranges of distance in a manner that a display color of a close range of distance is different from a display color of another range of distance which is farther from the user's vehicle than said close range of distance.
5. The driving support apparatus according to claim 4, wherein, when moving objects as said at least one moving object move into the respective different ranges of distance, the informing processor changes display colors of the respective different ranges of distance into which the moving objects move, or display colors of both an image of the moving objects and the respective different ranges of distance into which the moving objects move, into colors corresponding to the respective different ranges of distance into which the moving objects move.
6. The driving support apparatus according to claim 4, wherein:
when the moving object moves into a first range of distance, the informing processor changes a display color of the first
67
range of distance or display colors of both an image of the moving object and the first range of distance, into a first display color; and
when the moving object moves into a second range of 5 distance being closer to the user's vehicle than the first range of distance, the informing processor changes a display color of the second range of distance or display colors of both an image of the moving object and the second range of distance, into a second display color being different from the first 10 display color, while changing the display color of the first range of distance through which the moving object moves into the second range of distance, into the second display color that differs in tone from the first display color.
15 7 . The driving support apparatus according to claim 1, wherein, when the moving object moves into the one or more ranges of distance, the informing processor causes the display device to display, in a display screen displaying the image of nearby surroundings of the user's vehicle, a translucent blur with a
20 color on an area corresponding to a direction from which the moving object approaches the user's vehicle,
the informing processor enlarging a size of the area displaying the translucent blur as the moving object approaches the user's vehicle.
25
8. The driving support apparatus according to claim 7, wherein the display screen of the display device is comprised of a plurality of image layers, wherein
the informing processor causes the translucent blur to
30 appear on one layer of the image layers as the moving object
68
approaches the user's vehicle, in a manner that a level of the one layer is lower than or equal to a level of another layer of the image layers which displays a predetermined important display item.
9. The driving support apparatus according to claim 1,
wherein:
the threshold lines that define the one or more ranges of distance arranged at a side of the user's vehicle appear in the image of nearby surroundings of the user's vehicle so as to extend along a road on which the user ' s vehicle is traveling; and
when the moving object moves into the one or more ranges of distance arranged at the side of the user's vehicle or in front or back of the user's vehicle, the informing processor changes a display color of a region which is defined by the extending threshold lines and includes the one or more ranges of distance.
10. The driving support apparatus according to claim 3, wherein, when the moving obj ect moves into the one or more ranges of distance, the informing processor changes a size and contents of a display item associated with drawing of attention of the driver, in a manner that a size and contents corresponding to one range of distance close to the user's vehicle is different from a size and contents corresponding to another range of distance which is farther from the user ' s vehicle than said one range of distance.
11. The driving support apparatus according to claim 2,
69
wherein the informing processor causes the display device to display the image of nearby surroundings of the user's vehicle representing a plan-view image of both the user's vehicle and the moving object or a bird's-eye-view image of both the user's vehicle and the moving object.
12. The driving support apparatus according to claim 11, wherein, as the moving object approaches the user's vehicle, the informing processor moves a location of a viewpoint from which the user's vehicle is viewed in a plan-view or a bird's-eye-view, closer to the user's vehicle.
13. The driving support apparatus according to claim 11, wherein, as the moving object approaches the user's vehicle, the informing processor moves, in a horizontal direction, a location of a viewpoint from which the user's vehicle is viewed in a plan-view or a bird' s-eye-view such that the image of nearby surroundings of the user's vehicle shows the moving object.
14. The driving support apparatus according to claim 1, wherein the informing processor causes both an image indicating a driving maneuver which requires attention and a display item associated with drawing of attention of the driver to appear in the image of nearby surroundings of the user's vehicle.
15. The driving support apparatus according to claim 1, wherein the acquiring unit acquires the positional relationship between said at least one moving object and the one or more ranges of distance from the user's vehicle, using a lane width of a traffic route which is specified based on at least one of
70
information detected from a road on which the user's vehicle is traveling and existing road data associated with the road.
16. A driving support method comprising: 5 in an acquiring unit, acquiring a positional relationship between at least one moving object and one or more ranges of distance from a user's vehicle, each of the one or more ranges of distance being a range in which a notification of approach of the moving object to the user's vehicle is to be provided;
10 and
in an informing processor, causing a display device to display an image of nearby surroundings of the user's vehicle including images of the user's vehicle and threshold lines that define the one or more ranges of distance, and changing the image
15 of nearby surroundings of the user's vehicle in accordance with the positional relationship acquired in the acquiring unit.
| # | Name | Date |
|---|---|---|
| 1 | 201747020254-IntimationOfGrant16-11-2022.pdf | 2022-11-16 |
| 1 | Translated Copy of Priority Document [09-06-2017(online)].pdf | 2017-06-09 |
| 2 | 201747020254-PatentCertificate16-11-2022.pdf | 2022-11-16 |
| 2 | PROOF OF RIGHT [09-06-2017(online)].pdf | 2017-06-09 |
| 3 | Power of Attorney [09-06-2017(online)].pdf | 2017-06-09 |
| 3 | 201747020254-FORM 3 [09-07-2021(online)].pdf | 2021-07-09 |
| 4 | Form 5 [09-06-2017(online)].pdf | 2017-06-09 |
| 4 | 201747020254-FORM 3 [07-12-2020(online)].pdf | 2020-12-07 |
| 5 | Form 3 [09-06-2017(online)].pdf | 2017-06-09 |
| 5 | 201747020254-2. Marked Copy under Rule 14(2) [01-06-2020(online)].pdf | 2020-06-01 |
| 6 | Form 18 [09-06-2017(online)].pdf_179.pdf | 2017-06-09 |
| 6 | 201747020254-ABSTRACT [01-06-2020(online)].pdf | 2020-06-01 |
| 7 | Form 18 [09-06-2017(online)].pdf | 2017-06-09 |
| 7 | 201747020254-CLAIMS [01-06-2020(online)].pdf | 2020-06-01 |
| 8 | Form 1 [09-06-2017(online)].pdf | 2017-06-09 |
| 8 | 201747020254-COMPLETE SPECIFICATION [01-06-2020(online)].pdf | 2020-06-01 |
| 9 | 201747020254-DRAWING [01-06-2020(online)].pdf | 2020-06-01 |
| 9 | Drawing [09-06-2017(online)].pdf | 2017-06-09 |
| 10 | 201747020254-FER_SER_REPLY [01-06-2020(online)].pdf | 2020-06-01 |
| 10 | Description(Complete) [09-06-2017(online)].pdf_178.pdf | 2017-06-09 |
| 11 | 201747020254-OTHERS [01-06-2020(online)].pdf | 2020-06-01 |
| 11 | Description(Complete) [09-06-2017(online)].pdf | 2017-06-09 |
| 12 | 201747020254-Retyped Pages under Rule 14(1) [01-06-2020(online)].pdf | 2020-06-01 |
| 12 | CLAIMS UNDER RULE 1 (PROVISIO) OF RULE 20 [09-06-2017(online)].pdf | 2017-06-09 |
| 13 | 201747020254-FORM 3 [26-05-2020(online)].pdf | 2020-05-26 |
| 13 | 201747020254.pdf | 2017-06-14 |
| 14 | 201747020254-FER.pdf | 2019-12-02 |
| 14 | Correspondence by Agent_Executed Form1_23-06-2017.pdf | 2017-06-23 |
| 15 | 201747020254-FORM 3 [25-04-2019(online)].pdf | 2019-04-25 |
| 15 | Other Document [05-07-2017(online)].pdf | 2017-07-05 |
| 16 | 201747020254-FORM 3 [11-05-2018(online)].pdf | 2018-05-11 |
| 16 | Marked Copy [05-07-2017(online)].pdf | 2017-07-05 |
| 17 | Form 13 [05-07-2017(online)].pdf | 2017-07-05 |
| 17 | 201747020254-FORM 3 [22-11-2017(online)].pdf | 2017-11-22 |
| 18 | Description(Complete) [05-07-2017(online)].pdf | 2017-07-05 |
| 18 | Description(Complete) [05-07-2017(online)].pdf_101.pdf | 2017-07-05 |
| 19 | Description(Complete) [05-07-2017(online)].pdf | 2017-07-05 |
| 19 | Description(Complete) [05-07-2017(online)].pdf_101.pdf | 2017-07-05 |
| 20 | 201747020254-FORM 3 [22-11-2017(online)].pdf | 2017-11-22 |
| 20 | Form 13 [05-07-2017(online)].pdf | 2017-07-05 |
| 21 | 201747020254-FORM 3 [11-05-2018(online)].pdf | 2018-05-11 |
| 21 | Marked Copy [05-07-2017(online)].pdf | 2017-07-05 |
| 22 | 201747020254-FORM 3 [25-04-2019(online)].pdf | 2019-04-25 |
| 22 | Other Document [05-07-2017(online)].pdf | 2017-07-05 |
| 23 | Correspondence by Agent_Executed Form1_23-06-2017.pdf | 2017-06-23 |
| 23 | 201747020254-FER.pdf | 2019-12-02 |
| 24 | 201747020254-FORM 3 [26-05-2020(online)].pdf | 2020-05-26 |
| 24 | 201747020254.pdf | 2017-06-14 |
| 25 | 201747020254-Retyped Pages under Rule 14(1) [01-06-2020(online)].pdf | 2020-06-01 |
| 25 | CLAIMS UNDER RULE 1 (PROVISIO) OF RULE 20 [09-06-2017(online)].pdf | 2017-06-09 |
| 26 | 201747020254-OTHERS [01-06-2020(online)].pdf | 2020-06-01 |
| 26 | Description(Complete) [09-06-2017(online)].pdf | 2017-06-09 |
| 27 | 201747020254-FER_SER_REPLY [01-06-2020(online)].pdf | 2020-06-01 |
| 27 | Description(Complete) [09-06-2017(online)].pdf_178.pdf | 2017-06-09 |
| 28 | 201747020254-DRAWING [01-06-2020(online)].pdf | 2020-06-01 |
| 28 | Drawing [09-06-2017(online)].pdf | 2017-06-09 |
| 29 | 201747020254-COMPLETE SPECIFICATION [01-06-2020(online)].pdf | 2020-06-01 |
| 29 | Form 1 [09-06-2017(online)].pdf | 2017-06-09 |
| 30 | Form 18 [09-06-2017(online)].pdf | 2017-06-09 |
| 30 | 201747020254-CLAIMS [01-06-2020(online)].pdf | 2020-06-01 |
| 31 | Form 18 [09-06-2017(online)].pdf_179.pdf | 2017-06-09 |
| 31 | 201747020254-ABSTRACT [01-06-2020(online)].pdf | 2020-06-01 |
| 32 | Form 3 [09-06-2017(online)].pdf | 2017-06-09 |
| 32 | 201747020254-2. Marked Copy under Rule 14(2) [01-06-2020(online)].pdf | 2020-06-01 |
| 33 | Form 5 [09-06-2017(online)].pdf | 2017-06-09 |
| 33 | 201747020254-FORM 3 [07-12-2020(online)].pdf | 2020-12-07 |
| 34 | Power of Attorney [09-06-2017(online)].pdf | 2017-06-09 |
| 34 | 201747020254-FORM 3 [09-07-2021(online)].pdf | 2021-07-09 |
| 35 | PROOF OF RIGHT [09-06-2017(online)].pdf | 2017-06-09 |
| 35 | 201747020254-PatentCertificate16-11-2022.pdf | 2022-11-16 |
| 36 | 201747020254-IntimationOfGrant16-11-2022.pdf | 2022-11-16 |
| 36 | Translated Copy of Priority Document [09-06-2017(online)].pdf | 2017-06-09 |
| 1 | search_26-11-2019.pdf |