WO2011108198A1 - 車両の周辺監視装置 - Google Patents
車両の周辺監視装置 Download PDFInfo
- Publication number
- WO2011108198A1 WO2011108198A1 PCT/JP2011/000803 JP2011000803W WO2011108198A1 WO 2011108198 A1 WO2011108198 A1 WO 2011108198A1 JP 2011000803 W JP2011000803 W JP 2011000803W WO 2011108198 A1 WO2011108198 A1 WO 2011108198A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- vehicle
- alarm
- objects
- display
- Prior art date
Links
- 238000012806 monitoring device Methods 0.000 title claims description 8
- 238000001514 detection method Methods 0.000 claims abstract description 14
- 238000003384 imaging method Methods 0.000 claims description 11
- 238000012544 monitoring process Methods 0.000 claims description 5
- 238000000034 method Methods 0.000 description 36
- 230000008569 process Effects 0.000 description 29
- 239000013598 vector Substances 0.000 description 14
- 238000012545 processing Methods 0.000 description 13
- 230000005484 gravity Effects 0.000 description 9
- 238000004364 calculation method Methods 0.000 description 8
- 238000013459 approach Methods 0.000 description 5
- 238000012937 correction Methods 0.000 description 5
- 238000005070 sampling Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 230000004397 blinking Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000002250 progressing effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
- B60R1/24—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view in front of the vehicle
-
- B60K35/29—
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/30—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles providing vision in the non-visible spectrum, e.g. night or infrared vision
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3602—Input other than that of destination using image analysis, e.g. detection of road signs, lanes, buildings, real preceding vehicles using a camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G1/00—Traffic control systems for road vehicles
- G08G1/16—Anti-collision systems
- G08G1/166—Anti-collision systems for active traffic, e.g. moving vehicles, pedestrians, bikes
-
- B60K2360/191—
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/106—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using night vision cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/20—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of display used
- B60R2300/207—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of display used using multi-purpose displays, e.g. camera image and navigation or video on same display
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/307—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing virtually distinguishing relevant parts of a scene from the background of the scene
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8093—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for obstacle warning
Definitions
- the present invention relates to an apparatus for monitoring the periphery of a vehicle, and more specifically to an apparatus for detecting and displaying an object around the vehicle.
- Patent Document 1 proposes a system that detects an object around a vehicle using an infrared camera, and notifies the driver using video and sound if the object is determined to be a pedestrian. Has been.
- a head-up display (HUD) is provided in a vehicle, and an image captured by an infrared camera is displayed on the HUD.
- HUD head-up display
- the head-up display since the head-up display is arranged in front of the driver, the head-up display has an advantage that the amount of line-of-sight movement is relatively small for the driver to visually recognize, but may not be mounted on a general vehicle. Many.
- the spread of display devices attached to dashboards is progressing. In particular, as navigation devices become more popular, devices are becoming more popular for displaying various information such as map information provided by a navigation function.
- a vehicle periphery monitoring device is mounted on a vehicle, an imaging unit that images the periphery of the vehicle, and an object around the vehicle based on an image captured by the imaging unit
- a detection means for detecting an object a means for displaying the captured image on a display device, and a warning display on the displayed image for the presence of the detected object
- Alarm means for performing an alarm In the first mode, the warning means highlights the position where the object exists for each of the objects when the number of objects detected by the detection means is equal to or less than a predetermined number at the same time. If the number of objects detected by the detection means is larger than the predetermined number at the same time, one alarm is displayed for the entire displayed image without performing the alarm display of the first form.
- a second form of alarm display that performs predetermined highlighting is performed.
- the present invention if the detected object is a relatively small number, highlighting is performed for each position of the object, so that the object is individually highlighted. However, if there are a relatively large number of detected objects, one highlighting is performed on the entire display image. Therefore, compared to a display form in which individual highlighting is performed when there are a large number of detected objects, the display form can be made easier to see and easier to judge. The driver can instantly recognize that there are many objects to be noted around the vehicle by performing one predetermined highlighting on the displayed image, and thus the displayed image is displayed. It is possible to pay attention to the driving operation without being confused by the image.
- the alarm means further has a predetermined image at a predetermined position in the displayed image when the number of objects simultaneously detected by the detection means is larger than the predetermined number. Is displayed in a superimposed manner, thereby performing highlighting according to the second embodiment.
- the predetermined image is highlighted and displayed at a predetermined position in the image. Therefore, compared to a display form in which individual highlighting is performed when there are a large number of detected objects, the display form can be made easier to see and easier to judge. The driver can instantly recognize that there are a plurality of objects to be noted around the vehicle by superimposing one predetermined image on the displayed image.
- the alarm means when the number of objects simultaneously detected by the detecting means is greater than the predetermined number, the alarm means further alerts all objects larger than the predetermined number.
- the emphasis display according to the second form is performed by determining the priority of the image and performing the emphasis display only on the position where the object determined to have the highest priority exists.
- highlighting is performed only on the position of the object determined to have the highest priority regarding the alarm among the objects. Therefore, compared to a display form in which individual highlighting is performed when there are a large number of detected objects, the display form can be made easier to see and easier to judge. In addition, since highlighting is performed only on an object having the highest priority, the driver can instantly recognize which object should be most careful.
- the display device is a display device of a car navigation device.
- the display device of the car navigation device can be effectively used to notify the driver of the presence of an object existing around the vehicle.
- the block diagram which shows the structure of the vehicle periphery monitoring apparatus according to one Example of this invention.
- 3 is a flowchart illustrating a process in an image processing unit according to an embodiment of the present invention.
- 7 is a flowchart showing a process in an image processing unit according to another embodiment of the present invention.
- FIG. 5 is a flowchart illustrating an alarm determination process according to an embodiment of the present invention.
- FIG. 1 is a block diagram showing a configuration of a vehicle periphery monitoring device using a display device of a navigation device according to an embodiment of the present invention
- FIG. 2 shows attachment of the display device and a camera to the vehicle.
- the vehicle is equipped with a navigation device, and the navigation device includes a navigation unit 5 and a display device 4.
- the display device 4 passes through the center of the vehicle handle (steering wheel) 21 and extends in the front-rear direction of the vehicle. It is attached so as to be visible to the driver at a position separated by a predetermined distance with respect to (shown to extend in the direction).
- the display device 4 is embedded in the dashboard 23 of the vehicle.
- the navigation unit 5 is realized by a computer having a central processing unit (CPU) and a memory.
- the navigation unit 5 receives, for example, a GPS signal for measuring the position of the vehicle 10 using an artificial satellite via a communication device (not shown) provided in the navigation unit 5, and receives the GPS signal. Based on this, the current position of the vehicle 10 is detected.
- the navigation unit 5 stores the current position in the map information around the vehicle (which can be stored in a storage device of the navigation device or can be received from a predetermined server via the communication device). Is displayed on the display screen 25 of the display device 4.
- the display screen 25 of the display device 4 constitutes a touch panel, and the occupant inputs the destination to the navigation unit 5 via the touch panel or another input device 27 such as a key or a button. Can do.
- the navigation unit 5 can calculate the optimal route of the vehicle to the destination, superimpose an image showing the optimal route on the map information, and display it on the display screen 25 of the display device 4.
- navigation devices are equipped with various other functions such as providing traffic information and facility guidance in the vicinity of the vehicle.
- any appropriate navigation device can be used. .
- the vehicle periphery monitoring device is mounted on the vehicle 10 and detects objects around the vehicle based on two infrared cameras 1R and 1L capable of detecting far infrared rays and image data captured by the cameras 1R and 1L. And a speaker 3 that generates an alarm with sound or voice based on the detection result.
- the display device 4 is used for displaying an image obtained through the imaging of the camera 1R or 1L and for causing the driver to recognize the presence of an object around the vehicle.
- the periphery monitoring device includes a yaw rate sensor 6 that detects the yaw rate of the vehicle, and a vehicle speed sensor 7 that detects the traveling speed (vehicle speed) of the vehicle, and the detection results of these sensors are sent to the image processing unit 2. It is done.
- the cameras 1R and 1L are symmetrical with respect to the central axis passing through the center of the vehicle width at the front portion of the vehicle 10 so as to image the front of the vehicle 10. It is arranged in the position.
- the two cameras 1R and 1L are fixed to the vehicle so that their optical axes are parallel to each other and their height from the road surface is equal.
- the infrared cameras 1R and 1L have a characteristic that the level of the output signal becomes higher (that is, the luminance in the captured image becomes higher) as the temperature of the object is higher.
- the image processing unit 2 includes an A / D conversion circuit that converts an input analog signal into a digital signal, an image memory that stores a digitized image signal, a central processing unit (CPU) that performs various arithmetic processing, and a data RAM (Random Access Memory) used to store data, ROM (Read Only Memory) that stores programs to be executed by the CPU and data to be used (including tables and maps), driving signals for the speaker 3, and display for the display device 4
- An output circuit for outputting signals and the like is provided.
- the output signals of the cameras 1R and 1L are converted into digital signals and input to the CPU.
- the display device 4 of the navigation device is used to display to the driver the display of an image obtained through imaging by the cameras 1R and 1L and the presence of a predetermined object detected from the image. Used for display for notification (alarm).
- the display device 4 is a predetermined distance from the handle 21 in the vehicle width direction. Since it is provided at a position far away from the HUD, the amount of line-of-sight movement is large for the driver to visually recognize the screen of the display device 4, and thus the time required for visual recognition is also long.
- a display that is easier to understand than a HUD, that is, a display that can be recognized in a shorter time is desired.
- the present invention makes it possible to display such a form.
- a predetermined number for example, 1
- the position of the object is determined.
- highlighting if more than a predetermined number of objects are present in the image, one highlighting is performed on the entire displayed image. This specific method will be described below.
- FIG. 3 is a flowchart showing a process executed by the image processing unit 2 according to the first embodiment of the present invention. The process is performed at predetermined time intervals.
- the output signals of the cameras 1R and 1L (that is, captured image data) are received as input, A / D converted, and stored in the image memory.
- the stored image data is a gray scale image including luminance information.
- step S14 the right image captured by the camera 1R is used as a reference image (alternatively, the left image may be used as a reference image), and the image signal is binarized. Specifically, a process of setting a region brighter than the luminance threshold value ITH to “1” (white) and a dark region to “0” (black) is performed.
- the luminance threshold value ITH can be determined by any appropriate technique.
- this binarization processing an object having a temperature higher than a predetermined temperature such as a living body is extracted as a white region.
- the luminance threshold value ITH can be determined by any appropriate technique.
- step S15 the binarized image data is converted into run-length data.
- the coordinates of the start point (the leftmost pixel of each line) of the white area (referred to as a line) in each pixel row, and the start point to the end point (each The run length data is represented by the length (expressed by the number of pixels) up to the pixel on the right end of the line.
- the y-axis is taken in the vertical direction in the image
- the x-axis is taken in the horizontal direction.
- steps S16 and S17 the object is labeled and the object is extracted. That is, of the lines converted into run length data, a line having a portion overlapping in the y direction is regarded as one object, and a label is given thereto. Thus, one or a plurality of objects are extracted.
- step S17 you may perform the determination process whether the extracted target object is a pedestrian.
- This determination process can be realized by any appropriate method. For example, using known pattern matching, the similarity between the object extracted as described above and a predetermined pattern representing a pedestrian is calculated, and if the similarity is high, the object is determined to be a pedestrian. be able to. Examples of such determination processing are described in, for example, Japanese Patent Application Laid-Open Nos. 2007-241740 and 2007-334751.
- step S31 it is determined whether or not the object (as described above, a pedestrian here) is located within a predetermined alarm range from the vehicle.
- a predetermined alarm range is shown in FIG.
- the area AR0 indicates an imaging range that is an area that can be imaged by the cameras 1R and 1L.
- the processing in steps S11 to S17 in FIG. 3 is performed on the captured image corresponding to the imaging range AR0.
- the area AR1 corresponds to an area corresponding to a range obtained by adding a margin ⁇ (for example, about 50 to 100 cm) on both sides of the vehicle width ⁇ of the vehicle 10, in other words, the central axis of the vehicle 10 in the vehicle width direction. It is an area having a width of ( ⁇ / 2 + ⁇ ) on both sides, and is an approach determination area with a high possibility of collision if the object continues to exist as it is.
- the areas AR2 and AR3 are areas where the absolute value of the X coordinate is larger than the approach determination area (outside in the lateral direction of the approach determination area), and an intrusion determination that may cause an object in this area to enter the access determination area It is an area.
- Z1 that defines the size of the areas AR1 to AR3 in the distance direction can be set to a predetermined value in this embodiment, for example.
- the determination in step S31 is performed with the approach determination area AR1 as the alarm range.
- the position of the center of gravity of the object may be obtained, and if the position of the center of gravity is included in the alarm range, it may be determined that the object is located in the alarm range.
- step S31 if no object is located within the alarm range, the process proceeds to step S32, and alarm output is prohibited. That is, normal display is performed in which the grayscale image acquired in step S13 is output on the display device 4 without outputting an alarm.
- step S32 determines whether the number of objects determined to be located within the alarm range is greater than a predetermined number (for example, 1). To do. If the number of objects is not greater than the predetermined number (that is, if it is equal to or less than the predetermined number), the determination in step S33 is No and the process proceeds to step S34 to output an alarm. Specifically, a gray scale image is displayed on the display device 4 and an alarm output is realized by individually highlighting each object position within the alarm range in the image. .
- the form in which the position of each object is individually highlighted is referred to as a first form.
- step S32 If it is determined in step S32 that the number of objects determined to be within the alarm range is greater than the predetermined number, the process proceeds to step S35 and an alarm is output. Specifically, a grayscale image is displayed on the display device 4, and one highlight display is performed on the entire display image, thereby realizing an alarm output.
- a form in which one highlighted display is performed on the entire displayed image is referred to as a second form.
- the first mode is to output an alarm in a mode (individual emphasis mode) that highlights the position of the target for each target. It is.
- the second form is a form in which an alarm is output in a form in which one emphasis display is performed for each image displayed on the display device 4 (an overall emphasis form).
- step S33 a schematic example of temporal transition of a screen displayed on the display device 4 according to the process of FIG. 4 is shown.
- An area surrounded by the line 101 and the line 102 indicates an alarm range.
- the “predetermined number” in step S33 is a value of 1.
- the actual grayscale image displayed on the display device 4 is displayed with a higher brightness than the background (that is, white). It should be noted that it is drawn so as not to follow such an actual display mode (thus, in the figure, the pedestrian is drawn blacker than the background). The same applies to the following drawings.
- (B) After that, at time t2 of (b), it is determined that five pedestrians 111 to 115 exist within the warning range. Since the number of pedestrians to be alarmed is 5, display in the second form, that is, display in the form of overall emphasis is performed.
- one icon image as indicated by reference numeral 121 is superimposed and displayed on the display image.
- the icon image 121 is a single highlight displayed on the entire display image.
- a person is drawn in a triangle, which indicates that a large number of pedestrians are present in the vicinity.
- the icon image 121 is preferably displayed so as to draw the driver's attention. For example, a triangle can be displayed in red, and the icon image can be lit or blinked.
- the alarm output is switched between the first form and the second form depending on whether or not the number of objects is larger than a predetermined number.
- one highlighting is performed on the entire display image.
- individual highlighting such as a frame is performed for each position of a plurality of objects. If there are many such highlights, the entire screen becomes difficult to see. It can be difficult for the driver to quickly determine which object to watch out for.
- one highlighting is performed for each display image, so that the display form is easier for the driver to see and easier to judge. Therefore, the driver can instantly recognize that there are a plurality of objects to be noted around by visually recognizing the one highlight.
- the second form of performing one highlight display on the entire display image is realized by superimposing and displaying an icon image as indicated by reference numeral 121.
- the overall highlighting is not limited to this, and may take other forms.
- FIG. 7 shows another example of overall highlighting.
- the entire frame according to the second mode is emphasized by displaying the outer frame 131 so as to surround the outer edge of the displayed image.
- the outer frame 131 can be colored to attract the driver's attention (for example, red or yellow), and the outer frame 131 can be lit or blinked.
- the overall brightness of the captured image is lowered (for example, the brightness value of each pixel is lowered by a predetermined value), and the image is displayed in a form with low contrast.
- the icon image 121 is superimposed and displayed.
- only the icon image 121 is displayed without displaying the captured image.
- the icon images 123 of (b) and (c) can also be turned on or blinked. As shown in (b) and (c), by making it difficult to see the entire captured image together with the display of the icon image 121, the driver can be urged to gaze at the front.
- FIG. 8 is a flowchart showing a process executed by the image processing unit 2 according to the second embodiment of the present invention. The process is performed at predetermined time intervals.
- Steps S11 to S17 are the same as those in FIG.
- Steps S18 to S26 are processes for calculating the relative movement vector of the object. Details of the relative movement vector calculation process are described in, for example, Japanese Patent Laid-Open No. 2001-6096, and will be briefly described here.
- step S18 for each of the objects extracted in step S17, the center of gravity G, area S of the object, and the aspect ratio ASPECT of the rectangle circumscribing the object, that is, the circumscribed rectangle, are calculated.
- the area S is calculated by integrating the lengths of run length data for the same object.
- the coordinates of the center of gravity G are calculated as the x coordinate of a line that bisects the area S in the x direction and the y coordinate of a line that bisects the area S in the y direction.
- the aspect ratio ASPECT is calculated as a ratio Dy / Dx between the length Dy in the y direction and the length Dx in the x direction of the circumscribed square. Note that the position of the center of gravity G may be substituted by the position of the center of gravity of the circumscribed rectangle.
- step S19 tracking of the target object for the time (tracking), that is, recognition of the same target object is performed at every predetermined sampling period.
- the sampling period may be the same as the period in which the process is executed. Specifically, when the time obtained by discretizing the time t as an analog quantity with the sampling period is k and the object A is extracted at the time k, the object A and the time at the next sampling period are considered. The identity with the object B extracted at (k + 1) is determined. The identity determination can be performed according to a predetermined condition.
- the difference between the x and y coordinates of the position of the center of gravity G on the images of the objects A and B is smaller than a predetermined allowable value
- the area on the image of the object B on the image of the object A If the ratio of the aspect ratio of the circumscribed rectangle of the object B to the aspect ratio of the circumscribed rectangle of the object A is smaller than the predetermined allowable value, the objects A and B Can be determined to be the same.
- the position of the object (in this embodiment, the position coordinates of the center of gravity G) is stored in the memory as time-series data together with the assigned label.
- step S20 the vehicle speed VCAR detected by the vehicle speed sensor 7 and the yaw rate YR detected by the yaw rate sensor 6 are read, and the yaw rate YR is integrated over time to calculate the turning angle of the vehicle 10, that is, the turning angle ⁇ r.
- steps S21 to S23 a process of calculating the distance z from the vehicle 10 to the object is performed in parallel with the processes of steps S19 and S20. Since this calculation requires a longer time than steps S19 and S20, it may be executed in a cycle longer than steps S19 and S20 (for example, a cycle about three times the execution cycle of steps S11 to S20).
- step S21 one of the objects to be tracked by the binarized image of the reference image (in this example, the right image) is selected, and this is surrounded by a search image R1 (here, a circumscribed rectangle). Let the image area be a search image).
- step S22 an image of the same object as the search image R1 (hereinafter referred to as a corresponding image) is searched for in the left image. Specifically, it can be performed by executing a correlation calculation between the search image R1 and the left image.
- the correlation calculation is performed according to the following formula (1). This correlation calculation is performed using a grayscale image, not a binary image.
- the search image R1 has M ⁇ N pixels, and IR (m, n) is a luminance value at the position of the coordinates (m, n) in the search image R1, and IL (a + m ⁇ M , B + n ⁇ N) is a luminance value at the position of the coordinates (m, n) in the local area having the same shape as the search image R1 with the predetermined coordinates (a, b) in the left image as a base point.
- the position of the corresponding image is specified by determining the position where the luminance difference sum C (a, b) is minimized by changing the coordinates (a, b) of the base point.
- a region to be searched may be set in advance, and a correlation calculation may be performed between the search image R1 and the region.
- step S23 the distance dR (number of pixels) between the centroid position of the search image R1 and the image center line (a line that bisects the captured image in the x direction) LCTR of the captured image, the centroid position of the corresponding image, and the image center line LCTR Distance dL (number of pixels) is obtained and applied to Equation (2) to calculate the distance z to the object of the vehicle 10.
- B is the base line length, that is, the distance in the x direction (horizontal direction) between the center position of the image sensor of the camera 1R and the center position of the image sensor of the camera 1L (that is, the distance between the optical axes of both cameras).
- F represents the focal length of the lenses provided in the cameras 1R and 1L
- p represents the pixel interval of the image sensors of the cameras 1R and 1L.
- step S24 the coordinates (x, y) in the image of the position of the object (as described above, the position of the center of gravity G in this embodiment) and the distance z calculated by the expression (2) are expressed in the expression (3).
- the real space coordinates (X, Y, Z) are set to the origin O as the midpoint position (position fixed to the vehicle) of the attachment positions of the cameras 1R and 1L.
- the coordinates on the image are expressed in a coordinate system in which the center of the image is the origin, the horizontal direction is the x axis, and the vertical direction is the y axis.
- (xc, yc) is the coordinate (x, y) on the right image based on the relative position relationship between the mounting position of the camera 1R and the origin O of the real space coordinate system and the image. Are converted into coordinates in a virtual image that coincides with the center of the image.
- F is a ratio between the focal length F and the pixel interval p.
- step S25 a turning angle correction is performed to correct a positional shift on the image due to the turning of the vehicle.
- the vehicle 10 turns, for example, by a turning angle ⁇ r in the left direction, the image obtained by the camera is shifted in the x direction (positive direction) by ⁇ x. Therefore, this is corrected.
- the real space coordinates (X, Y, Z) are applied to Equation (4) to calculate the corrected coordinates (Xr, Yr, Zr).
- the calculated real space position data (Xr, Yr, Zr) is stored in the memory in time series in association with each object.
- the corrected coordinates are indicated as (X, Y, Z).
- 1) indicating the direction of the approximate straight line LMV, the straight line represented by the equation (5) is obtained.
- u is a parameter that takes an arbitrary value.
- Xav, Yav, and Zav are the average value of the X coordinate, the average value of the Y coordinate, and the average value of the Z coordinate of the real space position data string, respectively.
- FIG. 10 is a diagram for explaining the approximate straight line LMV.
- P (0), P (1), P (2),..., P (N-2), P (N-1) represent time-series data after turning angle correction
- the numerical value in () attached to P indicating the coordinates of each data point indicates that the data is past data as the value increases.
- P (0) indicates the latest position coordinates
- P (1) indicates the position coordinates one sample period before
- P (2) indicates the position coordinates two sample periods before.
- X (j), Y (j), Z (j) and the like in the following description.
- a more detailed method for calculating the approximate straight line LMV is described in Japanese Patent Laid-Open No. 2001-6096.
- the corrected position coordinates Pv (0) (Xv (0) are obtained by using the expression (6) obtained by applying the Z coordinates Z (0) and Z (n ⁇ 1) to the expression (5a).
- Yv (0), Zv (0)) and Pv (N-1) (Xv (N-1), Yv (N-1), Zv (N-1)).
- a vector from the position coordinates Pv (N ⁇ 1) calculated by Expression (6) toward Pv (0) is calculated as a relative movement vector.
- step S27 an alarm determination is performed using the relative movement vector.
- FIG. 11 is a flowchart of the alarm determination process executed in step S27.
- step S41 it is determined whether or not the number of objects extracted from the captured image and having the relative movement vector calculated as described above is greater than a predetermined number (for example, 1). If this determination is No, in step S42, the alarm is output in the first form described above. That is, the grayscale image acquired in step S13 is displayed on the display device 4, and an alarm output is realized by performing individual emphasis display on the position of each object in the grayscale image.
- a predetermined number for example, a predetermined number (for example, 1).
- step S41 if the number of objects is greater than the predetermined number, the process from step S43 is executed, and among these objects, the object with the highest priority for the alarm is determined.
- step S43 for each object, the arrival time TTC of the object to the vehicle is compared with a predetermined value TTC_TH.
- the relative speed Vs of the object with respect to the host vehicle in the Z direction is calculated using the Zv value, as shown in Equation (7).
- the arrival time TTC is calculated as shown in the equation (8), and whether or not this is less than a predetermined value TTC_TH (for example, 2 to 5 seconds) as shown in the equation (9). to decide.
- Vs (Zv (N ⁇ 1) ⁇ Zv (0)) / ⁇ T (7)
- TTC Zv (0) / Vs (8)
- Zv (0) is attached to indicate that the latest distance detection value (v is data after correction by the approximate straight line LMV, but the Z coordinate is the same value as before correction.
- Zv (N ⁇ 1) is a distance detection value before time ⁇ T. In this way, it is determined whether each object has a possibility of reaching the host vehicle in the Z direction within a predetermined time TTC_TH.
- the predetermined distance Z1 between the areas AR1 to AR3 shown in FIG. 4 corresponds to Vs ⁇ TTC_TH. Therefore, the objects that satisfy the equation (9) are objects that exist in the areas AR1 to AR3.
- step S43 If there is no target that satisfies Equation (9), it indicates that there is no target that may reach the host vehicle in the Z direction within the predetermined time TTC_TH, and therefore the determination in step S43 is No. Proceeding to step S44, it is determined that there is no high priority object. In this case, the process proceeds to step S45, and alarm output is prohibited. That is, normal display is performed in which the grayscale image acquired in step S13 is output on the display device 4 without outputting an alarm.
- step S43 determines whether or not the moving direction of the object is toward the host vehicle for each of the objects that satisfy Expression (9).
- the Xv value of the relative movement vector of the object is used.
- the relative movement vector represents the movement from Xv (N ⁇ 1) to Xv (0) in the X direction within the period of ⁇ T. Further, as shown in FIG. 9, the X coordinate value has the position of the host vehicle as the origin.
- step S48 the target object determined to be heading toward the host vehicle is determined to have high priority.
- the determination in step S47 is further performed, and the Xv (0) of the target object is a predetermined range (X_TH1 to X_TH2) centered on the host vehicle. To determine whether it is located between.
- X_TH1 ⁇ Xv (0) ⁇ X_TH2 it indicates that the target object may not collide with the host vehicle but may collide with the host vehicle if it exists as it is. Therefore, if there is at least one object for which X_TH1 ⁇ Xv (0) ⁇ X_TH2 is established, the process proceeds to step S48, and it is determined that the priority of the object is high. If there is no target that satisfies X_TH1 ⁇ Xv (0) ⁇ X_TH2, the process proceeds to step S44, and it is determined that there is no target with high priority.
- step S48 the arrival time TTC is less than or equal to the predetermined value and the object heading toward the own vehicle, and the arrival time TTC is less than or equal to the predetermined value and the predetermined range in the vehicle width direction centering on the own vehicle It is determined that the object existing inside has a high priority.
- step S49 when there are a plurality of objects determined to have high priority in step S48, the object having the highest priority is determined. This is performed in order to determine an object that the driver should be most careful about, and any appropriate method can be used as a technique for that purpose.
- the arrival time TTC and the latest X value that is, Xv (0) are used as parameters, and the degree of collision possibility is expressed by the parameters.
- a value obtained by multiplying the parameter TTC value and the absolute value of the parameter Xv (0) can be considered as an index representing the degree of collision possibility.
- the degree of possibility of collision may be expressed using the arrival time TTC, Xv (0), and the velocity in the X direction as parameters.
- the velocity Xs in the X direction can be calculated by (Xv (N ⁇ 1) ⁇ Xv (0)) / ⁇ T.
- the possibility of collision is obtained by multiplying the parameter TTC value, the absolute value of the parameter Xv (0), and the inverse of the speed Xs in the X direction.
- a value representing the degree of the calculation can be calculated, and the object having the smallest multiplication value can be determined as the object having the highest priority.
- any of the parameters of the arrival time TTC, Xv (0), and the velocity in the X direction may be used.
- other parameters for example, the position Zv (0) of the object in the Z direction
- FIG. 12 a state in which a plurality of objects 151 to 154 are present in the imaging range AR0 of the vehicle 10 is shown.
- the arrows attached to the objects 151 to 154 indicate the relative movement vectors of the objects.
- TTC_TH a predetermined value
- the area between the line passing through X_TH1 and the line passing through X_TH2 on the X axis in the figure corresponds to the approach determination area AR1 in FIG.
- the objects 151 to 153 are moving toward the host vehicle in the X direction, as can be seen from the directions of their relative movement vectors. Therefore, the determination in step S46 is Yes for these objects. Further, although the object 154 is not moving toward the host vehicle in the X direction, it is determined in step S47 that the object 154 exists within the predetermined range X_TH1 to X_TH2. Therefore, all of the objects 151 to 154 are determined to be high priority objects.
- the object having the highest priority among the objects 151 to 154 is determined. For example, as described above, using the arrival time TTC and the absolute value of Xv (0) as parameters, a value indicating the degree of collision possibility is calculated for each object, and the value having the highest degree of collision possibility is calculated.
- the determined object can be determined as the highest priority object.
- step S50 if the highest priority object is determined in this way, the process proceeds to step S50, and an alarm is output in the second form. Specifically, a grayscale image is displayed on the display device 4 and highlighting is performed only on the position of the object determined to have the highest priority in the image. In this way, one highlighting is performed on the entire displayed image.
- the highlighting is performed only for the object with the highest priority, i.e. the driver should be most careful.
- the display form can be made easier for the driver to see and easier to judge. Therefore, the driver can instantly recognize the object to be most careful.
- FIG. 13 shows a schematic example of the temporal transition of the screen displayed on the display device 4 according to the process of FIG.
- the regions surrounded by the lines 161 and 162 correspond to the lines indicating X_TH1 and X_TH2 shown in step S47 and FIG. 12, respectively.
- the “predetermined number” in step S41 is assumed to be a value 1.
- step S41 Since the number of detected pedestrians 171 is 1 at time t1 in (a), the determination in step S41 is No, and an alarm output according to the first mode is performed. Therefore, warning output is performed by highlighting the position of the pedestrian with a frame.
- the frame according to the first form in the case of (a) and the frame according to the second form in the case of (b) may be different in color, shape and the like.
- the frame according to the second form can be made to have a color and shape that draws the driver's attention more than the frame according to the first form.
- the former may be a frame display with a thin red line
- the latter may be a frame display with a red thick line
- the former may be a red frame display
- the latter may be blinking a red frame display. It may be a thing.
- highlighting is performed by adding a frame to the object to be alarmed, but a form different from the frame may be used.
- highlighting may be performed by superimposing another predetermined image such as an icon on the position of the object.
- the highest priority object is highlighted and the outer frame 131 in FIG. 7A is displayed and lit or blinked.
- the luminance of the area other than the object with the highest priority may be lowered.
- the image is displayed with a high contrast, and the driver's attention to the object with the highest priority is further increased. Can be strongly aroused.
- step S41 in FIG. 10 when step S41 in FIG. 10 is No (that is, the number of objects is equal to or less than a predetermined number), the object exists within the predetermined alarm range described in the first embodiment. Determine whether or not. If it is within the alarm range, the alarm is output in the first form as described in step S34 in FIG.
- step S41 in FIG. 10 is Yes (ie, there are more objects than the predetermined number)
- the object is determined to be a predetermined object. Determine if it is within the alarm range. If it is within the alarm range, as described in step S35 of FIG.
- Step S41 is Yes and it is determined that there is an object with a high priority, in Step S50 of FIG. 10, 1 for only the object with the highest priority as shown in FIG.
- the alarm output according to the second form is performed by two emphasis displays.
- the predetermined number is not limited to 1, but it is preferable to check in advance how many emphasis displays are made to feel that the screen is difficult to see, and is preferably set to a number smaller than the emphasis display number at which it is difficult to see.
- the display device 4 uses a display device of a navigation device.
- a display form that can be recognized by the driver in a short time is adopted, even if a display device arranged on the left or right side of the driver is used, an object that avoids the influence on the driving operation. An object notification can be performed.
- other display devices may be used, and the display form of the present invention may be applied to a head-up display as in the past.
- an alarm sound may be output via the speaker 3 when the alarm is output in the processes of FIGS. 3 and 10.
- the alarm sound may be arbitrary, may be a simple buzzer sound, or may be a voice message.
- the present invention can be applied to other cameras (for example, a visible camera).
Abstract
Description
(X-Xav)/lx=(Y-Yav)/ly=(Z-Zav)/lz
(5a)
Vs=(Zv(N-1)―Zv(0))/ΔT (7)
TTC=Zv(0)/Vs (8)
TTC≦TTC_TH (9)
2 画像処理ユニット
3 スピーカ
4 表示装置
Claims (4)
- 車両に搭載され、該車両の周辺を撮像する撮像手段と、
前記撮像手段により撮像された画像に基づいて、前記車両の周辺の対象物を検出する検出手段と、
前記撮像された画像を表示装置上に表示する手段と、
前記検出された対象物の存在について、前記表示される画像上で警報表示を行うことにより、運転者に警報を行う警報手段と、を備える車両の周辺監視装置であって、
前記警報手段は、同時に前記検出手段によって検出された対象物の数が所定数以下である場合には、該対象物のそれぞれについて、該対象物が存在する位置に対して強調表示を行う第1の形態の警報表示を行い、同時に前記検出手段によって検出された対象物の数が前記所定数より多い場合には、前記第1の形態の警報表示を行うことなく、前記表示される画像全体に対して1つの所定の強調表示を行う第2の形態の警報表示を行う、ことを特徴とする周辺監視装置。 - 前記警報手段は、さらに、同時に前記検出手段によって検出された対象物の数が前記所定数より多い場合には、前記表示される画像内の所定位置に、所定の画像を重畳表示させることによって、前記第2の形態による強調表示を行う、
請求項1に記載の周辺監視装置。 - 前記警報手段は、さらに、同時に前記検出手段によって検出された対象物の数が前記所定数より多い場合には、該所定数より多い対象物全てに対して、警報についての優先度を判定し、最も優先度が高いと判定された対象物が存在する位置に対してのみ強調表示を行うことによって、前記第2の形態による強調表示を行う、
請求項1に記載の周辺監視装置。 - 前記表示装置は、カーナビゲーション装置の表示装置である、
請求項1から3のいずれかに記載の周辺監視装置。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11750312.8A EP2544161B1 (en) | 2010-03-03 | 2011-02-14 | Surrounding area monitoring device for vehicle |
CN201180011493.3A CN102782741B (zh) | 2010-03-03 | 2011-02-14 | 车辆周围监测装置 |
US13/580,864 US9073484B2 (en) | 2010-03-03 | 2011-02-14 | Surrounding area monitoring apparatus for vehicle |
JP2012502982A JP5706874B2 (ja) | 2010-03-03 | 2011-02-14 | 車両の周辺監視装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010-047161 | 2010-03-03 | ||
JP2010047161 | 2010-03-03 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011108198A1 true WO2011108198A1 (ja) | 2011-09-09 |
Family
ID=44541869
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/000803 WO2011108198A1 (ja) | 2010-03-03 | 2011-02-14 | 車両の周辺監視装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US9073484B2 (ja) |
EP (1) | EP2544161B1 (ja) |
JP (1) | JP5706874B2 (ja) |
CN (1) | CN102782741B (ja) |
WO (1) | WO2011108198A1 (ja) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013098919A (ja) * | 2011-11-04 | 2013-05-20 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2015128201A (ja) * | 2013-12-27 | 2015-07-09 | トヨタ自動車株式会社 | 車両用情報表示装置及び車両用情報表示方法 |
JP2017016457A (ja) * | 2015-07-02 | 2017-01-19 | パナソニックIpマネジメント株式会社 | 表示制御装置、投影装置、表示制御プログラム、および記録媒体 |
WO2017022496A1 (ja) * | 2015-08-04 | 2017-02-09 | 株式会社デンソー | 運転者に支援画像を提示する装置及びその方法 |
JP2017074871A (ja) * | 2015-10-15 | 2017-04-20 | 日立建機株式会社 | 車両周囲障害物検出装置 |
JP2017138817A (ja) * | 2016-02-04 | 2017-08-10 | トヨタ自動車株式会社 | 車両用注意喚起装置 |
JP2018508893A (ja) * | 2015-02-27 | 2018-03-29 | トヨタ モーター エンジニアリング アンド マニュファクチャリング ノース アメリカ,インコーポレイティド | 状況環境情報を提供するためのシステム及び方法 |
JP2018097704A (ja) * | 2016-12-15 | 2018-06-21 | トヨタ自動車株式会社 | 車両用注意喚起装置 |
JP2019089394A (ja) * | 2017-11-13 | 2019-06-13 | 矢崎総業株式会社 | 自動運転時情報伝達方法および車両用情報提示装置 |
JP2020024570A (ja) * | 2018-08-07 | 2020-02-13 | 日立オートモティブシステムズ株式会社 | 運転支援装置 |
WO2024042762A1 (ja) * | 2022-08-23 | 2024-02-29 | マクセル株式会社 | ヘッドアップディスプレイ装置および映像データの処理方法 |
Families Citing this family (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2827318A4 (en) * | 2012-03-12 | 2016-01-13 | Honda Motor Co Ltd | DEVICE FOR MONITORING VEHICLE PERIPHERY |
DE102012210798A1 (de) * | 2012-06-26 | 2014-01-02 | Robert Bosch Gmbh | Kommunikationsvorrichtung und Kommunikationsverfahren für ein Fahrzeug |
JP5888164B2 (ja) | 2012-07-20 | 2016-03-16 | トヨタ自動車株式会社 | 車両周辺監視装置、車両周辺監視システム |
JP5957359B2 (ja) * | 2012-10-19 | 2016-07-27 | 日立オートモティブシステムズ株式会社 | ステレオ画像処理装置及びステレオ画像処理方法 |
US10547774B2 (en) * | 2013-01-09 | 2020-01-28 | Sony Corporation | Image processing device, image processing method, and program |
US9514650B2 (en) * | 2013-03-13 | 2016-12-06 | Honda Motor Co., Ltd. | System and method for warning a driver of pedestrians and other obstacles when turning |
JP5991647B2 (ja) | 2013-03-28 | 2016-09-14 | 株式会社デンソー | 車両用周辺監視制御装置 |
JP6232759B2 (ja) * | 2013-06-07 | 2017-11-22 | ソニー株式会社 | 情報処理装置、接近対象物通知方法及びプログラム |
JP6205923B2 (ja) * | 2013-07-11 | 2017-10-04 | 株式会社デンソー | 走行支援装置 |
TW201529381A (zh) * | 2014-01-17 | 2015-08-01 | Primax Electronics Ltd | 行車影像輔助系統 |
CN104802708A (zh) * | 2014-01-23 | 2015-07-29 | 致伸科技股份有限公司 | 行车影像辅助系统 |
JP6267972B2 (ja) * | 2014-01-23 | 2018-01-24 | 日立建機株式会社 | 作業機械の周囲監視装置 |
US9598012B2 (en) * | 2014-03-11 | 2017-03-21 | Toyota Motor Engineering & Manufacturing North America, Inc. | Surroundings monitoring system for a vehicle |
KR102310286B1 (ko) * | 2014-11-07 | 2021-10-07 | 현대모비스 주식회사 | 특정물체 감지 장치 및 특정물체 감지 방법 |
CN107004363B (zh) * | 2014-12-10 | 2020-02-18 | 三菱电机株式会社 | 图像处理装置及车载显示系统及显示装置及图像处理方法 |
WO2016143306A1 (ja) | 2015-03-10 | 2016-09-15 | 株式会社Jvcケンウッド | 警告装置、警告方法及び警告プログラム |
JP6224029B2 (ja) * | 2015-05-21 | 2017-11-01 | 富士通テン株式会社 | 画像処理装置および画像処理方法 |
US20180090002A1 (en) * | 2015-08-03 | 2018-03-29 | Mitsubishi Electric Corporation | Display control apparatus, display device, and display control method |
US10140870B1 (en) | 2015-09-30 | 2018-11-27 | Waymo Llc | Occupant facing vehicle display |
US10906530B2 (en) | 2015-11-10 | 2021-02-02 | Hyundai Motor Company | Automatic parking system and automatic parking method |
US10919574B2 (en) | 2015-11-10 | 2021-02-16 | Hyundai Motor Company | Automatic parking system and automatic parking method |
KR101892026B1 (ko) * | 2015-11-10 | 2018-08-27 | 현대자동차주식회사 | 차량의 원격 주차 제어 방법 및 장치 |
US10606257B2 (en) | 2015-11-10 | 2020-03-31 | Hyundai Motor Company | Automatic parking system and automatic parking method |
EP3381180B1 (en) * | 2015-11-24 | 2023-07-05 | Samsung Electronics Co., Ltd. | Photographing device and method of controlling the same |
KR101916993B1 (ko) * | 2015-12-24 | 2018-11-08 | 엘지전자 주식회사 | 차량용 디스플레이 장치 및 그 제어방법 |
CN109844839A (zh) * | 2016-06-10 | 2019-06-04 | 大陆汽车系统公司 | 用于自主车道变换操纵的状况分析的系统和方法 |
CN107016495A (zh) * | 2017-03-21 | 2017-08-04 | 乐蜜科技有限公司 | 城市区域相关性的确定方法、装置和终端设备 |
JP6819431B2 (ja) * | 2017-04-12 | 2021-01-27 | トヨタ自動車株式会社 | 注意喚起装置 |
JP6834860B2 (ja) * | 2017-09-05 | 2021-02-24 | トヨタ自動車株式会社 | 衝突防止装置、衝突防止方法、衝突防止プログラム、記録媒体 |
CN109455140B (zh) * | 2017-09-06 | 2022-03-01 | 丰田自动车株式会社 | 图像显示装置 |
WO2019058492A1 (ja) * | 2017-09-22 | 2019-03-28 | マクセル株式会社 | 表示システム及び表示方法 |
CN108172004A (zh) * | 2017-11-30 | 2018-06-15 | 广州华夏职业学院 | 一种用于车辆中红绿灯提醒装置及系统 |
CN108693548B (zh) * | 2018-05-18 | 2021-10-22 | 中国科学院光电研究院 | 一种基于场景目标识别的导航方法及系统 |
KR20200005282A (ko) | 2018-07-06 | 2020-01-15 | 현대모비스 주식회사 | 미러리스 자동차의 측방 영상 처리 장치 및 방법 |
JP7143661B2 (ja) * | 2018-07-24 | 2022-09-29 | トヨタ自動車株式会社 | 情報処理システム、プログラム、及び制御方法 |
JP7180421B2 (ja) * | 2019-02-04 | 2022-11-30 | トヨタ自動車株式会社 | 車両制御装置 |
US11351991B2 (en) * | 2019-03-25 | 2022-06-07 | Zoox, Inc. | Prediction based on attributes |
WO2020206380A1 (en) * | 2019-04-05 | 2020-10-08 | FLIR Belgium BVBA | Passage planning and navigation systems and methods |
US11488472B2 (en) * | 2020-10-15 | 2022-11-01 | Ford Global Technologies, Llc | Vehicle sound emission control |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001006096A (ja) | 1999-06-23 | 2001-01-12 | Honda Motor Co Ltd | 車両の周辺監視装置 |
JP2003284057A (ja) | 2002-01-18 | 2003-10-03 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2007159036A (ja) * | 2005-12-08 | 2007-06-21 | Denso Corp | 車両用表示装置 |
JP2007241740A (ja) | 2006-03-09 | 2007-09-20 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2007334751A (ja) | 2006-06-16 | 2007-12-27 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2009098956A (ja) * | 2007-10-17 | 2009-05-07 | Mazda Motor Corp | 車両用運転支援装置 |
Family Cites Families (70)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3738949B2 (ja) * | 1999-07-08 | 2006-01-25 | 三菱自動車工業株式会社 | 車両周辺領域監視装置 |
US6281806B1 (en) * | 2000-10-12 | 2001-08-28 | Ford Global Technologies, Inc. | Driver road hazard warning and illumination system |
US7180475B2 (en) * | 2001-06-07 | 2007-02-20 | Infocus Corporation | Method and apparatus for wireless image transmission to a projector |
AU2003295318A1 (en) * | 2002-06-14 | 2004-04-19 | Honda Giken Kogyo Kabushiki Kaisha | Pedestrian detection and tracking with night vision |
JP3941650B2 (ja) * | 2002-09-30 | 2007-07-04 | 横河電機株式会社 | 監視システム |
EP1491857A1 (en) * | 2003-06-26 | 2004-12-29 | Harman Becker Automotive Systems GmbH | Method for assisting navigation and navigation system |
JP2005212709A (ja) * | 2004-02-02 | 2005-08-11 | Yoichi Takebayashi | 車両運転支援システム |
US20070230792A1 (en) * | 2004-04-08 | 2007-10-04 | Mobileye Technologies Ltd. | Pedestrian Detection |
JP4128562B2 (ja) * | 2004-11-30 | 2008-07-30 | 本田技研工業株式会社 | 車両周辺監視装置 |
EP1696669B1 (en) * | 2005-02-24 | 2013-07-03 | Aisin Seiki Kabushiki Kaisha | Vehicle surrounding monitoring device |
DE102005041535A1 (de) * | 2005-08-31 | 2007-03-01 | Siemens Ag | Night-Vision-System, bei dem die Umgebung mittels einer Aufnahmevorrichtung aufgenommen und mittels einer Anzeige zumindest teilweise dargestellt wird |
SE529304C2 (sv) * | 2005-09-06 | 2007-06-26 | Gm Global Tech Operations Inc | Metod och system för förbättrande av trafiksäkerhet |
JP2007148835A (ja) * | 2005-11-28 | 2007-06-14 | Fujitsu Ten Ltd | 物体判別装置、報知制御装置、物体判別方法および物体判別プログラム |
GB2436916B (en) * | 2006-03-29 | 2008-09-24 | John Turnbull | Warning System |
JP4830594B2 (ja) * | 2006-04-05 | 2011-12-07 | トヨタ自動車株式会社 | 物体検出装置 |
US20080008355A1 (en) * | 2006-07-05 | 2008-01-10 | Fujitsu Ten Limited | Image recognition apparatus, image recognition method, and vehicle control apparatus |
US8085140B2 (en) * | 2006-09-04 | 2011-12-27 | Panasonic Corporation | Travel information providing device |
JP2008152438A (ja) * | 2006-12-15 | 2008-07-03 | Mitsubishi Fuso Truck & Bus Corp | 車両用表示装置 |
WO2008087706A1 (ja) * | 2007-01-16 | 2008-07-24 | Pioneer Corporation | 車両用表示装置、車両用表示方法及び車両用表示プログラム |
US8462988B2 (en) * | 2007-01-23 | 2013-06-11 | Valeo Schalter Und Sensoren Gmbh | Method and system for universal lane boundary detection |
DE102008038731A1 (de) * | 2008-08-12 | 2010-02-18 | Continental Automotive Gmbh | Verfahren zur Erkennung ausgedehnter statischer Objekte |
US8395529B2 (en) * | 2009-04-02 | 2013-03-12 | GM Global Technology Operations LLC | Traffic infrastructure indicator on head-up display |
US8547298B2 (en) * | 2009-04-02 | 2013-10-01 | GM Global Technology Operations LLC | Continuation of exterior view on interior pillars and surfaces |
US8358224B2 (en) * | 2009-04-02 | 2013-01-22 | GM Global Technology Operations LLC | Point of interest location marking on full windshield head-up display |
US8817090B2 (en) * | 2009-04-02 | 2014-08-26 | GM Global Technology Operations LLC | Luminance uniformity compensation of vector projection display |
US8269652B2 (en) * | 2009-04-02 | 2012-09-18 | GM Global Technology Operations LLC | Vehicle-to-vehicle communicator on full-windshield head-up display |
US8427395B2 (en) * | 2009-04-02 | 2013-04-23 | GM Global Technology Operations LLC | Full-windshield hud enhancement: pixelated field of view limited architecture |
US8912978B2 (en) * | 2009-04-02 | 2014-12-16 | GM Global Technology Operations LLC | Dynamic vehicle system information on full windshield head-up display |
US8629903B2 (en) * | 2009-04-02 | 2014-01-14 | GM Global Technology Operations LLC | Enhanced vision system full-windshield HUD |
US8629784B2 (en) * | 2009-04-02 | 2014-01-14 | GM Global Technology Operations LLC | Peripheral salient feature enhancement on full-windshield head-up display |
US8564502B2 (en) * | 2009-04-02 | 2013-10-22 | GM Global Technology Operations LLC | Distortion and perspective correction of vector projection display |
US8482486B2 (en) * | 2009-04-02 | 2013-07-09 | GM Global Technology Operations LLC | Rear view mirror on full-windshield head-up display |
US8350724B2 (en) * | 2009-04-02 | 2013-01-08 | GM Global Technology Operations LLC | Rear parking assist on full rear-window head-up display |
US20100253595A1 (en) * | 2009-04-02 | 2010-10-07 | Gm Global Technology Operations, Inc. | Virtual controls and displays by laser projection |
US8977489B2 (en) * | 2009-05-18 | 2015-03-10 | GM Global Technology Operations LLC | Turn by turn graphical navigation on full windshield head-up display |
US8164543B2 (en) * | 2009-05-18 | 2012-04-24 | GM Global Technology Operations LLC | Night vision on full windshield head-up display |
WO2011043006A1 (ja) * | 2009-10-07 | 2011-04-14 | パナソニック株式会社 | 制御装置および車両周囲監視装置 |
EP2491344B1 (en) * | 2009-10-22 | 2016-11-30 | TomTom Global Content B.V. | System and method for vehicle navigation using lateral offsets |
JP5441626B2 (ja) * | 2009-11-06 | 2014-03-12 | 日立オートモティブシステムズ株式会社 | 車載用マルチアプリ実行装置 |
JP2011118482A (ja) * | 2009-11-30 | 2011-06-16 | Fujitsu Ten Ltd | 車載装置および認知支援システム |
US8232872B2 (en) * | 2009-12-03 | 2012-07-31 | GM Global Technology Operations LLC | Cross traffic collision alert system |
US8537030B2 (en) * | 2010-02-15 | 2013-09-17 | Ford Global Technologies, Llc | Pedestrian alert system and method |
JP5382198B2 (ja) * | 2010-02-22 | 2014-01-08 | トヨタ自動車株式会社 | 危険度算出装置 |
WO2011108218A1 (ja) * | 2010-03-01 | 2011-09-09 | 本田技研工業株式会社 | 車両の周辺監視装置 |
US9211837B2 (en) * | 2010-03-01 | 2015-12-15 | Concorde Invest Ab | Pedestrian warning device for a vehicle |
DE102010010912A1 (de) * | 2010-03-10 | 2010-12-02 | Daimler Ag | Fahrerassistenzvorrichtung mit optischer Darstellung erfasster Objekte |
JP5218458B2 (ja) * | 2010-03-23 | 2013-06-26 | 株式会社デンソー | 車両接近報知システム |
WO2011132388A1 (ja) * | 2010-04-19 | 2011-10-27 | 本田技研工業株式会社 | 車両の周辺監視装置 |
WO2011148417A1 (ja) * | 2010-05-26 | 2011-12-01 | 三菱電機株式会社 | 車外への音発生装置 |
DE102010040803A1 (de) * | 2010-09-15 | 2012-03-15 | Continental Teves Ag & Co. Ohg | Visuelles Fahrerinformations- und Warnsystem für einen Fahrer eines Kraftfahrzeugs |
JP5554261B2 (ja) * | 2011-02-24 | 2014-07-23 | アルパイン株式会社 | 立体物検出装置および立体物検出方法 |
US9233659B2 (en) * | 2011-04-27 | 2016-01-12 | Mobileye Vision Technologies Ltd. | Pedestrian collision warning system |
WO2012147548A1 (ja) * | 2011-04-28 | 2012-11-01 | 本田技研工業株式会社 | 車両用物体検出装置 |
DE102011076112A1 (de) * | 2011-05-19 | 2012-11-22 | Bayerische Motoren Werke Aktiengesellschaft | Verfahren und Vorrichtung zum Erkennen eines möglichenKollisionsobjektes |
US8994823B2 (en) * | 2011-07-05 | 2015-03-31 | Toyota Jidosha Kabushiki Kaisha | Object detection apparatus and storage medium storing object detection program |
JP5664784B2 (ja) * | 2011-07-21 | 2015-02-04 | トヨタ自動車株式会社 | 車両用情報伝達装置 |
US8611598B2 (en) * | 2011-07-26 | 2013-12-17 | Harman International (China) Holdings Co., Ltd. | Vehicle obstacle detection system |
JP2013029451A (ja) * | 2011-07-29 | 2013-02-07 | Ricoh Co Ltd | 付着物検出装置及び付着物検出方法 |
DE102011081367A1 (de) * | 2011-08-23 | 2013-02-28 | Robert Bosch Gmbh | Verfahren zur Steuerung einer Lichtaussendung eines Scheinwerfers eines Fahrzeugs |
US20130093583A1 (en) * | 2011-10-14 | 2013-04-18 | Alan D. Shapiro | Automotive panel warning and protection system |
DE102011084618A1 (de) * | 2011-10-17 | 2013-04-18 | Robert Bosch Gmbh | Fahrerassistenzsystem |
CN103858156B (zh) * | 2011-10-18 | 2015-04-15 | 本田技研工业株式会社 | 车辆周围监测装置 |
US9146898B2 (en) * | 2011-10-27 | 2015-09-29 | Magna Electronics Inc. | Driver assist system with algorithm switching |
US9235990B2 (en) * | 2011-11-25 | 2016-01-12 | Honda Motor Co., Ltd. | Vehicle periphery monitoring device |
US8824733B2 (en) * | 2012-03-26 | 2014-09-02 | Tk Holdings Inc. | Range-cued object segmentation system and method |
US10089537B2 (en) * | 2012-05-18 | 2018-10-02 | Magna Electronics Inc. | Vehicle vision system with front and rear camera integration |
US8659408B2 (en) * | 2012-05-22 | 2014-02-25 | Delphi Technologies, Inc. | Object detection system and method using a camera and a multiple zone temperature sensor |
US9139133B2 (en) * | 2012-05-31 | 2015-09-22 | GM Global Technology Operations LLC | Vehicle collision warning system and method |
JP5710752B2 (ja) * | 2012-11-13 | 2015-04-30 | 株式会社東芝 | 検出装置、方法及びプログラム |
US9743002B2 (en) * | 2012-11-19 | 2017-08-22 | Magna Electronics Inc. | Vehicle vision system with enhanced display functions |
-
2011
- 2011-02-14 EP EP11750312.8A patent/EP2544161B1/en not_active Not-in-force
- 2011-02-14 US US13/580,864 patent/US9073484B2/en not_active Expired - Fee Related
- 2011-02-14 CN CN201180011493.3A patent/CN102782741B/zh not_active Expired - Fee Related
- 2011-02-14 WO PCT/JP2011/000803 patent/WO2011108198A1/ja active Application Filing
- 2011-02-14 JP JP2012502982A patent/JP5706874B2/ja not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001006096A (ja) | 1999-06-23 | 2001-01-12 | Honda Motor Co Ltd | 車両の周辺監視装置 |
JP2003284057A (ja) | 2002-01-18 | 2003-10-03 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2007159036A (ja) * | 2005-12-08 | 2007-06-21 | Denso Corp | 車両用表示装置 |
JP2007241740A (ja) | 2006-03-09 | 2007-09-20 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2007334751A (ja) | 2006-06-16 | 2007-12-27 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2009098956A (ja) * | 2007-10-17 | 2009-05-07 | Mazda Motor Corp | 車両用運転支援装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2544161A4 |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2013098919A (ja) * | 2011-11-04 | 2013-05-20 | Honda Motor Co Ltd | 車両周辺監視装置 |
JP2015128201A (ja) * | 2013-12-27 | 2015-07-09 | トヨタ自動車株式会社 | 車両用情報表示装置及び車両用情報表示方法 |
JP2018508893A (ja) * | 2015-02-27 | 2018-03-29 | トヨタ モーター エンジニアリング アンド マニュファクチャリング ノース アメリカ,インコーポレイティド | 状況環境情報を提供するためのシステム及び方法 |
JP2017016457A (ja) * | 2015-07-02 | 2017-01-19 | パナソニックIpマネジメント株式会社 | 表示制御装置、投影装置、表示制御プログラム、および記録媒体 |
US10464484B2 (en) | 2015-08-04 | 2019-11-05 | Denso Corporation | Apparatus for presenting support images to a driver and method thereof |
WO2017022496A1 (ja) * | 2015-08-04 | 2017-02-09 | 株式会社デンソー | 運転者に支援画像を提示する装置及びその方法 |
JP2017034543A (ja) * | 2015-08-04 | 2017-02-09 | 株式会社デンソー | 車載表示制御装置、車載表示制御方法 |
JP2017074871A (ja) * | 2015-10-15 | 2017-04-20 | 日立建機株式会社 | 車両周囲障害物検出装置 |
JP2017138817A (ja) * | 2016-02-04 | 2017-08-10 | トヨタ自動車株式会社 | 車両用注意喚起装置 |
JP2018097704A (ja) * | 2016-12-15 | 2018-06-21 | トヨタ自動車株式会社 | 車両用注意喚起装置 |
JP2019089394A (ja) * | 2017-11-13 | 2019-06-13 | 矢崎総業株式会社 | 自動運転時情報伝達方法および車両用情報提示装置 |
JP2020024570A (ja) * | 2018-08-07 | 2020-02-13 | 日立オートモティブシステムズ株式会社 | 運転支援装置 |
JP7189695B2 (ja) | 2018-08-07 | 2022-12-14 | 日立Astemo株式会社 | 運転支援装置 |
WO2024042762A1 (ja) * | 2022-08-23 | 2024-02-29 | マクセル株式会社 | ヘッドアップディスプレイ装置および映像データの処理方法 |
Also Published As
Publication number | Publication date |
---|---|
US20120320212A1 (en) | 2012-12-20 |
EP2544161A4 (en) | 2013-07-24 |
CN102782741A (zh) | 2012-11-14 |
US9073484B2 (en) | 2015-07-07 |
JP5706874B2 (ja) | 2015-04-22 |
CN102782741B (zh) | 2015-08-19 |
JPWO2011108198A1 (ja) | 2013-06-20 |
EP2544161A1 (en) | 2013-01-09 |
EP2544161B1 (en) | 2014-12-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5706874B2 (ja) | 車両の周辺監視装置 | |
JP5503728B2 (ja) | 車両の周辺監視装置 | |
JP5577398B2 (ja) | 車両の周辺監視装置 | |
JP4410292B1 (ja) | 車両の周辺監視装置 | |
JP4456086B2 (ja) | 車両周辺監視装置 | |
EP2759999B1 (en) | Apparatus for monitoring surroundings of vehicle | |
JP4173902B2 (ja) | 車両周辺監視装置 | |
JP4528283B2 (ja) | 車両周辺監視装置 | |
WO2011132388A1 (ja) | 車両の周辺監視装置 | |
JP2003134508A (ja) | 車両用情報提供装置 | |
JP2008027309A (ja) | 衝突判定システム、及び衝突判定方法 | |
JP4988781B2 (ja) | 車両の周辺監視装置 | |
JP2007334511A (ja) | 対象物検出装置、車両、対象物検出方法、並びに対象物検出用プログラム | |
JP5192007B2 (ja) | 車両の周辺監視装置 | |
JP5292047B2 (ja) | 車両の周辺監視装置 | |
JP5192009B2 (ja) | 車両の周辺監視装置 | |
JP4176558B2 (ja) | 車両周辺表示装置 | |
JP4823753B2 (ja) | 車両の周辺監視装置 | |
JP2010191666A (ja) | 車両の周辺監視装置 | |
JP4943403B2 (ja) | 車両の周辺監視装置 | |
JP5430633B2 (ja) | 車両周辺監視装置 | |
JP2007311866A (ja) | 対象物検出装置、車両、対象物検出方法、並びに対象物検出用プログラム | |
JP2006185431A (ja) | 車両周辺監視装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180011493.3 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11750312 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13580864 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012502982 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011750312 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |