US7141796B2 - Vehicle information providing apparatus - Google Patents

Vehicle information providing apparatus Download PDF

Info

Publication number
US7141796B2
US7141796B2 US10/273,314 US27331402A US7141796B2 US 7141796 B2 US7141796 B2 US 7141796B2 US 27331402 A US27331402 A US 27331402A US 7141796 B2 US7141796 B2 US 7141796B2
Authority
US
United States
Prior art keywords
zone
search
image
brightness
area
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US10/273,314
Other versions
US20030083790A1 (en
Inventor
Hiroshi Hattori
Nobuharu Nagaoka
Masahito Watanabe
Takayuki Tsuji
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Arriver Software AB
Original Assignee
Honda Motor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honda Motor Co Ltd filed Critical Honda Motor Co Ltd
Assigned to HONDA GIKEN KOGYO KABUSHIKI KAISHA reassignment HONDA GIKEN KOGYO KABUSHIKI KAISHA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HATTORI, HIROSHI, NAGAOKA, NOBUHARU, TSUJI, TAKAYUKI, WATANABE, MASAHITO
Publication of US20030083790A1 publication Critical patent/US20030083790A1/en
Application granted granted Critical
Publication of US7141796B2 publication Critical patent/US7141796B2/en
Assigned to VEONEER SWEDEN AB reassignment VEONEER SWEDEN AB ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HONDA MOTOR CO., LTD
Assigned to ARRIVER SOFTWARE AB reassignment ARRIVER SOFTWARE AB ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VEONEER SWEDEN AB
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/182Network patterns, e.g. roads or rivers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads

Definitions

  • the present invention relates to a vehicle information providing apparatus that detects and displays, for example, a physical object present in the vicinity of a vehicle.
  • an apparatus in which an image from one or a plurality of infrared cameras mounted on the front of the vehicle is displayed at a position visible from the driver's seat to complement the front visual field of the driver.
  • the image displayed to the driver is displayed on an image display apparatus such as a NAVIDisplay disposed in the dashboard of the vehicle, a HUD (Head Up Display) that displays information at a position on the front window that does not obscure the front visual field of the driver, a meter integrated display that is integrated with a meter that displays the travel state of the automobile numerically, or the like.
  • Japanese Unexamined Patent Application, First Publication, Hei 11-328364 discloses an apparatus that records the environment in the vicinity of the vehicle and displays the result to the driver in this manner.
  • the temperature of the exposed skin of the head of the detected object is higher than the temperature of the other parts, and first the position of the head of the detected object is identified using the fact that it is comparatively easy to record as an image of an infrared camera.
  • the zone corresponding to the body of the detected object is determined based on the information of the identified position of the head. Thereby, for example, caution can be prompted by displaying to the driver not only the head of the pedestrian, but also the entire body.
  • the detected object in the case that the detected object is recorded from the front, the amount of infrared radiation of the skin of the face is large in comparison to other parts, and thus the position of the head of the detected object is easy to detect.
  • detecting the zone of the entire body therefrom is easy, in such a case as the detected object being recorded from the back, the parts other than the face are extracted, and it is difficult to identify the position of the head.
  • the temperature of parts of the body of the detected object change depending on the conditions and the environment. Examples are parts whose temperature rises due to direct exposure to the sun and parts whose temperature is lowered due to being blown on by the wind. Consequently, due to the increase or decrease in temperature, it is not always the case that the position of the head can be identified from the recorded image.
  • the zone that corresponds to the body identified from the part recognized as the position of the head may be different from the zone in which the detected object is actually present.
  • a first aspect of the vehicle information providing apparatus of the present invention which displays an image recorded by an infrared camera, comprises an extracted zone setting device (for example, step S 1 to step S 18 in the embodiments) that sets a present zone of the physical object detected by multivalued processing of the image as the extracted zone (for example, the first extracted zone 51 ), a search area setting device (for example, step S 21 to step S 24 in the embodiments) that sets the search area (for example, search areas 54 to 58 in the embodiments) in the vicinity of the extracted zone, and a physical object recognition device (for example, step S 25 to step S 37 in the embodiments) that searches for changes in brightness in the searched area, and provides a highlighted display of the area where there is a change in brightness along with the extracted zone as the same object.
  • an extracted zone setting device for example, step S 1 to step S 18 in the embodiments
  • a search area setting device for example, step S 21 to step S 24 in the embodiments
  • search area setting device for example, step S 21 to step S 24 in
  • the vehicle information providing apparatus having the structure described above sets the search area by setting as the reference the extracted zone set by the extracted zone setting device and searching the zone exhibiting a change in brightness in the vicinity of the extracted zone. Thereby, the zone in which a physical object identical to the object recorded in the extracted zone can be provided in a highlighted display along with the extracted zone.
  • a second aspect of the vehicle information providing apparatus of the present invention comprises two infrared cameras and a parallax calculating device (for example, step S 34 in the embodiments) that finds the parallax of the images recorded by the two cameras, and the physical object recognizing device gives a highlighted display of the zone having a change in brightness and parallax identical to the extracted zone as an identical physical body.
  • a parallax calculating device for example, step S 34 in the embodiments
  • the vehicle information providing apparatus having the structure described above recognizes a zone having a change in brightness and parallax identical to the extracted area as the area in which a physical body identical to the physical body that was taken in the extracted area, and this can be given a highlighted display along with the extracted area.
  • the physical object recognizing device stops searching for changes in brightness in the case that the adjacent search area (for example, search areas 52 and 53 in the embodiments) set above or below the extracted zone exceeds the range of the image.
  • the adjacent search area for example, search areas 52 and 53 in the embodiments
  • the physical object recognizing device gives a highlighted display of only an extracted zone without a search of the physical object by determining that the large part of the physical object in the extracted zone is recorded in the case that the zone of the image is exceeded when the search area setting device sets the search area above or below the extracted zone.
  • FIG. 1 is a block diagram showing the structure of the vehicle information providing apparatus according to an embodiment of the present invention.
  • FIG. 2 is a drawing showing the installation positions of the infrared cameras, sensors, display, and the like on the vehicle.
  • FIG. 3 is a flowchart showing the entire operation of the vehicle information providing apparatus according to the same embodiment.
  • FIGS. 4A and 4B show the gray scale image taken by the infrared camera and the binary image thereof.
  • FIGS. 5A , 5 B, and 5 C are drawings showing the conversion processing to a run length data and labeling.
  • FIGS. 6A and 6B are drawings showing the time tracking of the object.
  • FIG. 7 is a drawing showing the rotation angle correction of the object image.
  • FIGS. 8A and 8B are drawings showing the search area in the right image and the search area set in the left image.
  • FIG. 9 is a drawing showing the correlative operation processing in which the search area is set as the object.
  • FIGS. 10A and 10B are drawings showing the method of calculating the parallax between objects in the distance calculation of the object.
  • FIGS. 11A and 11B are drawings showing the offset of the position of the object in the image generated by the turning of the car.
  • FIG. 12 is a drawing showing the case in which a collision may easily occur.
  • FIG. 13 is a drawing showing the zone division in front of the vehicle.
  • FIG. 14 is a flowchart showing the highlighted image output processing operation according to the first embodiment of the same example.
  • FIGS. 15A and 15B are drawings showing an example of a zone division result in the image of the same example.
  • FIGS. 16A and 16B is a drawing showing the zone searched sequence in the image according to a first embodiment of the same example.
  • FIGS. 17A and 17B are drawings showing the zone searched sequence in the image according to a first embodiment of the same example.
  • FIG. 18 is a flowchart showing the brightness change search processing operation of the vehicle information providing apparatus of the same example.
  • FIGS. 19A and 19B are drawing showings the zone search sequence in an image according to a second and third embodiment of the example.
  • FIG. 20 is a drawing showing the zone search sequence in an image according to a fourth embodiment of the example.
  • FIGS. 21A and 21B are drawings showing an example of the zone division results in an image of the same example.
  • FIG. 22 is a drawing showing an example of the zone division results in an image of the same example.
  • FIG. 1 is a block diagram showing the structure of the vehicle information providing apparatus according to an embodiment of the present invention.
  • reference numeral 1 is an image processing unit providing a CPU (central control unit) that controls the vehicle information providing apparatus according to the present embodiment, and wherein two infrared cameras 2 R and 2 L that can detect infrared radiation, a yaw rate sensor 3 that detects the yaw rate of the vehicle, a vehicle speed sensor 4 that detects the travel speed (vehicle speed) of the vehicle, and brake sensors 5 that detect the operation of the brakes are connected together.
  • CPU central control unit
  • the image processing unit 1 detects a moving physical object such as a pedestrian or animal in front of the vehicle from an infrared image of the vicinity of the vehicle and a signal indicating the travel state of the vehicle, and issues a warning when it determines that the possibility of a collision is high.
  • a speaker 6 for issuing a warning by voice and an image display apparatus 7 including, for example, a meter integrated Display having a meter integrated for showing numerically the travel condition of the vehicle, a NAVIDisplay mounted in the dashboard of the vehicle, or a HUD (head up display) 7 a that displays information at a position on the front window at a position that does not obscure the front visual field of the driver, are connected.
  • the image display apparatus 7 is for displaying the image recorded by the infrared cameras 2 R and 2 L, and informing the driver about objects with which there is a high danger of collision.
  • the image processing unit 1 comprises an A/D conversion circuit for converting input analog signals into digital signals, image memory for storing the digitalized image signal, a CPU (central processing unit) that carries out each of the operation processes, RAM (random access memory) that the CPU uses to store data that is being processed, ROM (read only memory) for storing programs executed by the CPU, tables, maps, or the like, a drive signal for the speaker 6 , and an output circuit that outputs display signals or the like from the HUD 7 a or the like.
  • Each of the output signals of the infrared cameras 2 R and 2 L, the yaw rate sensor 3 , the vehicle speed sensor 4 , and the brake sensors 5 are formed so as to be converted to digital signals and then input into the CPU.
  • the infrared cameras 2 R and 2 L are mounted at a positions substantially symmetrical with respect to the center in the transverse direction of the vehicle 10 , the optical axis of the two infrared cameras 2 R and 2 L are mutually parallel, and their heights above the surface of the road are fixed so as to be equal. Moreover, the infrared cameras 2 R and 2 L have the characteristic that the output signal level becomes higher (the brightness increases) the higher the temperature of the object.
  • the HUD 7 a is mounted so that the display image is displayed at a position on the window of the vehicle 10 that does not obscure the visual field of the driver.
  • FIG. 3 is a flowchart showing the processing sequence in the image processing unit 1 of the vehicle information providing apparatus according to the same embodiment.
  • the image processing unit 1 records the infrared image, which is the output signal of the infrared cameras 2 R and 2 L (step S 1 ), carries out A/D conversion thereon (step S 2 ), and stores the gray scale image in the image memory (step S 3 ). Moreover, here the right image is obtained by the infrared camera 2 R and the left image is obtained by the infrared camera 2 L. In addition, because the horizontal position of an identical object on the display screen is displayed out of alignment in the right image and the left image, the distance to the object can be calculated using this misalignment (parallax).
  • the right image obtained from the infrared camera 2 R is made the reference image, and binary processing of this image signal is carried out.
  • a 1 (white) is assigned to a zone brighter than a brightness threshold value ITH, and a 0 (black) is assigned to a dark zone.
  • FIG. 4A shows a gray scale image obtained by the infrared camera 2 r, and by carrying out binary processing thereon, the image shown in FIG. 4B is obtained.
  • the physical object surrounded by the frame from P 1 to P 4 for example, will be an object (below, referred to as a “high brightness zone”) displayed as white on the display screen.
  • processing is carried out in which the binary image data is converted to run length data (step S 5 ).
  • FIG. 5A is a drawing to explain this, and in this figure the zone that has become white due to binary conversion is shown as the lines L 1 to L 8 .
  • Lines L 1 to L 8 all have a width of 1 pixel in the y direction, and while they are actually arranged without a space between them in the y direction, they have been separated for the sake of the explanation.
  • the lines L 1 to L 8 respectively have the lengths 2 pixels, 2 pixels, 3 pixels, 8 pixels, 7 pixels, 8 pixels, 8 pixels, and 8 pixels.
  • the run length data is shown by the coordinates of the start point of each of the lines (the point on the left end of each line) and the length (number of pixels) from the start point to the end point (the point on the right end of each line).
  • line L 3 comprises the 3 pixels (x 3 , y 5 ), (x 4 , y 5 ), and (x 5 , y 5 ), and thus (x 3 , y 5 , 3 ) becomes the run length data.
  • step S 7 the processing in which the object is extracted is carried out (step S 7 ). That is, as shown in FIG. 5B , among the lines L 1 to L 8 that have been converted to run length data, the lines L 1 to L 3 , which are the parts overlapping in the y direction, are treated as one object 1 , lines L 4 to L 8 are treated as one object 2 , and the object labels 1 and 2 are added to the run length data.
  • the high brightness zones shown in FIG. 4B are respectively recognized as objects 1 through 4 .
  • step S 8 the aspect ratio ASPECT of the circumscribed quadrangle represented by the broken lines is calculated.
  • the surface zone S is calculated by adding the lengths of the run length data for the same object.
  • the coordinate of the center of gravity G is calculated as the x coordinate of the line that bisects the surface zone S in the x direction, and the y coordinate of the line that bisects it in the y direction.
  • the aspect ratio ASPECT is calculated as the Dy/Dx ratio of Dy and Dx shown in FIG. 5C .
  • the position of the center of gravity G can be substituted for by the position of the center of gravity of the circumscribed quadrangle.
  • step S 9 next recognition of the time trace, that is, the sampling of each cycle, of the same object is carried out (step S 9 ).
  • k serves as the time during which time t, an analogue amount, is made discrete at a sampling cycle, and as shown in FIG. 6A , in the case that objects A and B are extracted at time k, obj cots C and D extracted at time (k+1) are determined to be identical to obj cots A and B.
  • objects A and B are determined to be identical to objects C and D, and objects C and D have their labels changed respectively to objects A and B.
  • the object A and the object C satisfy the conditions for the identification of identity described above and the object B and the object D satisfy the conditions for the identification of identity described above, and thus the objects C and D are respectively recognized to be the objects A and B.
  • the positions coordinates (of the center of gravity) of each of the recognized objects is stored in the memory as time series position data to be used on later calculation processing.
  • steps S 4 to S 9 explained above is carried out on a binary reference image (in this embodiment, the right image).
  • the velocity VCAR detected by the velocity sensor 4 and the yaw rate YR detected by the yaw rate sensor 3 are read, and as shown in FIG. 7 , the turning angle ⁇ r of the vehicle 10 is calculated by integrating the yaw rate YR with respect to time (step S 10 ).
  • step S 9 and step S 10 are carried out in parallel, and in steps S 11 through S 13 , processing that calculates the distance z between the object and the vehicle 10 is carried out. Because this calculation requires a longer time than step S 9 and step S 10 , they are executed at a longer cycle than step S 9 and S 10 (for example, at a cycle about three times the execution cycle of steps S 1 to S 10 ).
  • the search image R 1 (here, the entire zone surrounded by the circumscribed quadrangle is made the searched image) is extracted from the right image (step S 11 ).
  • the vehicle search area in which the image corresponding to the searched image (below, referred to as the “corresponding image”) from the left image is set, and the corresponding image is extracted by executing the correlation calculation (step S 12 ).
  • the search region R 2 in the left image is set, and the brightness difference total value C (a, b), which indicates the degree of the correlation with the searched image R 1 in the search region R 2 , is calculated by the Eq. 1 shown below, and the zone in which this total value C (a, b) becomes minimum is extracted as the corresponding image.
  • this correlation calculation is carried out using the gray scale image, not the binary image.
  • a zone R 2 a (shown by the broken line in FIG. 8B ) that is narrower than the search region R 2 is set to serve as the search area.
  • IR(m, n) is the brightness value of the position of the coordinate (m, n) in the search area R 1 shown in FIG. 9
  • IL(a+m ⁇ M, b+n ⁇ N) is the brightness value of the position of the coordinate (m, n) in the search one R 1 and the local zone R 3 having the same shape, where the coordinates (a, b) in the search area are the base points.
  • the position of the corresponding image is defined by finding the position at which the total value C (a, b) of the brightness difference is minimized by changing the coordinates (a, b) of the base point.
  • step S 12 Due to the processing in step S 12 , as shown in FIG. 10A and FIG. 10B , because the searched image R 1 and the corresponding image R 4 corresponding to this object are extracted, next the distance dR (number of pixels) between the position of the center of gravity of the searched image R 1 and the image center line LCTR and the distance dL (number of pixels) between the position of the center of gravity of the corresponding image R 4 and the image center line LCTR are found, and by applying the following Eq. 2, the distance z between the vehicle 10 and the object is calculated (step S 13 ).
  • dR number of pixels
  • dL number of pixels
  • B is the base line length, that is, the distance in the horizontal direction between the center position of the photographic element of the infrared camera 2 R and the center position of the photographic element of the infrared camera 2 L (the separation of the light beam axis of both infrared cameras);
  • F is the focal distance of the lenses of the infrared cameras 2 R and 2 L,
  • p is the pixel separation in the photographic element of the infrared cameras 2 R and 2 L, and
  • step S 10 When the calculation of the turning angle ⁇ r in step S 10 and the calculation of the distance to the object in step S 13 have completed, the coordinates (x, y) in the image and the distance z calculated by Eq. 2 are applied to the following Eq. 3, and converted to real spatial coordinates (X, Y, Z) (step S 14 ).
  • the real spatial coordinates (X, Y, Z) have as their origin O the position of the center point of the installation position of the infrared cameras 2 R and 2 L (the position at which they are fixed on the vehicle 10 ), they are fixed as shown in the figure, and the coordinates in the image are determined by x in the horizontal direction and y in the vertical direction, where the center of the image serves as the origin.
  • (xc, yc) are the coordinates (x, y) of the right image that have been converted to coordinates of a virtual image in which the real spatial origin O and the center of the image have been made to coincide based on the relative positional relationship between the installation position of the infrared camera 2 R and the real special origin O.
  • f is the ratio of the focus distance F and the pixel interval p.
  • turning angle compensation is carried out in order to compensate the positional shift in the image due to the turning of the vehicle 10 (step S 15 ).
  • the turning angle compensation is a process to compensate this.
  • the real spatial coordinate system (X, Y, Z) is applied, and the compensated coordinates (Xr, Yr, Zr) are calculated.
  • the calculated real spatial position data (Xr, Yr, Zr) is associated with each object and stored in memory.
  • the coordinates after turning angle compensation are denoted (X, Y, Z).
  • u is a parameter that takes an arbitrary value
  • Xav, Yav, and Zav are respectively the average values of the X coordinate, Y coordinate, and Z coordinate of the real spatial position data sequence.
  • the numerical value in the parenthesis added to P which denotes the coordinates of each of the data point, indicates that the larger the value, the older the data.
  • P( 0 ) denotes the most recent position coordinate
  • P( 1 ) denotes the position coordinate of one sample cycle back
  • P( 2 ) denotes the position coordinate two sample cycles back.
  • the Z coordinates Z( 0 ), Z(N ⁇ 1) to the Eq. 5a above, that is, the following Eq.
  • the relative motion vector is found as the vector from the position coordinate Pv(N ⁇ 1) calculated in Eq. 8 towards Pv( 0 ).
  • step S 16 when the relative motion vector has been found, next the possibility of a collision with the detected object is determined, and a warning determination process, which issues a warning when the possibility is high, is executed (step S 17 ).
  • the warning determining processing is processing that determines the possibility of a collision between the vehicle 10 and the detected object by using any one of the collision detecting processing, processing to determine whether or not the object is in an approach determination zone, or invasive collision determination processing, which are described below. Below, as shown in FIG. 12 , the explanation will treat an example wherein there is an animal 20 being approached at a speed Vp from a substantially 90° with respect to the direction of forward motion of the vehicle 10 .
  • the image processing unit 1 calculates the relative velocity Vs in the Z direction using the following Eq. 7 from the animal 20 approaching the distance Zv( 0 ) from the distance Zv(N ⁇ 1) during the time ⁇ T, and carries out collision determination processing.
  • the collision determination processing is processing that determines whether there is a possibility of a collision when the following Equations 8 and 9 are satisfied.
  • Vs ( Zv ( N ⁇ a ) ⁇ Zv (0))/ ⁇ T Eq.7 Zv (0)/ Vs ⁇ T Eq. 8
  • Zv( 0 ) is the most recent distance detection value (v is attached in order to indicate that this is data after compensation using the approximately straight line LMV, while the Z coordinate is a value identical to that before compensation), and Zv(N ⁇ 1) is the detected distance value before the time ⁇ T.
  • T is an allowable time and signifies that the possibility of a collision is determined time T before the predicted collision time, and is about 2 to 5 seconds, for example.
  • H is a predetermined height that defines the range of the Y direction, that is the height direction, and is set, for example, to about twice the height of the vehicle 10 .
  • AR 1 is the zone corresponding to the range having added the allowance ⁇ (for example, about 50 to 100 cm) to both sides of the width ⁇ of the vehicle 10 , or in other words, the zone having a width ( ⁇ /2+ ⁇ ) on both sides of the axle at the center part in the width direction of vehicle 10 , and if the object continues to be present as-is, the possibility of a collision is extremely high.
  • these zones are called approach determination zones.
  • the zones AR 2 and AR 3 are zones (in the outside transverse direction of the approach determination zone) in which the absolute value of the X coordinate is larger than the approach determination zone, an invasive collision determination, described below, is made about the object inside this zone, and thus this is called the invasive determination zone.
  • these zones have a predetermined height H in the Y direction, as shown in the above Eq. 9.
  • the invasive collision determination processing in step S 23 distinguishes whether or not the difference between xc( 0 ), which is the most recent x coordinate on the image (the character c, as will be explained below, is attached in order to signify that it is a coordinate on which compensation has been carried out that makes the center position of the image align with the real spatial origin point O) and xc(N ⁇ 1), which is the x coordinate before the time ⁇ T, satisfies the following Eq. 10, and in the case that it is satisfied, it is determined that the possibility of a collision is high.
  • step S 17 in any of the collision determining processing, the processing for determining whether there is an object in the approach determining zone, or the invasive collision determining processing, in the case that it is determined that there is no possibility of collision between the vehicle 10 and the detected object (NO in step S 17 ), the processing returns to step S 1 , and the processing described above repeats.
  • step S 17 in the case that it is determined that there is a possibility of a collision between the vehicle 10 and the detected object by any of the collision determining processing, the processing to determined whether there is an object in the approach determining zone, or the invasive collision determining processing (YES in step S 17 ), the processing proceeds to the warning output determining processing of step S 18 .
  • step S 18 the warning output determining processing, that is, determining whether or not to carry out the warning output, is carried out (step S 18 ).
  • the warning output determination process determines whether or not the driver of the vehicle 10 is carrying out a braking action from the output BR of the brake sensor 5 .
  • step S 18 the processing returns to step S 1 , and the processing described above is repeated.
  • step S 19 a warning message is issued via speaker 3 (step S 19 ), and at the same time, a gray scale image obtained, for example, from the infrared camera 2 R is displayed on the image display apparatus 7 , and the highlighted display frame is set on the object being approached. Thereby, the object is displayed to the driver of the vehicle 10 as a highlighted image (step S 20 ).
  • the predetermined threshold value GTH is determined by the following Eq. 11. This is the value corresponding to the condition in which the vehicle 10 stops at a running distance equal to or less than the distance Zv( 0 ) in the case that the acceleration Gs during the braking action is maintained as-is.
  • step S 20 shown in the flowchart in FIG. 3 will be explained with reference to the flowcharts shown in FIG. 14 and FIG. 18 and the drawings shown in FIG. 15 and FIG. 17 .
  • FIG. 14 is a flowchart showing the entire operation of the highlighted image output processing operation according to the first embodiment.
  • the size of the area (below, referred to as the “first detected area”) of the object detected by binarization and correlation operation is set as the reference, a plurality of areas are established by partitioning the zone of the first detected area in the vertical direction such that their size is identical to that of the first detected area, and this serves as the search area (step S 21 ).
  • step S 22 it is determined whether or not the partitioned search area is in the zone of the reference image (fight image) (step S 22 ), the processing returns to step S 21 until the search area protrudes from the reference image, and then repeats the area partition of the zone (establishing the search area) (YES in step S 22 ).
  • step S 22 when the search area protrudes from the reference image (NO in step S 22 ), the number of areas partitioned in the reference image is calculated (step S 23 ).
  • step S 24 it is determined whether or not the number of areas in the reference image zone is greater than one.
  • step S 24 in the case that the number of areas in the reference image zone is one (the case in which the number of areas is one in the first detected area in the reference image zone and the number of areas is 2 in the search area protruding vertically from the reference image zone, making a total of three) (NO in step S 24 ), the first detected area is set as the highlighted display area (step S 25 ).
  • FIGS. 15A and 15B are drawings showing an example of an image in the case that the object 30 is a human and the number of areas in the reference image zone 50 is one.
  • FIG. 15A is the case in which the entire body of the object 30 has been taken in the first detected area 51 , and whether this is the entire body can be determined by the area size and distance of the object 30 .
  • 15 B is the case in which a part of the body of the object 30 has been taken in the first detected area 51 .
  • the first detected area 51 in the reference image zone 50 and the search areas 52 and 53 that vertically protrude from the reference image zone 50 are established, making a total of three areas.
  • the search area in the row designated by the register j is determined in the longitudinal direction for each search area, and the brightness change search processing that sets the flag in the search area in which a change in brightness in the image is recognized is carried out (step S 27 ). Moreover, the details of the brightness change search process will be described below.
  • FIG. 16A shows an image row in the case that the object 30 is a person and the number of areas in the reference image zone 50 is greater than one.
  • a row is formed comprising the first detected area 51 , the search areas 54 and 55 that are present in the vertical direction thereto and protrude from the reference image zone 50 , and further, a plurality of search areas 56 present in the reference image zone 50 , for a total of M areas.
  • step S 28 it is determined whether or not there is a flag indicating that a change in brightness was recognized in the image in the search area of the row designated by the register j (step S 28 ).
  • step S 32 brightness change search processing is carried out.
  • step S 33 it is determined whether or not there is a flag indicating that a change in brightness has been recognized in the image in the search area in the row designated by the register j (step S 33 ).
  • step S 33 in the case that there is no flag indicating that a change in brightness has been recognized in the image in the search area in the designated row (NO in step S 33 ), it is determined that the search in the reference image zone 50 has completed, and next the parallax calculation is carried out on the search area for which a flag has been set (step S 34 ).
  • step S 35 the flag of the search area having a parallax that differs from the first search area 51 (a differing distance from the vehicle 10 ) is cleared.
  • highlighted display is carried out such that the first search area 51 and the search area having a set flag are included (step S 36 ).
  • the highlighted display output set in step S 25 and the highlighted display output set in step S 36 are output to the image display apparatus 7 (step S 37 ), and the highlighted display output processing is completed.
  • FIG. 17A shows the state in which the highlighted display frame 59 that includes the first detected area 51 and the search area that includes a search area recognized to have a change in brightness is set.
  • FIG. 17B when output to the image display apparatus 7 , the highlighted display frame 59 that includes the entire object, which is easily recognized as having an object shape, is set on the displayed gray scale image.
  • the brightness change search processing is processing in which it is determined whether or not a change in brightness in the image in the search area can be recognized in each search area in the row designated by the register j, and sets a flag in a search area where a change in brightness is recognized in the image.
  • the search area in the row is searched in the direction of the row (longitudinal direction) with respect to the designated row.
  • a search is carried out to determine whether or not a change in brightness in the gray scale image of the search area is recognized (step S 42 ).
  • step S 43 it is determined whether or not the change in brightness in the gray scale image of the designated area is large (step S 43 ), and in the case that the change in brightness is large (YES in step S 43 ), the flag is set for the corresponding search area (step S 44 ).
  • step S 43 the processing returns to step S 45 without taking any action.
  • step S 46 because the total number in each row of the first detected area 51 and the search area is M areas, the value of the register i is determined (step S 46 ), and the brightness change search processing returns to step S 42 until the value of i is larger than M, and the processing described above repeats (NO in step S 46 ).
  • step S 46 when the value of i becomes large (i>M) (YES in step S 46 ), the brightness change search processing ends.
  • FIG. 19A is a drawing for explaining the second embodiment of the highlighted image output processing.
  • the first embodiment enlarges the search area in the row direction
  • enlargement is carried out in the row direction by making all search areas included vertically in M areas serve as objects.
  • a search is carried out by enlarging only the area of a row where a search area having a set flag is present.
  • the brightness change processing is carried out as follows:
  • FIG. 19B is a drawing for explaining the third embodiment of the highlighted image output processing.
  • the area made by extending the search area in the horizontal direction is set using the size of the first detected area 51 as the reference, and in the case that the search area is expanded in the row direction, in the row that is the same as the first detected area 51 , the column in which a change in brightness has been recognized is the object.
  • the search area is expanded in the row direction, in the row that is the same as the first detected area 51 , the column in which a change in brightness has been recognized is the object.
  • FIG. 20 is a drawing for explaining the fourth embodiment of the highlighted image output processing.
  • the reference image zone 50 is partitioned by a search area having a predetermined size, and the brightness change search processing is carried out.
  • the brightness change search processing is executed using an area somewhat larger than the first detected area 51 .
  • the brightness change search processing is carried out in the reference image zone 50 irrespective of the size of the first detected area 51 , and thus although the processing load on the image processing unit 1 can be largely reduced, in the case that the difference in the size of the search area and the size of the object 30 is great, it is necessary to consider the case that this may entail a decrease in the detection precision and a decrease in the detection speed.
  • the parallax calculation between the images of the two infrared cameras 2 R and 2 L can be carried out first, and then only the portion of the search area having identical parallax needs to be expanded.
  • the objects can be distinguished and extracted.
  • the search area has a parallax value that differs from the first detected area 51 , and a plurality of search area having such parallax values are present.
  • the portion excluding the hidden area can be given a highlighted display.
  • FIG. 21A is an example of the partition of the reference image using medium sized search areas that have been set based on the first detected area 51 , which has detected the torso of the object 30 .
  • FIG. 21B is an example of the partition of a reference image using small sized search areas that have been set based on the first search area 51 , which has detected the end of the hand of the object 30 .
  • FIG. 22 is an example of the partition of the reference image using a transversely large sized search area set based on the first detected area 51 , which has detected both shoulders and the chest portion of the object 30 .
  • the transverse size of the search area can be adjusted by the distance to the object found from the parallax of the image.
  • monitoring the front of the vehicle were given, but monitoring can be carried out in any direction, such as the back of the vehicle.
  • infrared camera 2 R and 2 L were provided, and the distance to the object found using parallax is also used as determination material for extracting identical physical objects.
  • the image processing unit 1 includes the extracted zone setting device, the search zone setting device, the physical object recognition device, and the parallax calculating device. More concretely, steps S 1 to S 18 in FIG. 3 correspond to the extracted zone setting device, steps S 21 to S 24 in FIG. 14 correspond to the search zone setting device, and steps S 25 to S 37 in FIG. 14 correspond to the physical body recognizing device. Furthermore, step S 34 in FIG. 14 corresponds to the parallax calculating device.
  • the vehicle information providing apparatus of the present embodiments sets the search area in the vicinity of the vehicle using a first detected area 51 , which is the first area to be detected, as the reference, and in the case that a change in brightness is recognized in the search area and two infrared cameras are used, the search area having a parallax identical to the first search area 51 is determined to be the image of a physical object identical to that in the first search area 51 , Thereby, it is possible to extract an image area that takes in the entire object reliably.
  • a search zone is set using the extracted zone set by the extraction zone setting device as a reference, and by searching the zone that has a change in brightness in the vicinity of the extracted zone, a zone in which is it appears that a physical body can be recorded that is identical to the physical body recorded in the extracted zone can be given a highlighted display along with the extracted zone. Therefore, when the search zone is set based on the size of the extracted zone, the size of the search zone can be suitably set with respect to the object by using the size of the extraction zone that is proportional to the size of the object. Thereby, the effect is obtained that the detection speed of detecting the entire physical object by searching in the image zone using the search zone is increased.
  • a zone having a change in brightness and a parallax identical to the extracted zone is recognized as a zone in which it is assumed that the physical object can be taken that is identical to the physical object taken in the extracted zone, and this can be given a highlighted display along with the extracted zone.
  • the physical object recognition device determines that a large part of the physical object has been recorded in the extracted zone, and carries out highlighted display of only the extracted zone without carrying out a search of the physical object.

Abstract

A vehicle information providing apparatus determines and displays the range of presence of a person or other object based on the characteristics of images recorded by infrared cameras. An image processing unit carries out brightness change search processing in the infrared camera image zone represented by a gray scale using a search area wherein the size of a first detected area detected by binarization and correlation operation serves as the reference. In addition, the parallax between the first detected area and the searched area having a change in brightness is compared, a search area having a parallax that differs from that of the first searched area is eliminated, and a search area having a parallax identical to that of the first searched area is provided a highlighted display as an object identical to that of the first detected area.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to a vehicle information providing apparatus that detects and displays, for example, a physical object present in the vicinity of a vehicle.
2. Background of the Invention
Conventionally, in order to notify the driver of a vehicle about an obstacle such as a pedestrian on the travel path, an apparatus is known in which an image from one or a plurality of infrared cameras mounted on the front of the vehicle is displayed at a position visible from the driver's seat to complement the front visual field of the driver. The image displayed to the driver is displayed on an image display apparatus such as a NAVIDisplay disposed in the dashboard of the vehicle, a HUD (Head Up Display) that displays information at a position on the front window that does not obscure the front visual field of the driver, a meter integrated display that is integrated with a meter that displays the travel state of the automobile numerically, or the like.
In addition, for example, Japanese Unexamined Patent Application, First Publication, Hei 11-328364 discloses an apparatus that records the environment in the vicinity of the vehicle and displays the result to the driver in this manner. In this apparatus, the temperature of the exposed skin of the head of the detected object is higher than the temperature of the other parts, and first the position of the head of the detected object is identified using the fact that it is comparatively easy to record as an image of an infrared camera. The zone corresponding to the body of the detected object is determined based on the information of the identified position of the head. Thereby, for example, caution can be prompted by displaying to the driver not only the head of the pedestrian, but also the entire body.
However, in the conventional apparatus described above, in the case that the detected object is recorded from the front, the amount of infrared radiation of the skin of the face is large in comparison to other parts, and thus the position of the head of the detected object is easy to detect. However, although detecting the zone of the entire body therefrom is easy, in such a case as the detected object being recorded from the back, the parts other than the face are extracted, and it is difficult to identify the position of the head.
In addition, the temperature of parts of the body of the detected object change depending on the conditions and the environment. Examples are parts whose temperature rises due to direct exposure to the sun and parts whose temperature is lowered due to being blown on by the wind. Consequently, due to the increase or decrease in temperature, it is not always the case that the position of the head can be identified from the recorded image. The zone that corresponds to the body identified from the part recognized as the position of the head may be different from the zone in which the detected object is actually present.
In consideration of the problems described above, it is an object of the present invention to provide a vehicle information providing apparatus that determines and displays the range of presence of a person from the characteristics of images recorded by an infrared camera.
SUMMARY OF THE INVENTION
In order to solve the problems described above, a first aspect of the vehicle information providing apparatus of the present invention, which displays an image recorded by an infrared camera, comprises an extracted zone setting device (for example, step S 1 to step S 18 in the embodiments) that sets a present zone of the physical object detected by multivalued processing of the image as the extracted zone (for example, the first extracted zone 51), a search area setting device (for example, step S 21 to step S 24 in the embodiments) that sets the search area (for example, search areas 54 to 58 in the embodiments) in the vicinity of the extracted zone, and a physical object recognition device (for example, step S 25 to step S 37 in the embodiments) that searches for changes in brightness in the searched area, and provides a highlighted display of the area where there is a change in brightness along with the extracted zone as the same object.
The vehicle information providing apparatus having the structure described above sets the search area by setting as the reference the extracted zone set by the extracted zone setting device and searching the zone exhibiting a change in brightness in the vicinity of the extracted zone. Thereby, the zone in which a physical object identical to the object recorded in the extracted zone can be provided in a highlighted display along with the extracted zone.
A second aspect of the vehicle information providing apparatus of the present invention comprises two infrared cameras and a parallax calculating device (for example, step S 34 in the embodiments) that finds the parallax of the images recorded by the two cameras, and the physical object recognizing device gives a highlighted display of the zone having a change in brightness and parallax identical to the extracted zone as an identical physical body.
The vehicle information providing apparatus having the structure described above recognizes a zone having a change in brightness and parallax identical to the extracted area as the area in which a physical body identical to the physical body that was taken in the extracted area, and this can be given a highlighted display along with the extracted area.
In a third aspect of the vehicle information providing apparatus of the present invention, the physical object recognizing device stops searching for changes in brightness in the case that the adjacent search area (for example, search areas 52 and 53 in the embodiments) set above or below the extracted zone exceeds the range of the image.
In a vehicle information providing apparatus of the present invention having the structure, the physical object recognizing device gives a highlighted display of only an extracted zone without a search of the physical object by determining that the large part of the physical object in the extracted zone is recorded in the case that the zone of the image is exceeded when the search area setting device sets the search area above or below the extracted zone.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram showing the structure of the vehicle information providing apparatus according to an embodiment of the present invention.
FIG. 2 is a drawing showing the installation positions of the infrared cameras, sensors, display, and the like on the vehicle.
FIG. 3 is a flowchart showing the entire operation of the vehicle information providing apparatus according to the same embodiment.
FIGS. 4A and 4B show the gray scale image taken by the infrared camera and the binary image thereof.
FIGS. 5A, 5B, and 5C are drawings showing the conversion processing to a run length data and labeling.
FIGS. 6A and 6B are drawings showing the time tracking of the object.
FIG. 7 is a drawing showing the rotation angle correction of the object image.
FIGS. 8A and 8B are drawings showing the search area in the right image and the search area set in the left image.
FIG. 9 is a drawing showing the correlative operation processing in which the search area is set as the object.
FIGS. 10A and 10B are drawings showing the method of calculating the parallax between objects in the distance calculation of the object.
FIGS. 11A and 11B are drawings showing the offset of the position of the object in the image generated by the turning of the car.
FIG. 12 is a drawing showing the case in which a collision may easily occur.
FIG. 13 is a drawing showing the zone division in front of the vehicle.
FIG. 14 is a flowchart showing the highlighted image output processing operation according to the first embodiment of the same example.
FIGS. 15A and 15B are drawings showing an example of a zone division result in the image of the same example.
FIGS. 16A and 16B is a drawing showing the zone searched sequence in the image according to a first embodiment of the same example.
FIGS. 17A and 17B are drawings showing the zone searched sequence in the image according to a first embodiment of the same example.
FIG. 18 is a flowchart showing the brightness change search processing operation of the vehicle information providing apparatus of the same example.
FIGS. 19A and 19B are drawing showings the zone search sequence in an image according to a second and third embodiment of the example.
FIG. 20 is a drawing showing the zone search sequence in an image according to a fourth embodiment of the example.
FIGS. 21A and 21B are drawings showing an example of the zone division results in an image of the same example.
FIG. 22 is a drawing showing an example of the zone division results in an image of the same example.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
Below, the embodiments of the present invention will be explained with reference to the drawings.
FIG. 1 is a block diagram showing the structure of the vehicle information providing apparatus according to an embodiment of the present invention.
In FIG. 1, reference numeral 1 is an image processing unit providing a CPU (central control unit) that controls the vehicle information providing apparatus according to the present embodiment, and wherein two infrared cameras 2R and 2L that can detect infrared radiation, a yaw rate sensor 3 that detects the yaw rate of the vehicle, a vehicle speed sensor 4 that detects the travel speed (vehicle speed) of the vehicle, and brake sensors 5 that detect the operation of the brakes are connected together. Thereby, the image processing unit 1 detects a moving physical object such as a pedestrian or animal in front of the vehicle from an infrared image of the vicinity of the vehicle and a signal indicating the travel state of the vehicle, and issues a warning when it determines that the possibility of a collision is high.
In addition, in the image processing unit 1, a speaker 6 for issuing a warning by voice, and an image display apparatus 7 including, for example, a meter integrated Display having a meter integrated for showing numerically the travel condition of the vehicle, a NAVIDisplay mounted in the dashboard of the vehicle, or a HUD (head up display) 7 a that displays information at a position on the front window at a position that does not obscure the front visual field of the driver, are connected. The image display apparatus 7 is for displaying the image recorded by the infrared cameras 2R and 2L, and informing the driver about objects with which there is a high danger of collision.
In addition, the image processing unit 1 comprises an A/D conversion circuit for converting input analog signals into digital signals, image memory for storing the digitalized image signal, a CPU (central processing unit) that carries out each of the operation processes, RAM (random access memory) that the CPU uses to store data that is being processed, ROM (read only memory) for storing programs executed by the CPU, tables, maps, or the like, a drive signal for the speaker 6, and an output circuit that outputs display signals or the like from the HUD 7 a or the like. Each of the output signals of the infrared cameras 2R and 2L, the yaw rate sensor 3, the vehicle speed sensor 4, and the brake sensors 5 are formed so as to be converted to digital signals and then input into the CPU.
In addition, as shown in FIG. 2, on the front of the vehicle 10, the infrared cameras 2R and 2L are mounted at a positions substantially symmetrical with respect to the center in the transverse direction of the vehicle 10, the optical axis of the two infrared cameras 2R and 2L are mutually parallel, and their heights above the surface of the road are fixed so as to be equal. Moreover, the infrared cameras 2R and 2L have the characteristic that the output signal level becomes higher (the brightness increases) the higher the temperature of the object.
In addition, the HUD 7 a is mounted so that the display image is displayed at a position on the window of the vehicle 10 that does not obscure the visual field of the driver.
Next, the operation of the present embodiment will be explained with reference to the figures.
FIG. 3 is a flowchart showing the processing sequence in the image processing unit 1 of the vehicle information providing apparatus according to the same embodiment.
First, the image processing unit 1 records the infrared image, which is the output signal of the infrared cameras 2R and 2L (step S 1), carries out A/D conversion thereon (step S 2), and stores the gray scale image in the image memory (step S 3). Moreover, here the right image is obtained by the infrared camera 2R and the left image is obtained by the infrared camera 2L. In addition, because the horizontal position of an identical object on the display screen is displayed out of alignment in the right image and the left image, the distance to the object can be calculated using this misalignment (parallax).
Next, the right image obtained from the infrared camera 2R is made the reference image, and binary processing of this image signal is carried out. In this binary processing, a 1 (white) is assigned to a zone brighter than a brightness threshold value ITH, and a 0 (black) is assigned to a dark zone.
FIG. 4A shows a gray scale image obtained by the infrared camera 2r, and by carrying out binary processing thereon, the image shown in FIG. 4B is obtained. Moreover, in FIG. 4B, the physical object surrounded by the frame from P1 to P4, for example, will be an object (below, referred to as a “high brightness zone”) displayed as white on the display screen.
When the binary image data from the infrared cameras is obtained, processing is carried out in which the binary image data is converted to run length data (step S 5).
FIG. 5A is a drawing to explain this, and in this figure the zone that has become white due to binary conversion is shown as the lines L1 to L8. Lines L1 to L8 all have a width of 1 pixel in the y direction, and while they are actually arranged without a space between them in the y direction, they have been separated for the sake of the explanation. In addition, the lines L1 to L8 respectively have the lengths 2 pixels, 2 pixels, 3 pixels, 8 pixels, 7 pixels, 8 pixels, 8 pixels, and 8 pixels. The run length data is shown by the coordinates of the start point of each of the lines (the point on the left end of each line) and the length (number of pixels) from the start point to the end point (the point on the right end of each line). For example, line L3 comprises the 3 pixels (x3, y5), (x4, y5), and (x5, y5), and thus (x3, y5,3) becomes the run length data.
Next, from the image data converted into run length data, by labeling the object (step S 6), the processing in which the object is extracted is carried out (step S7). That is, as shown in FIG. 5B, among the lines L1 to L8 that have been converted to run length data, the lines L1 to L3, which are the parts overlapping in the y direction, are treated as one object 1, lines L4 to L8 are treated as one object 2, and the object labels 1 and 2 are added to the run length data. By this processing, for example, the high brightness zones shown in FIG. 4B are respectively recognized as objects 1 through 4.
When the extraction of the objects has completed, as shown in FIG. 5C, next the center of gravity G, surface zone S, and the aspect ratio ASPECT of the circumscribed quadrangle represented by the broken lines is calculated (step S 8).
Here, the surface zone S is calculated by adding the lengths of the run length data for the same object. In addition, the coordinate of the center of gravity G is calculated as the x coordinate of the line that bisects the surface zone S in the x direction, and the y coordinate of the line that bisects it in the y direction. Furthermore, the aspect ratio ASPECT is calculated as the Dy/Dx ratio of Dy and Dx shown in FIG. 5C. Moreover, the position of the center of gravity G can be substituted for by the position of the center of gravity of the circumscribed quadrangle.
When the center of gravity, the surface zone, and the aspect ratio of the circumscribed quadrangle have been calculated, next recognition of the time trace, that is, the sampling of each cycle, of the same object is carried out (step S 9). In a time trace, k serves as the time during which time t, an analogue amount, is made discrete at a sampling cycle, and as shown in FIG. 6A, in the case that objects A and B are extracted at time k, obj cots C and D extracted at time (k+1) are determined to be identical to obj cots A and B. Specifically, when the following identity determination conditions 1 to 3 are satisfied, objects A and B are determined to be identical to objects C and D, and objects C and D have their labels changed respectively to objects A and B.
1) When the position coordinates of the center of gravity in the image of the object i (=A, B) at time k are set respectively to (xi (k), yi (k)) and the position coordinates of the center of gravity in the image of the object j (=C, D) at time (k+1) are set respectively to (xj (k+1), yj (k+1)), then |xj (k+1)−xi (k)|<Δx|yj (k+1)−yi (k)|<Δy, where Δx and Δy denote the allowable values of the amount of movement in the picture respectively in the x direction and the y direction.
2) When the surface zone of the object i (=A, B) in the image at time k is Si (k) and the surface zone of the object j (=C, D) in the image at time (k+1) is Sj (k+1), then Sj (k+1)/Si (k)<1±ΔS, where ΔS denotes the allowable values of the change in zone.
3) When the aspect ratio of the circumscribed quadrangle of the object i (=A, B) at time k is ASPECT i (k) and the aspect ratio of the circumscribed quadrangle of the object j (=C, D) is ASPECT j (k+1), then ASPECT j (k+1)/ASPECT i (k)<1±Δ ASPECT, where Δ ASPECT denotes the allowable values of the aspect ratio.
For example, when comparing FIG. 6A and FIG. 6B, although the size of each of the objects in the image becomes larger, the object A and the object C satisfy the conditions for the identification of identity described above and the object B and the object D satisfy the conditions for the identification of identity described above, and thus the objects C and D are respectively recognized to be the objects A and B. In this manner, the positions coordinates (of the center of gravity) of each of the recognized objects is stored in the memory as time series position data to be used on later calculation processing.
Moreover, the processing in steps S 4 to S 9 explained above is carried out on a binary reference image (in this embodiment, the right image).
Next, the velocity VCAR detected by the velocity sensor 4 and the yaw rate YR detected by the yaw rate sensor 3 are read, and as shown in FIG. 7, the turning angle θr of the vehicle 10 is calculated by integrating the yaw rate YR with respect to time (step S 10).
In contrast, the processing of step S 9 and step S 10 is carried out in parallel, and in steps S 11 through S 13, processing that calculates the distance z between the object and the vehicle 10 is carried out. Because this calculation requires a longer time than step S 9 and step S 10, they are executed at a longer cycle than step S 9 and S 10 (for example, at a cycle about three times the execution cycle of steps S 1 to S 10).
First, by selecting one among the objects that is traced out by the binary image of the reference object (the right image), as shown in FIG. 8A, in the search image R1 (here, the entire zone surrounded by the circumscribed quadrangle is made the searched image) is extracted from the right image (step S 11).
Next, the vehicle search area in which the image corresponding to the searched image (below, referred to as the “corresponding image”) from the left image is set, and the corresponding image is extracted by executing the correlation calculation (step S 12). Specifically, as shown in FIG. 8B, depending on each of the peak coordinates of the searched image R1, the search region R2 in the left image is set, and the brightness difference total value C (a, b), which indicates the degree of the correlation with the searched image R1 in the search region R2, is calculated by the Eq. 1 shown below, and the zone in which this total value C (a, b) becomes minimum is extracted as the corresponding image. Note that this correlation calculation is carried out using the gray scale image, not the binary image.
In addition, when there is past position data for the identical physical body, based on this position data, a zone R2 a (shown by the broken line in FIG. 8B) that is narrower than the search region R2 is set to serve as the search area.
C ( a , b ) = n = 0 N - 1 m = 0 M - 1 | I L ( a + m - M , b + n - N ) - I R ( m , n ) | Eq . 1
Here, IR(m, n) is the brightness value of the position of the coordinate (m, n) in the search area R1 shown in FIG. 9 and IL(a+m−M, b+n−N) is the brightness value of the position of the coordinate (m, n) in the search one R1 and the local zone R3 having the same shape, where the coordinates (a, b) in the search area are the base points. The position of the corresponding image is defined by finding the position at which the total value C (a, b) of the brightness difference is minimized by changing the coordinates (a, b) of the base point.
Due to the processing in step S 12, as shown in FIG. 10A and FIG. 10B, because the searched image R1 and the corresponding image R4 corresponding to this object are extracted, next the distance dR (number of pixels) between the position of the center of gravity of the searched image R1 and the image center line LCTR and the distance dL (number of pixels) between the position of the center of gravity of the corresponding image R4 and the image center line LCTR are found, and by applying the following Eq. 2, the distance z between the vehicle 10 and the object is calculated (step S 13).
z = B x F ( d L + d R ) × p = B x F Δ d × p Eq . 2
Here, B is the base line length, that is, the distance in the horizontal direction between the center position of the photographic element of the infrared camera 2R and the center position of the photographic element of the infrared camera 2L (the separation of the light beam axis of both infrared cameras); F is the focal distance of the lenses of the infrared cameras 2R and 2L, p is the pixel separation in the photographic element of the infrared cameras 2R and 2L, and Δd (=dR+dL) is the amount of parallax.
When the calculation of the turning angle θr in step S 10 and the calculation of the distance to the object in step S 13 have completed, the coordinates (x, y) in the image and the distance z calculated by Eq. 2 are applied to the following Eq. 3, and converted to real spatial coordinates (X, Y, Z) (step S 14).
Here, as shown in FIG. 2, the real spatial coordinates (X, Y, Z) have as their origin O the position of the center point of the installation position of the infrared cameras 2R and 2L (the position at which they are fixed on the vehicle 10), they are fixed as shown in the figure, and the coordinates in the image are determined by x in the horizontal direction and y in the vertical direction, where the center of the image serves as the origin.
[ X Y Z ] = [ x c × z / f y c × z / f z ] Eq . 3
where f=F/p.
Here, (xc, yc) are the coordinates (x, y) of the right image that have been converted to coordinates of a virtual image in which the real spatial origin O and the center of the image have been made to coincide based on the relative positional relationship between the installation position of the infrared camera 2R and the real special origin O. In addition, f is the ratio of the focus distance F and the pixel interval p.
In addition, when the real spatial coordinates have been found, turning angle compensation is carried out in order to compensate the positional shift in the image due to the turning of the vehicle 10 (step S 15).
As shown in FIG. 7, when the vehicle turns, for example, at a turning angle θr in the left direction during the time interval from time k to (k+1), a shift in the x direction by an amount equivalent to Δx, as shown in FIG. 11, occurs in the image obtained by the camera, and the turning angle compensation is a process to compensate this. Specifically, in the following Eq. 4, the real spatial coordinate system (X, Y, Z) is applied, and the compensated coordinates (Xr, Yr, Zr) are calculated. The calculated real spatial position data (Xr, Yr, Zr) is associated with each object and stored in memory. Moreover, in the following explanation, the coordinates after turning angle compensation are denoted (X, Y, Z).
[ X r Y r Z r ] = [ cos θ r 0 - sin θ r 0 1 0 sin θ r 0 cos θ r ] [ X Y Z ] Eq . 4
When the turning angle compensation for the real coordinates has completed, next, the approximately straight line LMV corresponding to the relative motion vector between the object and the vehicle 10 is found from N real spatial position data (for example, N=10) after turning angle compensation obtained during the monitoring period ΔT for one and the same object, that is from the time series data, (step S 16).
Concretely, when the direction vector L, which denotes the direction of the approximately straight line LMV, is equal to (lx, ly, lz) where (|L|=1), the straight line represented by the following Eq. 5 is found.
X=u·lx+Xav
Y=u·ly+Yav
Z=u·lz+Zav
X a v = j = 0 N - 1 X ( j ) / N Y a v = j = 0 N - 1 Y ( j ) / N Z a v = j = 0 N - 1 Z ( j ) / N Eq . 5
Here, u is a parameter that takes an arbitrary value, and Xav, Yav, and Zav are respectively the average values of the X coordinate, Y coordinate, and Z coordinate of the real spatial position data sequence.
Moreover, when the parameter u is eliminated, Eq. 5 becomes to Eq. 5a:
(X−Xav)/1x=(Y−Yav)/1y=(Z−Zav)/1z  Eq. 5a
In addition, in the case, for example, that P(0), P(1), P(2), . . . , P(n−2), P(N−1) denote the time series data after turning angle compensation, the approximately straight line LMV passes through the average position coordinate Pav=(Zav, Yav, Zav) of the time sequence data, and is found as the straight line which is characterized in that the average value of the square of the distance from each of the data points is minimal.
Here, the numerical value in the parenthesis added to P, which denotes the coordinates of each of the data point, indicates that the larger the value, the older the data. For example, P(0) denotes the most recent position coordinate, P(1) denotes the position coordinate of one sample cycle back, and P(2) denotes the position coordinate two sample cycles back.
Next, when the most recent position coordinate P(0)=(X(0), Y(0), Z(0)), the position coordinate P(N−1)=(X (N−1), Y(N−1), Z(N−1)) of the (N−1) sample back (before time ΔT) is compensated to a position on the approximately straight line LMV. Concretely, by applying the Z coordinates Z(0), Z(N−1) to the Eq. 5a above, that is, the following Eq. 6, the position coordinates after compensation Pv(0)=(Xv(0), Yv(0), Zv(0)) and Pv(N−1)=(Xv(N−1), Yv(N−1), Zv(N−1)) are found.
X v ( j ) = ( Z ( j ) - Z a v ) × l x l z - X a v Y v ( j ) = ( Z ( j ) - Z a v ) × l y l z - Y a v Eq . 6
Zv(j)=Z(j)
j=0,N−1
The relative motion vector is found as the vector from the position coordinate Pv(N−1) calculated in Eq. 8 towards Pv(0).
By finding the relative motion vector by calculating the approximately straight line that approximates the relative motion locus of the object with respect to the vehicle 10 from a plurality (N) of data within the monitoring period ΔT in this manner, the influence of position detection error can be reduced, and the possibility of a collision with the object can be more correctly predicted.
In addition, in step S 16, when the relative motion vector has been found, next the possibility of a collision with the detected object is determined, and a warning determination process, which issues a warning when the possibility is high, is executed (step S 17).
The warning determining processing (step S 17) is processing that determines the possibility of a collision between the vehicle 10 and the detected object by using any one of the collision detecting processing, processing to determine whether or not the object is in an approach determination zone, or invasive collision determination processing, which are described below. Below, as shown in FIG. 12, the explanation will treat an example wherein there is an animal 20 being approached at a speed Vp from a substantially 90° with respect to the direction of forward motion of the vehicle 10.
Collision Determining Processing
First, the image processing unit 1 calculates the relative velocity Vs in the Z direction using the following Eq. 7 from the animal 20 approaching the distance Zv(0) from the distance Zv(N−1) during the time ΔT, and carries out collision determination processing. The collision determination processing is processing that determines whether there is a possibility of a collision when the following Equations 8 and 9 are satisfied.
Vs=(Zv(N−a)−Zv(0))/ΔT  Eq.7
Zv(0)/Vs≦T  Eq. 8
|Yv(0)|≦H  Eq. 9
Here, Zv(0) is the most recent distance detection value (v is attached in order to indicate that this is data after compensation using the approximately straight line LMV, while the Z coordinate is a value identical to that before compensation), and Zv(N−1) is the detected distance value before the time ΔT. In addition, T is an allowable time and signifies that the possibility of a collision is determined time T before the predicted collision time, and is about 2 to 5 seconds, for example. In addition, H is a predetermined height that defines the range of the Y direction, that is the height direction, and is set, for example, to about twice the height of the vehicle 10.
Processing for determining whether the object is in an approach determination zone
Here, it is determined whether or not an object is present in the approach determination zone. For example, in FIG. 13, the zone that can be monitored by the infrared cameras 2R and 2L is indicated by the zone AR0 in the circumscribed triangle indicated by the bold solid line, and zones AR1, AR2, and AR3 in the zone AR0, which are closer to the vehicle 10 than Z1=Vs×T serve as the warning zones.
Here, AR1 is the zone corresponding to the range having added the allowance β (for example, about 50 to 100 cm) to both sides of the width α of the vehicle 10, or in other words, the zone having a width (α/2+β) on both sides of the axle at the center part in the width direction of vehicle 10, and if the object continues to be present as-is, the possibility of a collision is extremely high. Thus, these zones are called approach determination zones. The zones AR2 and AR3 are zones (in the outside transverse direction of the approach determination zone) in which the absolute value of the X coordinate is larger than the approach determination zone, an invasive collision determination, described below, is made about the object inside this zone, and thus this is called the invasive determination zone. Moreover, these zones have a predetermined height H in the Y direction, as shown in the above Eq. 9.
Invasive Collision Determining Processing
Concretely, the invasive collision determination processing in step S 23 distinguishes whether or not the difference between xc(0), which is the most recent x coordinate on the image (the character c, as will be explained below, is attached in order to signify that it is a coordinate on which compensation has been carried out that makes the center position of the image align with the real spatial origin point O) and xc(N−1), which is the x coordinate before the time ΔT, satisfies the following Eq. 10, and in the case that it is satisfied, it is determined that the possibility of a collision is high.
- α · f 2 ( 1 Z v ( 0 ) - 1 Z v ( N - 1 ) ) x c ( 0 ) - x c ( N - 1 ) α · f 2 ( 1 Z v ( 0 ) - 1 Z v ( N - 1 ) ) Eq . 10
Moreover, as shown in FIG. 14, in the case that there is an animal progressing in a direction that is at an angle of approximately 90° with respect to the direction of the progress of the vehicle 10, when Xv(N−1)/Zv(N−1)=Xv(0)/Zr(0), or in other words, when the ratio of the velocity Vp and the relative velocity Vs of the animal is Vp/Vs=Xr(N−1)/Zr(N−1), the bearing θd viewing the animal 20 from the vehicle 10 becomes constant, and the possibility of a collision becomes high. Eq. 10 determines this possibility taking into account the width α of the vehicle 10.
In the warning determining processing (step S 17), in any of the collision determining processing, the processing for determining whether there is an object in the approach determining zone, or the invasive collision determining processing, in the case that it is determined that there is no possibility of collision between the vehicle 10 and the detected object (NO in step S 17), the processing returns to step S 1, and the processing described above repeats.
In addition, in the warning determining processing (step S 17), in the case that it is determined that there is a possibility of a collision between the vehicle 10 and the detected object by any of the collision determining processing, the processing to determined whether there is an object in the approach determining zone, or the invasive collision determining processing (YES in step S 17), the processing proceeds to the warning output determining processing of step S 18.
In step S 18, the warning output determining processing, that is, determining whether or not to carry out the warning output, is carried out (step S 18).
The warning output determination process determines whether or not the driver of the vehicle 10 is carrying out a braking action from the output BR of the brake sensor 5.
In the case that the driver of the vehicle 10 is carrying out a braking action, the acceleration Gs (positive in the deceleration direction) generated thereby is calculated, and when this acceleration Gs is larger than a predetermined threshold value GTH, it is determined that a collision can be avoided by the braking action, and the warning determination processing completes (NO in step S 18), the processing returns to step S 1, and the processing described above is repeated.
Thereby, when an appropriate braking action is carried out, no warning is issued, and the driver will not be excessively annoyed.
In addition, when the acceleration Gs is equal to or less than a predetermined threshold GTH, and additionally, if the driver of the vehicle 10 is not carrying out a braking action, the flow immediately proceeds to the processing in step S 19 (YES in step S 18). Because the possibility of contact with the object is high, a warning message is issued via speaker 3 (step S 19), and at the same time, a gray scale image obtained, for example, from the infrared camera 2R is displayed on the image display apparatus 7, and the highlighted display frame is set on the object being approached. Thereby, the object is displayed to the driver of the vehicle 10 as a highlighted image (step S 20).
Moreover, the predetermined threshold value GTH is determined by the following Eq. 11. This is the value corresponding to the condition in which the vehicle 10 stops at a running distance equal to or less than the distance Zv(0) in the case that the acceleration Gs during the braking action is maintained as-is.
G T H = V s 2 2 × Z v ( 0 ) Eq . 8
Next, the highlighted display output processing in step S 20 shown in the flowchart in FIG. 3 will be explained with reference to the flowcharts shown in FIG. 14 and FIG. 18 and the drawings shown in FIG. 15 and FIG. 17.
First Embodiment
FIG. 14 is a flowchart showing the entire operation of the highlighted image output processing operation according to the first embodiment.
In FIG. 14, first the size of the area (below, referred to as the “first detected area”) of the object detected by binarization and correlation operation is set as the reference, a plurality of areas are established by partitioning the zone of the first detected area in the vertical direction such that their size is identical to that of the first detected area, and this serves as the search area (step S 21).
In addition, in establishing the plurality of search areas by partitioning the zone of the first detected area vertically, it is determined whether or not the partitioned search area is in the zone of the reference image (fight image) (step S 22), the processing returns to step S 21 until the search area protrudes from the reference image, and then repeats the area partition of the zone (establishing the search area) (YES in step S 22).
In contrast, when the search area protrudes from the reference image (NO in step S 22), the number of areas partitioned in the reference image is calculated (step S 23).
In addition, it is determined whether or not the number of areas in the reference image zone is greater than one (step S 24).
In step S 24, in the case that the number of areas in the reference image zone is one (the case in which the number of areas is one in the first detected area in the reference image zone and the number of areas is 2 in the search area protruding vertically from the reference image zone, making a total of three) (NO in step S 24), the first detected area is set as the highlighted display area (step S 25).
FIGS. 15A and 15B are drawings showing an example of an image in the case that the object 30 is a human and the number of areas in the reference image zone 50 is one. FIG. 15A is the case in which the entire body of the object 30 has been taken in the first detected area 51, and whether this is the entire body can be determined by the area size and distance of the object 30. In contrast, 15B is the case in which a part of the body of the object 30 has been taken in the first detected area 51. In either case, the first detected area 51 in the reference image zone 50 and the search areas 52 and 53 that vertically protrude from the reference image zone 50 are established, making a total of three areas.
Next, in step S 24, in the case that the number of areas in the reference image zone is more than one (YES in step S 24), the search area setting is carried out by making the row that includes the first detected area 51 the reference, and the register j for distinguishing rows is reset (j=0) (step S 26).
In addition, it is determined whether or not a change in the brightness of the image in the search area is recognized, the search area in the row designated by the register j is determined in the longitudinal direction for each search area, and the brightness change search processing that sets the flag in the search area in which a change in brightness in the image is recognized is carried out (step S 27). Moreover, the details of the brightness change search process will be described below.
FIG. 16A shows an image row in the case that the object 30 is a person and the number of areas in the reference image zone 50 is greater than one. For example, in FIG. 16A, a row is formed comprising the first detected area 51, the search areas 54 and 55 that are present in the vertical direction thereto and protrude from the reference image zone 50, and further, a plurality of search areas 56 present in the reference image zone 50, for a total of M areas. In addition, because this row serves as the reference, this row is set to j=0.
Next, it is determined whether or not there is a flag indicating that a change in brightness was recognized in the image in the search area of the row designated by the register j (step S 28).
In step S 28, in the case that there is a flag indicating that a change in brightness was recognized in the image in the search area in the designated row (YES in step S 28), the adjacent row to be searched presently is designated by increasing j by 1 (j=j+1) (step S 29), the processing returns to step S 27, and the brightness change search processing is carried out on the new row (step S 27).
In addition, in step S 28, in the case that there is no flag indicating that a change in brightness has been recognized in the image in the search area in the designated row (NO in step S 28), j is reset (j=0) (step S 30), j is decreased by 1 (j=j−1), and thereby a row on the opposite side of the one searched above is designated (step S 31).
Additionally, like step S 27 described above, brightness change search processing is carried out (step S 32).
Next, like step S 28 described above, it is determined whether or not there is a flag indicating that a change in brightness has been recognized in the image in the search area in the row designated by the register j (step S 33).
In step S 33, in the case that there is a flag indicating that a change in brightness has been recognized in the image in the search area in the designated row (YES in step S 33), the processing returns to step S 31, j is decreased by 1 (j=j−1), and thereby the adjacent row to be search presently is designated, and in step S 32 brightness change search processing is carried out on the new row (step S 32).
FIG. 16B is a drawing showing the case in which the search area searched up to j=2 where a recognized change in brightness is not present, and thus j=−1 on the opposite side of j=0 is searched.
In addition, in step S 33, in the case that there is no flag indicating that a change in brightness has been recognized in the image in the search area in the designated row (NO in step S 33), it is determined that the search in the reference image zone 50 has completed, and next the parallax calculation is carried out on the search area for which a flag has been set (step S 34).
Additionally, the flag of the search area having a parallax that differs from the first search area 51 (a differing distance from the vehicle 10) is cleared (step S 35).
In addition, when the search area having a differing parallax is eliminated, highlighted display is carried out such that the first search area 51 and the search area having a set flag are included (step S 36).
Moreover, the highlighted display output set in step S 25 and the highlighted display output set in step S 36 are output to the image display apparatus 7 (step S 37), and the highlighted display output processing is completed.
Based on FIG. 16A and FIG. 16B, FIG. 17A shows the state in which the highlighted display frame 59 that includes the first detected area 51 and the search area that includes a search area recognized to have a change in brightness is set. In addition, as shown in FIG. 17B, when output to the image display apparatus 7, the highlighted display frame 59 that includes the entire object, which is easily recognized as having an object shape, is set on the displayed gray scale image.
Next, the brightness change search processing will be explained using the flowchart shown in FIG. 18.
The brightness change search processing is processing in which it is determined whether or not a change in brightness in the image in the search area can be recognized in each search area in the row designated by the register j, and sets a flag in a search area where a change in brightness is recognized in the image. The search area in the row is searched in the direction of the row (longitudinal direction) with respect to the designated row.
First, in order to carry out a search in the direction of the column, the register i for distinguishing columns is set (i=1) (step S 41).
Next, for the search area in the row designated by the register i, a search is carried out to determine whether or not a change in brightness in the gray scale image of the search area is recognized (step S 42).
In addition, it is determined whether or not the change in brightness in the gray scale image of the designated area is large (step S 43), and in the case that the change in brightness is large (YES in step S 43), the flag is set for the corresponding search area (step S 44).
When a flag is set for a search area having a large change in brightness, i is increased by 1 (i=i+1) in order to carry out a search of the next column, and thereby the next column to be searched presently is designated.
In addition, in the case that the change in brightness of the gray scale image in the designated search area is not large (NO in step S 43), the processing returns to step S 45 without taking any action.
Moreover, as shown in FIGS. 16A and 16B, because the total number in each row of the first detected area 51 and the search area is M areas, the value of the register i is determined (step S 46), and the brightness change search processing returns to step S 42 until the value of i is larger than M, and the processing described above repeats (NO in step S 46).
Therefore, in step S 24 in the flowchart shown in FIG. 14, in the case that the number of areas in the reference image zone 50 is greater than 1 (YES in step S 24), brightness change search processing is executed on the search area in each of the rows (where the row j=0, search areas 54 and 55) that protrude from the reference image zone 50.
In addition, in step S 46, when the value of i becomes large (i>M) (YES in step S 46), the brightness change search processing ends.
Next, an example of a different implementation of the highlighted image output processing in the partition method of the area and the search method will be explained with reference to the drawings in FIGS. 19A, 19B and 20.
Second Embodiment
FIG. 19A is a drawing for explaining the second embodiment of the highlighted image output processing. In the case that the first embodiment enlarges the search area in the row direction, enlargement is carried out in the row direction by making all search areas included vertically in M areas serve as objects. In contrast, in the second embodiment, for the rows outside of the right and left j=±1, a search is carried out by enlarging only the area of a row where a search area having a set flag is present. In FIG. 19A, where j=1 and j=−1, brightness change search processing is carried out in the rows j=2 and j=−2 only in search area 57 at i=2, 3, 4, where a change in brightness has been recognized.
Specifically, in the case that the search areas are represented as A(i, j), the presence of a change in brightness is denoted 0 and the absence of a change in brightness is denoted X, the brightness change processing is carried out as follows:
  • A(1, −1): X, A(1, 0): X, A(1, 1): X→no area enlargement
  • A(2, −1): X, A(2, 0): 0, A(2, 1): X→no area enlargement
  • A(3, −1): 0, A(3, 0): 0, A(3, 1): 0→area enlargement present
  • A(4, −1): 0, A(4, 0): 0, A(4, 1): 0→area enlargement present
  • A(5, −1): 0, A(5, 0): 0, A(5, 1): 0→area enlargement present
  • A(6, −1): X, A(6, 0): X, A(6, 1): X→no area enlargement
  • A(M, −1): X, A(M, 0):X, A(M, 1): X→no area enlargement
Thereby, in the previous row, in a column in which there is no recognition of the change in brightness, the possibility that a change in brightness will be recognized on the other side is also small, and thus by omitting the unnecessary brightness change search processing, the processing load on the image processing unit is decreased.
Third Embodiment
FIG. 19B is a drawing for explaining the third embodiment of the highlighted image output processing. In the third embodiment, the area made by extending the search area in the horizontal direction is set using the size of the first detected area 51 as the reference, and in the case that the search area is expanded in the row direction, in the row that is the same as the first detected area 51, the column in which a change in brightness has been recognized is the object. In FIG. 19B, the area made by extending in the horizontal direction using the size of the first detected area 51 as the reference is represented as j=0, and in the row j=1 or j=−1, the search area 58 from i=1 to i=6, where a change in brightness has been recognized in the area where j=0, becomes the object of the brightness change search processing.
Thereby, processing that determines the column of the object of the brightness change search processing depending on the presence or absence of a brightness change when carrying out expansion in the row direction becomes unnecessary, and thus the processing load on the image processing unit 1 can be further decreased.
Fourth Embodiment
FIG. 20 is a drawing for explaining the fourth embodiment of the highlighted image output processing. In the fourth embodiment, irrespective of the size of the first detected area 51, the reference image zone 50 is partitioned by a search area having a predetermined size, and the brightness change search processing is carried out.
In FIG. 20, the brightness change search processing is executed using an area somewhat larger than the first detected area 51.
However, thereby, the brightness change search processing is carried out in the reference image zone 50 irrespective of the size of the first detected area 51, and thus although the processing load on the image processing unit 1 can be largely reduced, in the case that the difference in the size of the search area and the size of the object 30 is great, it is necessary to consider the case that this may entail a decrease in the detection precision and a decrease in the detection speed.
Moreover, when added to the first through fourth embodiments described above, in the case that the search area is expanded, the parallax calculation between the images of the two infrared cameras 2R and 2L can be carried out first, and then only the portion of the search area having identical parallax needs to be expanded.
In addition, because there are portions in which there is no change in brightness equal to or greater than a predetermined value even for the same physical object, vertical expansion that included this portion is also possible.
In addition, by clearing the flags of the areas having a parallax differing from the first detected area 51, even in the case that the objects are overlapping, the objects can be distinguished and extracted. For example, when something hidden is detected in the case that an object is present behind the vehicle, for example, there is a change in brightness in the search area, and at the same time, the search area has a parallax value that differs from the first detected area 51, and a plurality of search area having such parallax values are present. Thus, in the case that the distance of the area is found using the parallax values and it is thought that a person is present behind the body of the vehicle, the portion excluding the hidden area can be given a highlighted display.
Next, an example of the partition of the search area will be explained with reference to the drawings of the other examples of partitions.
Like the head portion of the object 30 explained in the first through fourth embodiments described above, FIG. 21A is an example of the partition of the reference image using medium sized search areas that have been set based on the first detected area 51, which has detected the torso of the object 30.
FIG. 21B is an example of the partition of a reference image using small sized search areas that have been set based on the first search area 51, which has detected the end of the hand of the object 30.
FIG. 22 is an example of the partition of the reference image using a transversely large sized search area set based on the first detected area 51, which has detected both shoulders and the chest portion of the object 30. As shown in FIG. 22, in the case that the size of the first detected are 51 in the transverse direction is large, the transverse size of the search area can be adjusted by the distance to the object found from the parallax of the image.
Moreover, in the embodiments described above, examples of monitoring the front of the vehicle were given, but monitoring can be carried out in any direction, such as the back of the vehicle.
In addition, in the embodiments described above, infrared camera 2R and 2L were provided, and the distance to the object found using parallax is also used as determination material for extracting identical physical objects. However, it is possible to provide one camera, and in this case, the distance to the object can be eliminated from the determination material used for extracting an identical physical object.
In addition, in the present embodiment, the image processing unit 1 includes the extracted zone setting device, the search zone setting device, the physical object recognition device, and the parallax calculating device. More concretely, steps S 1 to S 18 in FIG. 3 correspond to the extracted zone setting device, steps S 21 to S 24 in FIG. 14 correspond to the search zone setting device, and steps S 25 to S 37 in FIG. 14 correspond to the physical body recognizing device. Furthermore, step S 34 in FIG. 14 corresponds to the parallax calculating device.
As explained above, the vehicle information providing apparatus of the present embodiments sets the search area in the vicinity of the vehicle using a first detected area 51, which is the first area to be detected, as the reference, and in the case that a change in brightness is recognized in the search area and two infrared cameras are used, the search area having a parallax identical to the first search area 51 is determined to be the image of a physical object identical to that in the first search area 51, Thereby, it is possible to extract an image area that takes in the entire object reliably.
Therefore, in order to carry out conventional processing that extracts the head, which is assumed to have the highest surface temperature among the body members of a person, mistaken detection due to changes in the temperature of the body members of the detected object have been a problem. Examples of such parts are parts whose temperature had risen due to exposure to direct sunlight and the parts whose temperature had fallen due to being blown on by the wind. Thereby, even when the parts of the body or the like are first detected, there is the effect that the entire object (the entire body of the detected object) can always be taken.
As explained above, according to a first aspect of the vehicle information providing apparatus, a search zone is set using the extracted zone set by the extraction zone setting device as a reference, and by searching the zone that has a change in brightness in the vicinity of the extracted zone, a zone in which is it appears that a physical body can be recorded that is identical to the physical body recorded in the extracted zone can be given a highlighted display along with the extracted zone. Therefore, when the search zone is set based on the size of the extracted zone, the size of the search zone can be suitably set with respect to the object by using the size of the extraction zone that is proportional to the size of the object. Thereby, the effect is obtained that the detection speed of detecting the entire physical object by searching in the image zone using the search zone is increased.
According to a second aspect of the present invention, a zone having a change in brightness and a parallax identical to the extracted zone is recognized as a zone in which it is assumed that the physical object can be taken that is identical to the physical object taken in the extracted zone, and this can be given a highlighted display along with the extracted zone.
Therefore, even in the condition that a plurality of physical objects are overlapping, there is the effect that a plurality of objects can be separated and displayed by selecting a zone having the same parallax (distance to the object) as the image of the two cameras.
According to a third aspect of the vehicle information providing apparatus, in the case that the zone of the image is exceeded when the search zone setting device sets the search zone vertically with respect to the extraction zone, the physical object recognition device determines that a large part of the physical object has been recorded in the extracted zone, and carries out highlighted display of only the extracted zone without carrying out a search of the physical object.
Therefore, there is the effect that unnecessary processing can be avoided by stopping the search for physical bodies after it has been determined that a large part of the physical body has been taken in the extracted zone.
Although there have been described what are the present embodiments of the invention, it will be understood by persons skilled in the art that variations and modifications may be made thereto without departing from the spirit or essence of the invention.

Claims (9)

1. A vehicle information providing apparatus that displays an image recorded by at least one infrared camera, the apparatus comprising:
an extracted zone setting device that sets a present zone of a physical object image, which corresponds to a physical object detected by multivalued processing of said image, as an extracted zone;
a search zone setting device that sets a search zone for detecting changes in brightness in the vicinity of said extracted zone; and
a physical object recognition device that searches for changes in brightness in said search zone, and provides a highlighted display of a portion of the search zone in which there is a change in brightness, along with said extracted zones, as the physical object image.
2. A vehicle information providing apparatus according to claim 1, comprising:
two infrared cameras; and
a parallax calculating device that finds the parallax of the images recorded by said two cameras; and wherein
said physical object recognition device provides a highlighted display of the portion of the search zone in which there is a change in brightness and which has a parallax identical to said physical object image.
3. A vehicle information providing apparatus according to claim 1, wherein, in the case that the search zone is set adjacent to an upper edge or a lower edge of the extracted zone and exceeds a range of said image, said physical object recognition device stops searching for changes in brightness.
4. A vehicle information providing apparatus according to claim 1,
further including a warning device which provides a warning to a driver of the vehicle when there is a high danger of collision by the vehicle with an object recognized by the physical object recognition device.
5. A vehicle information providing apparatus according to claim 1,
wherein said search zone includes a plurality of search areas, and said physical object recognition device selectively searches for changes in brightness in said search areas.
6. A vehicle information providing apparatus according to claim 5,
wherein said search areas separated from each other in at least one of columns and rows, and said physical object recognition device searches for changes in brightness in various ones of said search areas by said at least one of columns and rows based on detected changes in brightness in others of said search areas.
7. A vehicle information providing apparatus according to claim 1,
wherein said search zone includes a plurality of search areas, and said physical object recognition device searches for changes in brightness only in various ones of said search areas depending on brightness changes detected in others of said search areas.
8. A vehicle information providing apparatus according to claim 1,
wherein a size of said search zone is set proportional to a size of said extracted zone.
9. A vehicle information providing apparatus according to claim 1, wherein
the highlighted display comprising a frame which surrounds both said portion of the search zone in which there is a change in brightness, and said physical object image.
US10/273,314 2001-10-29 2002-10-17 Vehicle information providing apparatus Active 2024-11-11 US7141796B2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2001-331287 2001-10-29
JP2001331287A JP3764086B2 (en) 2001-10-29 2001-10-29 Vehicle information providing device

Publications (2)

Publication Number Publication Date
US20030083790A1 US20030083790A1 (en) 2003-05-01
US7141796B2 true US7141796B2 (en) 2006-11-28

Family

ID=19146890

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/273,314 Active 2024-11-11 US7141796B2 (en) 2001-10-29 2002-10-17 Vehicle information providing apparatus

Country Status (4)

Country Link
US (1) US7141796B2 (en)
JP (1) JP3764086B2 (en)
DE (1) DE10247371B4 (en)
GB (1) GB2383222B (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050137750A1 (en) * 2003-12-23 2005-06-23 Samsung Electronics Co., Ltd. Method and apparatus for using rotational movement amount of mobile device and computer-readable recording medium for storing computer program
US20050231339A1 (en) * 2004-02-17 2005-10-20 Fuji Jukogyo Kabushiki Kaisha Outside-vehicle monitoring system
US20060115122A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060114320A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co. Ltd. Position detecting apparatus and method of correcting data therein
US20060115114A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115119A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115117A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co. Ltd. Position detecting apparatus and method of correcting data therein
US20060115118A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115126A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle vicinity monitoring apparatus
US20060115163A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Apparatus for and method of extracting image
US20060126896A1 (en) * 2004-11-30 2006-06-15 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060126898A1 (en) * 2004-11-30 2006-06-15 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060151223A1 (en) * 2002-11-16 2006-07-13 Peter Knoll Device and method for improving visibility in a motor vehicle
US20060204037A1 (en) * 2004-11-30 2006-09-14 Honda Motor Co., Ltd. Vehicle vicinity monitoring apparatus
US7969466B2 (en) 2004-11-30 2011-06-28 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
EP2354316A1 (en) * 2010-02-05 2011-08-10 Kässbohrer Geländefahrzeug AG Piste maintenance vehicle
US20120106786A1 (en) * 2009-05-19 2012-05-03 Toyota Jidosha Kabushiki Kaisha Object detecting device
US20130033600A1 (en) * 2011-08-01 2013-02-07 Hitachi, Ltd. Image Processing Device
US20130322691A1 (en) * 2012-06-01 2013-12-05 Ricoh Company, Ltd. Target recognition system and target recognition method executed by the target recognition system

Families Citing this family (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3987048B2 (en) * 2003-03-20 2007-10-03 本田技研工業株式会社 Vehicle periphery monitoring device
US7088310B2 (en) * 2003-04-30 2006-08-08 The Boeing Company Method and system for presenting an image of an external view in a moving vehicle
US7046259B2 (en) 2003-04-30 2006-05-16 The Boeing Company Method and system for presenting different views to passengers in a moving vehicle
US6866225B2 (en) 2003-04-30 2005-03-15 The Boeing Company Method and system for presenting moving simulated images in a moving vehicle
US8587664B2 (en) * 2004-02-02 2013-11-19 Rochester Institute Of Technology Target identification and location system and a method thereof
JP3934119B2 (en) * 2004-06-14 2007-06-20 本田技研工業株式会社 Vehicle periphery monitoring device
JP3987057B2 (en) * 2004-06-14 2007-10-03 本田技研工業株式会社 Vehicle periphery monitoring device
JP4449618B2 (en) * 2004-07-22 2010-04-14 株式会社デンソー Vehicle perimeter monitoring system
JP4246691B2 (en) 2004-11-30 2009-04-02 本田技研工業株式会社 Image information processing system, image information processing method, image information processing program, and automobile
JP4246690B2 (en) 2004-11-30 2009-04-02 本田技研工業株式会社 Image information processing system, image information processing method, image information processing program, and automobile
JP4530827B2 (en) * 2004-12-08 2010-08-25 ダイハツ工業株式会社 Image processing method and image processing apparatus
US7489236B2 (en) * 2005-04-29 2009-02-10 Chavarria Faustino V Pedestrian alert system for vehicles
US7528372B2 (en) 2005-10-19 2009-05-05 Csi Technology, Inc. Apparatus and method for infrared imaging with performance algorithm
JP2007174113A (en) * 2005-12-20 2007-07-05 Sumitomo Electric Ind Ltd Obstacle detection system and obstacle detection method
DE102006047777A1 (en) * 2006-03-17 2007-09-20 Daimlerchrysler Ag Virtual spotlight for marking objects of interest in image data
JP4910529B2 (en) * 2006-07-14 2012-04-04 住友電気工業株式会社 Obstacle detection system and obstacle detection method
GB2443664A (en) * 2006-11-10 2008-05-14 Autoliv Dev An infra red object detection system de-emphasizing non relevant hot objects
JP4887537B2 (en) * 2007-08-02 2012-02-29 本田技研工業株式会社 Vehicle periphery monitoring device
US8391557B2 (en) * 2007-09-14 2013-03-05 Magna International Inc. Object detection and ranging method
FR2947935B1 (en) * 2009-07-08 2012-03-02 Valeo Vision Sas METHOD FOR DEFINING A SEARCH WINDOW
JP5158063B2 (en) * 2009-12-02 2013-03-06 株式会社デンソー Vehicle display device
KR101726682B1 (en) * 2011-04-27 2017-04-13 한화테크윈 주식회사 A method and an apparatus for computing object information
CN103568990A (en) * 2012-07-23 2014-02-12 北京新岸线移动多媒体技术有限公司 Method and system for achieving vehicle safety warning
JP5895858B2 (en) * 2013-01-16 2016-03-30 トヨタ自動車株式会社 Object detection device and object detection method, alarm device and alarm method, driving support device and driving support method
CN104149685A (en) * 2013-05-13 2014-11-19 赵盾 Automobile warning device
JP5870993B2 (en) * 2013-12-27 2016-03-01 トヨタ自動車株式会社 Vehicle information display device and vehicle information display method
JP5942979B2 (en) * 2013-12-27 2016-06-29 トヨタ自動車株式会社 Vehicle information display device and vehicle information display method
DE102014003550A1 (en) * 2014-03-12 2014-09-18 Daimler Ag Method and device for assisting a driver in driving a vehicle
KR102451091B1 (en) 2015-07-08 2022-10-05 엘지이노텍 주식회사 Apparatus and method for providing the moving body around vehicle, and recording medium for recording program performing the method
CN106564428A (en) * 2015-10-12 2017-04-19 常州博显汽车电子有限公司 Braking warning system and warning method based on infrared rays
CN105291978B (en) * 2015-11-27 2016-09-14 安徽工程大学 Vehicle lane change auxiliary early warning system and control method thereof
WO2018151759A1 (en) 2017-02-20 2018-08-23 3M Innovative Properties Company Optical articles and systems interacting with the same
KR20200061370A (en) 2017-09-27 2020-06-02 쓰리엠 이노베이티브 프로퍼티즈 캄파니 Personal protective equipment management system using optical patterns for equipment and safety monitoring
TWI688502B (en) * 2018-02-14 2020-03-21 先進光電科技股份有限公司 Apparatus for warning of vehicle obstructions
CN109163707B (en) * 2018-09-06 2019-11-26 百度在线网络技术(北京)有限公司 Method for barrier perception, system, computer equipment, computer storage medium
WO2021008712A1 (en) * 2019-07-18 2021-01-21 Toyota Motor Europe Method for calculating information relative to a relative speed between an object and a camera
CN112106017A (en) * 2019-07-29 2020-12-18 深圳市大疆创新科技有限公司 Vehicle interaction method, device and system and readable storage medium
US20220242433A1 (en) * 2021-02-02 2022-08-04 GM Global Technology Operations LLC Saliency-based presentation of objects in an image
CN114283548A (en) * 2021-12-27 2022-04-05 北京科技大学天津学院 Fire continuous monitoring method and system for unmanned aerial vehicle

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0729014A (en) 1993-06-24 1995-01-31 Sharp Corp Image processor
JPH11328364A (en) 1998-05-13 1999-11-30 Nissan Motor Co Ltd Area estimating device and display processor
US6137531A (en) 1997-04-15 2000-10-24 Fujitsu Limited Detecting device for road monitoring
US6327536B1 (en) * 1999-06-23 2001-12-04 Honda Giken Kogyo Kabushiki Kaisha Vehicle environment monitoring system
US20030007074A1 (en) * 2001-06-28 2003-01-09 Honda Giken Kogyo Kabushiki Kaisha Vehicle zone monitoring apparatus
US6552742B1 (en) * 1999-09-16 2003-04-22 Fuji Jukogyo Kabushiki Kaisha Positional deviation adjusting apparatus of stereo image
US20050100192A1 (en) * 2003-10-09 2005-05-12 Kikuo Fujimura Moving object detection using low illumination depth capable computer vision

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3866328B2 (en) * 1996-06-06 2007-01-10 富士重工業株式会社 Vehicle peripheral three-dimensional object recognition device
JP4118452B2 (en) * 1999-06-16 2008-07-16 本田技研工業株式会社 Object recognition device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0729014A (en) 1993-06-24 1995-01-31 Sharp Corp Image processor
US6137531A (en) 1997-04-15 2000-10-24 Fujitsu Limited Detecting device for road monitoring
JPH11328364A (en) 1998-05-13 1999-11-30 Nissan Motor Co Ltd Area estimating device and display processor
US6327536B1 (en) * 1999-06-23 2001-12-04 Honda Giken Kogyo Kabushiki Kaisha Vehicle environment monitoring system
US6552742B1 (en) * 1999-09-16 2003-04-22 Fuji Jukogyo Kabushiki Kaisha Positional deviation adjusting apparatus of stereo image
US20030007074A1 (en) * 2001-06-28 2003-01-09 Honda Giken Kogyo Kabushiki Kaisha Vehicle zone monitoring apparatus
US20050100192A1 (en) * 2003-10-09 2005-05-12 Kikuo Fujimura Moving object detection using low illumination depth capable computer vision

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060151223A1 (en) * 2002-11-16 2006-07-13 Peter Knoll Device and method for improving visibility in a motor vehicle
US7747348B2 (en) * 2003-12-23 2010-06-29 Samsung Electronics Co., Ltd. Method and apparatus for using rotational movement amount of mobile device and computer-readable recording medium for storing computer program
US20050137750A1 (en) * 2003-12-23 2005-06-23 Samsung Electronics Co., Ltd. Method and apparatus for using rotational movement amount of mobile device and computer-readable recording medium for storing computer program
US20050231339A1 (en) * 2004-02-17 2005-10-20 Fuji Jukogyo Kabushiki Kaisha Outside-vehicle monitoring system
US7567687B2 (en) * 2004-02-17 2009-07-28 Fuji Jukogyo Kabushiki Kaisha Outside-vehicle monitoring system
US7515737B2 (en) * 2004-11-30 2009-04-07 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US7567688B2 (en) 2004-11-30 2009-07-28 Honda Motor Co., Ltd. Apparatus for and method of extracting image
US20060115118A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115126A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle vicinity monitoring apparatus
US20060115163A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Apparatus for and method of extracting image
US20060126896A1 (en) * 2004-11-30 2006-06-15 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060126898A1 (en) * 2004-11-30 2006-06-15 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115119A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060204037A1 (en) * 2004-11-30 2006-09-14 Honda Motor Co., Ltd. Vehicle vicinity monitoring apparatus
US7388476B2 (en) 2004-11-30 2008-06-17 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US7489805B2 (en) * 2004-11-30 2009-02-10 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115114A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US7526104B2 (en) 2004-11-30 2009-04-28 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US7545955B2 (en) * 2004-11-30 2009-06-09 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US7561719B2 (en) * 2004-11-30 2009-07-14 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20060115117A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co. Ltd. Position detecting apparatus and method of correcting data therein
US20060114320A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co. Ltd. Position detecting apparatus and method of correcting data therein
US7590263B2 (en) 2004-11-30 2009-09-15 Honda Motor Co., Ltd. Vehicle vicinity monitoring apparatus
US7599521B2 (en) 2004-11-30 2009-10-06 Honda Motor Co., Ltd. Vehicle vicinity monitoring apparatus
US7616806B2 (en) 2004-11-30 2009-11-10 Honda Motor Co., Ltd. Position detecting apparatus and method of correcting data therein
US7620237B2 (en) 2004-11-30 2009-11-17 Honda Motor Co., Ltd. Position detecting apparatus and method of correcting data therein
US20060115122A1 (en) * 2004-11-30 2006-06-01 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US7969466B2 (en) 2004-11-30 2011-06-28 Honda Motor Co., Ltd. Vehicle surroundings monitoring apparatus
US20120106786A1 (en) * 2009-05-19 2012-05-03 Toyota Jidosha Kabushiki Kaisha Object detecting device
US8897497B2 (en) * 2009-05-19 2014-11-25 Toyota Jidosha Kabushiki Kaisha Object detecting device
EP2354316A1 (en) * 2010-02-05 2011-08-10 Kässbohrer Geländefahrzeug AG Piste maintenance vehicle
US20130033600A1 (en) * 2011-08-01 2013-02-07 Hitachi, Ltd. Image Processing Device
US9165374B2 (en) * 2011-08-01 2015-10-20 Hitachi, Ltd. Image processing device that performs tracking control
US20130322691A1 (en) * 2012-06-01 2013-12-05 Ricoh Company, Ltd. Target recognition system and target recognition method executed by the target recognition system
US8977006B2 (en) * 2012-06-01 2015-03-10 Ricoh Company, Ltd. Target recognition system and target recognition method executed by the target recognition system

Also Published As

Publication number Publication date
JP2003134508A (en) 2003-05-09
US20030083790A1 (en) 2003-05-01
DE10247371B4 (en) 2007-11-22
GB2383222A (en) 2003-06-18
JP3764086B2 (en) 2006-04-05
GB2383222B (en) 2003-12-31
DE10247371A1 (en) 2003-07-24
GB0223887D0 (en) 2002-11-20

Similar Documents

Publication Publication Date Title
US7141796B2 (en) Vehicle information providing apparatus
US8144195B2 (en) Vehicle zone monitoring apparatus
US7474765B2 (en) Image recognition apparatus
US7969466B2 (en) Vehicle surroundings monitoring apparatus
US7483549B2 (en) Vehicle surroundings monitoring apparatus
US7388476B2 (en) Vehicle surroundings monitoring apparatus
US6327536B1 (en) Vehicle environment monitoring system
US7489805B2 (en) Vehicle surroundings monitoring apparatus
US6789015B2 (en) Vehicle environment monitoring system
JP4104867B2 (en) Night vision system
US20060126896A1 (en) Vehicle surroundings monitoring apparatus
US7403639B2 (en) Vehicle surroundings monitoring apparatus
US7526104B2 (en) Vehicle surroundings monitoring apparatus
US7515737B2 (en) Vehicle surroundings monitoring apparatus
JP4425852B2 (en) Vehicle periphery monitoring device
JP3916930B2 (en) Approach warning device
JP2001018738A (en) Apparatus for monitoring ambient condition of vehicle
JP3949628B2 (en) Vehicle periphery monitoring device
US7545955B2 (en) Vehicle surroundings monitoring apparatus
JP3961269B2 (en) Obstacle alarm device
JP4943403B2 (en) Vehicle periphery monitoring device
JP2004348645A (en) Infrared image recognition apparatus and alarm equipment using the same

Legal Events

Date Code Title Description
AS Assignment

Owner name: HONDA GIKEN KOGYO KABUSHIKI KAISHA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HATTORI, HIROSHI;NAGAOKA, NOBUHARU;WATANABE, MASAHITO;AND OTHERS;REEL/FRAME:013406/0350

Effective date: 20021002

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553)

Year of fee payment: 12

AS Assignment

Owner name: VEONEER SWEDEN AB, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HONDA MOTOR CO., LTD;REEL/FRAME:057336/0220

Effective date: 20210405

AS Assignment

Owner name: ARRIVER SOFTWARE AB, SWEDEN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VEONEER SWEDEN AB;REEL/FRAME:059596/0826

Effective date: 20211230