US20170041591A1 - Vehicle-Mounted Image Recognition Device - Google Patents

Vehicle-Mounted Image Recognition Device Download PDF

Info

Publication number
US20170041591A1
US20170041591A1 US15/102,282 US201415102282A US2017041591A1 US 20170041591 A1 US20170041591 A1 US 20170041591A1 US 201415102282 A US201415102282 A US 201415102282A US 2017041591 A1 US2017041591 A1 US 2017041591A1
Authority
US
United States
Prior art keywords
light source
imaging unit
unit
flickering
cycle
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/102,282
Inventor
Naoki KOROGI
Yuji Otsuka
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hitachi Astemo Ltd
Original Assignee
Hitachi Automotive Systems Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi Automotive Systems Ltd filed Critical Hitachi Automotive Systems Ltd
Assigned to Hitachi Automotive System, Ltd. reassignment Hitachi Automotive System, Ltd. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OTSUKA, YUJI, KOROGI, NAOKI
Publication of US20170041591A1 publication Critical patent/US20170041591A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/296Synchronisation thereof; Control thereof
    • H04N13/0296
    • G06K9/00825
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/141Control of illumination
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/582Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/584Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
    • H04N13/0239
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/45Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/57Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/71Circuitry for evaluating the brightness variation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/73Circuitry for compensating brightness variation in the scene by influencing the exposure time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene
    • H04N23/745Detection of flicker frequency or suppression of flicker wherein the flicker is caused by illumination, e.g. due to fluorescent tube illumination or pulsed LED illumination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • H04N5/2258
    • H04N5/23229
    • H04N5/2351
    • H04N5/2353
    • H04N5/2357

Definitions

  • the present invention relates to an in-vehicle camera, and particularly to an in-vehicle image recognition device having a function of simultaneously recognizing a plurality of targets such as a preceding vehicle, a white lane, a pedestrian, and a light source.
  • the LED used in the traffic signal and the traffic sign is flickered to be repeatedly on and off at a high speed when being turned on.
  • a cycle of the flickering is 100 Hz in a region of 50 Hz commercial power source or 120 Hz in a region of 60 Hz commercial power source. The flickering is performed at a high speed, and thus it appears to person as always turned on.
  • the emission light of the light source is not recorded in the image obtained from the camera.
  • the capturing cycle of the camera is close to an integer times the cycle of the flickering light source, the above problem continuously occurs in a plurality of frames.
  • PTL 1 discloses a technique in which a number of imaging units are provided to capture an image while deviating the exposure timings in a time shorter than a flickering cycle of the flickering light source and to observe a change in brightness of the light source in order to recognize the flickering light source.
  • an image is acquired in a cycle Tled shorter than the flickering cycle of the light source, and a recognition process is performed on each image, thereby significantly increasing a processing load.
  • a transfer amount of a bus may be significantly expanded.
  • An in-vehicle camera having an image recognition function is requested for simultaneously recognizing a number of objects such as a preceding vehicle, a pedestrian, and a white lane on a road. Therefore, when the processing load and the bus transfer amount are increased, it becomes difficult to be compatible with the other recognition processing in addition, since more imaging units are required, the cost is increased.
  • the invention has been made in view of the above problems, and an object thereof is to provide an in-vehicle imaging device which can improve a performance of recognizing the flickering light source at a low cost without causing a degradation of an image recognition application such as a preceding vehicle recognition and a white lane recognition.
  • an in-vehicle image recognition device in order to solve the above issue which processes an image captured by a pair of imaging units, wherein the same light source is captured by the pair of imaging units in a capturing cycle which is an integer times a predetermined flickering cycle, and a capturing cycle where exposure timings of the pair of imaging units are deviated by a half of the flickering cycle, and wherein, in a case where a difference between a pixel value of the light source captured by one imaging unit and a pixel value of the light source captured by the other imaging unit is equal to or more than a threshold, the light source is determined as a flickering light source.
  • the visibility of a flickering light source can be improved at a low cost without causing a degradation of the performance of an image recognition application such as a preceding vehicle recognition and a white lane recognition.
  • FIG. 1 is a block diagram for describing functions of an in-vehicle image recognition device.
  • FIG. 2 is a diagram illustrating an example of flickering of a flickering light source and exposure timings of a first imaging unit and a second imaging unit.
  • FIG. 3 is a flowchart illustrating an exemplary operation of an exposure timing adjustment unit.
  • FIG. 1 is a functional block diagram illustrating an example of an in-vehicle image recognition device in this embodiment.
  • An in-vehicle image recognition device 10 is a device which alternately captures an image at a different timing using a first imaging unit 11 a and a second imaging unit 11 b which are paired in an in-vehicle camera, detects a flickering light source from one or both of a first image 12 a obtained from the first imaging unit 11 a and a second image 12 b obtained from the second imaging unit 11 b, and performs exposure using the first imaging unit 11 a and the second imaging unit 11 b in synchronization with timing of turning on the flickering light source from next time, so that the flickering light source is recognized with high accuracy without causing degradation of the performance of an image recognition function.
  • the in-vehicle image recognition device 10 is a stereo camera which is mounted in a vehicle and, as described in the drawing, provided with the first imaging unit 11 a and the second imaging unit 11 b, a distance data generation unit 13 which generates distance information based on the first image 12 a and the second image 12 b acquired by the first imaging unit 11 a and the second imaging unit 11 b, a distance data storage unit 14 which stores the distance information indicating a distance to a capturing target, an image recognition unit 15 which performs image recognition, a light source recognition unit 16 which performs recognition of a light source, an exposure timing adjustment unit 17 which adjusts each timing of exposure, an external information reception unit 18 which receives external information, and a first exposure control unit 19 a and a second exposure control unit 19 b which control the exposure of the first imaging unit 11 a and the second imaging unit 11 b.
  • the first imaging unit 11 a and the second imaging unit 11 b are directed in the same direction and capture the same range.
  • the imaging units are separately disposed on right and left in the vicinity of a rear-view mirror of the vehicle to capture the forward side of the vehicle through a front class
  • the distance data generation unit 13 generates the distance information such as a parallax image stereoscopically obtained by the first image 12 a and the second image 12 b.
  • the image recognition unit 15 recognizes an object body such as a traffic signal, a traffic sign, a preceding vehicle, a pedestrian, and a bicycle using at least one of the first image 12 a and the second image 12 b. As a recognition method, the well-known pattern matching may be used.
  • the light source recognition unit 16 recognizes a light source from at least one of the first image 12 a and the second image 12 b, and recognizes whether the recognized light source is a flickering light source.
  • the exposure timing adjustment unit 17 adjusts the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b based on an image recognition result of the image recognition unit 15 and the external information received by the external information reception unit 18 .
  • the first exposure control unit 19 a and the second exposure control unit 19 b control the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b according to, for example, an instruction from the exposure timing adjustment unit 17 .
  • the exposure timings are adjusted to be exposure parameters optimized to the image recognition based on the information of the images acquired from the first imaging unit 11 a and the second imaging unit 11 b.
  • the external information reception unit 18 is connected to a vehicle network 20 and acquires the external information such as vehicle speed information 21 , a yaw rate 22 , GPS information 23 , time information 24 , and map data 25 using the external information reception unit 18 .
  • the vehicle image recognition device 10 When capturing the traffic signal using an LED as the flickering light source and an indicator such as the traffic sign of an electric light display type, the vehicle image recognition device 10 captures the images at exposure timings of the first imaging unit 11 a and the second imaging unit 11 b in synchronization with a flickering period of the LED.
  • capturing cycles Tc of the first imaging unit 11 a and the second imaging unit 11 b are set to be Nc times (an integer) of a flickering cycle Tl, and the exposure timing of the first imaging unit 11 a and the exposure timing of the second imaging unit 11 b are set to be deviated by Ns (an integer)+0.5 times the flickering cycle, so that the first image and the second image deviated by a half cycle with respect to the flickering cycle are acquired.
  • the light source in the ON state is recorded in at least any one of the first image 12 a and the second image 12 b.
  • the light source recognition unit 16 performs a recognition process in which the light source is extracted from the first image 12 a and the second image 12 b. Then, in a case where there is a difference equal to or more than a predetermined value between pixel values of the light source extracted from the first image 12 a and the second image 12 b, it is determined that the subject light source is a flickering light source.
  • the exposure timing adjustment unit 17 changes the exposure timing such that the exposure timing of one imaging unit is changed to be matched to the exposure timing of the other imaging unit in which the flickering light source is brightly captured among the first imaging unit 11 a and the second imaging unit 11 b, and adjusts the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b to be synchronized.
  • the distance data generation unit generates distance data indicating a distance from the vehicle to the light source based on the first image and the second image captured by the first imaging unit and the second imaging unit.
  • FIG. 2 is a diagram illustrating an example of flickering of the flickering light source and exposure timings of the first imaging unit and the second imaging unit.
  • the exposure timing adjustment unit 17 sets the capturing cycles Tc of the first imaging unit 11 a and the second imaging unit 11 b to be Nc (an integer) times the flickering cycle Tl. Therefore, even the flickering light source can be captured in the same brightness in every frame, and the brightness of the flickering light source illuminated on each frame can be made constant.
  • the flickering light source at the exposure timing of the other imaging unit necessarily comes to be bright since the exposure timing of the other imaging unit is deviated by the half cycle with respect to the flickering cycle Tl.
  • the light source recognition unit 16 can recognize that the light source is flickering by recognizing that there is a light source in the image obtained by an imaging unit and there is no light source in the image obtained by the other imaging unit. Therefore, it is possible to reliably capture the flickering light source from next time by causing the first imaging unit 11 a and the second imaging unit 11 b to perform exposure in synchronization with timing when the flickering light source is bright. Therefore, the flickering light source can be stereoscopically viewed, and accurate distance information up to the flickering light source can be obtained.
  • the flickering light source can be captured with a sufficient pixel value, and can be recognized similarly to an unflickering light source (a light source always turned on).
  • the light source can be recognized in synchronization with the exposure timing of any one of the first imaging unit 11 a and the second imaging unit 11 b. Therefore, the flickering light source can be recognized and the distance information can be obtained by performing the stereoscopic viewing in synchronization with the exposure timing of any one of the imaging units. In other words, the flickering light source can be recognized similarly to a case where the light always turned on is recognized.
  • any method such as a template matching and a classification using color information may be used as long as a feature can be extracted from the image, and a well-known method may be used.
  • FIG. 3 is a flowchart for describing an exemplary method of setting the exposure timing using the exposure timing adjustment unit.
  • the exposure timing adjustment unit 17 determines whether the stereoscopic distance data is generated based on the recognition result of the image recognition unit 15 and the information acquired from the external information reception unit 18 .
  • Step S 1 recognition result information is received from the image recognition unit 15 and the light source recognition unit 16 .
  • Step S 2 the external information such as the vehicle speed information 21 , the yaw rate 22 , the GPS information 23 , the time information 24 , and the map data 25 linked to the vehicle network 20 are acquired from the external information reception unit 18 .
  • Step S 3 it is determined whether the flickering cycle Tl of the flickering light source is 1/50 second. Since the capturing cycle Tc and a capture deviation time are determined based on the flickering cycle of the flickering light source, the flickering cycle of the flickering light source is necessarily determined.
  • Step S 3 in a case where it is determined that the flickering cycle of the flickering light source is 1/50 second, the flickering cycle of the flickering light source is set to 1/50 second in Step S 4 , and if not (that is, it is determined that the flickering cycle is not 1/50 second), the flickering cycle of the flickering power source is set to 1/60 second in Step S 5 .
  • the flickering cycle is set to any one of 1/50 and 1/60 second. For example, in a case where the pixel value of the light source is almost the same value in the image captured by the same imaging unit, it can be determined that the capturing cycle is an integer times an intended flickering cycle and the flickering cycle is correctly set.
  • the flickering cycle of the light source is different from the current setting, and thus the setting of the flickering cycle is switched to the other value.
  • the flickering cycle from next time is set to the cycle at the previous time.
  • the setting is switched.
  • the flickering cycle of the flickering light source may be determined based on a power source frequency of the flickering light source. Since the frequency of the commercial power source is determined depending on regions, the flickering cycle may be determined according to GPS information or a running location of a subject vehicle received in cooperation with a car navigation system. For example, in the case of Japan, the power source frequency in the east region of Japan is 50 Hz, and the power source frequency in the west region of Japan is 60 Hz. Therefore, the power source frequency of the flickering power source can be recognized by specifying the current location of the vehicle based on the external information such as the GPS information 23 and the map data 25 .
  • the flickering cycle of the flickering light source can be set to 1/60 second.
  • Step S 6 it is determined whether a light emission timing of the flickering light source is estimated.
  • the first imaging unit 11 a and the second imaging unit 11 b are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S 8 .
  • the flickering light source can reliably perform the capturing using both the first imaging unit 11 a and the second imaging unit 11 b. Further, the distance information up to the light source can be stereoscopically acquired.
  • Step S 7 it is determined whether the distance information is necessary in Step S 7 . While the acquirement of the distance information using the stereo camera is a feature, the distance information may be not necessary for all the frames. Therefore, the stereoscopic viewing (the imaging units are caused to perform the capturing at the same exposure timing) or the deviating of the exposure timings is selected according to a situation based on the recognition result acquired in Step S 1 and the external information acquired in Step S 2 . In a case where it is determined that the distance information is necessary, the first imaging unit and the second imaging unit are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S 8 . In a case where it is determined that the distance information is not necessary, the capturing is performed by deviating the exposure timings of the first imaging unit and the second imaging unit by a predetermined time in Step S 9 .
  • the capturing is performed at the same exposure timing based on the other applications with priority higher than that of the light emission timing of the flickering light source.
  • the preceding vehicle and the pedestrian are recognized from the image, and the result is acquired from the image recognition unit 15 .
  • the preceding vehicle is within a predetermined distance, or in a case where the pedestrian or the bicycle is near the advancing path of the subject vehicle, it is determined that the distance information is necessary for avoiding a risk.
  • a three-dimensional object can be effectively detected in the stereoscope viewing. Therefore, a counter is provided in the device, and the capturing may be performed alternately using the stereoscopic viewing and the deviating of the exposure timings in a predetermined cycle even in a case where there is no change in an ambient environment.
  • the exposure timings may be deviated only when the subject vehicle comes close to a predetermined distance or more. It is desirable that a parameter related to the exposure be optimal to recognize the preceding vehicle and the pedestrian. However, for example, in a case where it is determined that the light source is recognized with priority, a shutter opening time may be lengthened.
  • the shutter opening time and a pixel transfer amount of the imaging unit are not changed, so that the performance of the image recognition application is not damaged.
  • a processing load is not increased.
  • the recognition performance of the flickering light source can be improved at a low cost.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Studio Devices (AREA)
  • Traffic Control Systems (AREA)

Abstract

This vehicle-mounted image recognition device achieves improved performance in flashing light source recognition at low cost and without loss of performance in image recognition applications such as preceding vehicle recognition and white line recognition. At staggered exposure timing, imaging is performed alternately with a first imaging unit and a second imaging unit, which form a pair in a vehicle-mounted camera. A flashing light source is detected from one or both of the first image obtained from the first imaging unit and a second image obtained from the second imaging unit, and the following times, exposure of the first imaging unit and the second imaging unit is matched to when the flashing light source is on. Thereby, flashing light sources are recognized with high accuracy without loss of performance of the image recognition function.

Description

    TECHNICAL FIELD
  • The present invention relates to an in-vehicle camera, and particularly to an in-vehicle image recognition device having a function of simultaneously recognizing a plurality of targets such as a preceding vehicle, a white lane, a pedestrian, and a light source.
  • BACKGROUND ART
  • In recent years, light sources used in a traffic signal and a traffic sign of an electric light display type are replaced from the conventional incandescent lamps to LEDs. The LED has a merit on less power consumption and long life span compared to the incandescent lamp. In addition, there is no pseudo lighting considered as emission caused by the reflection of sunlight, so that the visibility is high and also the stability is improved.
  • The LED used in the traffic signal and the traffic sign is flickered to be repeatedly on and off at a high speed when being turned on. A cycle of the flickering is 100 Hz in a region of 50 Hz commercial power source or 120 Hz in a region of 60 Hz commercial power source. The flickering is performed at a high speed, and thus it appears to person as always turned on.
  • In a case where the exposure timing of an in-vehicle camera is matched to the OFF state of the flickering light source, the emission light of the light source is not recorded in the image obtained from the camera. In addition, in a case where the capturing cycle of the camera is close to an integer times the cycle of the flickering light source, the above problem continuously occurs in a plurality of frames.
  • In order to solve the above problem, for example, PTL 1 discloses a technique in which a number of imaging units are provided to capture an image while deviating the exposure timings in a time shorter than a flickering cycle of the flickering light source and to observe a change in brightness of the light source in order to recognize the flickering light source.
  • CITATION LIST Patent Literature
  • PTL 1: Japanese Patent Application Laid-Open No. 2007-286943
  • SUMMARY OF INVENTION Technical Problem
  • However, in the conventional technique described above, an image is acquired in a cycle Tled shorter than the flickering cycle of the light source, and a recognition process is performed on each image, thereby significantly increasing a processing load. In addition, since a number of images are transferred from an imaging unit to a recognition unit, a transfer amount of a bus may be significantly expanded. An in-vehicle camera having an image recognition function is requested for simultaneously recognizing a number of objects such as a preceding vehicle, a pedestrian, and a white lane on a road. Therefore, when the processing load and the bus transfer amount are increased, it becomes difficult to be compatible with the other recognition processing in addition, since more imaging units are required, the cost is increased.
  • The invention has been made in view of the above problems, and an object thereof is to provide an in-vehicle imaging device which can improve a performance of recognizing the flickering light source at a low cost without causing a degradation of an image recognition application such as a preceding vehicle recognition and a white lane recognition.
  • Solution to Problem
  • There is provided an in-vehicle image recognition device according to the present invention in order to solve the above issue which processes an image captured by a pair of imaging units, wherein the same light source is captured by the pair of imaging units in a capturing cycle which is an integer times a predetermined flickering cycle, and a capturing cycle where exposure timings of the pair of imaging units are deviated by a half of the flickering cycle, and wherein, in a case where a difference between a pixel value of the light source captured by one imaging unit and a pixel value of the light source captured by the other imaging unit is equal to or more than a threshold, the light source is determined as a flickering light source.
  • Advantageous Effects of Invention
  • According to the invention, the visibility of a flickering light source can be improved at a low cost without causing a degradation of the performance of an image recognition application such as a preceding vehicle recognition and a white lane recognition. Further, other objects, configurations, and effects will become clear through the explanation of the following embodiments.
  • BRIEF DESCRIPTION OF DRAWINGS
  • [FIG. 1] FIG. 1 is a block diagram for describing functions of an in-vehicle image recognition device.
  • [FIG. 2] FIG. 2 is a diagram illustrating an example of flickering of a flickering light source and exposure timings of a first imaging unit and a second imaging unit.
  • [FIG. 3] FIG. 3 is a flowchart illustrating an exemplary operation of an exposure timing adjustment unit.
  • DESCRIPTION OF EMBODIMENTS
  • Next, embodiments of the invention will be described below using the drawings. FIG. 1 is a functional block diagram illustrating an example of an in-vehicle image recognition device in this embodiment.
  • An in-vehicle image recognition device 10 is a device which alternately captures an image at a different timing using a first imaging unit 11 a and a second imaging unit 11 b which are paired in an in-vehicle camera, detects a flickering light source from one or both of a first image 12 a obtained from the first imaging unit 11 a and a second image 12 b obtained from the second imaging unit 11 b, and performs exposure using the first imaging unit 11 a and the second imaging unit 11 b in synchronization with timing of turning on the flickering light source from next time, so that the flickering light source is recognized with high accuracy without causing degradation of the performance of an image recognition function.
  • The in-vehicle image recognition device 10 is a stereo camera which is mounted in a vehicle and, as described in the drawing, provided with the first imaging unit 11 a and the second imaging unit 11 b, a distance data generation unit 13 which generates distance information based on the first image 12 a and the second image 12 b acquired by the first imaging unit 11 a and the second imaging unit 11 b, a distance data storage unit 14 which stores the distance information indicating a distance to a capturing target, an image recognition unit 15 which performs image recognition, a light source recognition unit 16 which performs recognition of a light source, an exposure timing adjustment unit 17 which adjusts each timing of exposure, an external information reception unit 18 which receives external information, and a first exposure control unit 19 a and a second exposure control unit 19 b which control the exposure of the first imaging unit 11 a and the second imaging unit 11 b.
  • The first imaging unit 11 a and the second imaging unit 11 b are directed in the same direction and capture the same range. For example, the imaging units are separately disposed on right and left in the vicinity of a rear-view mirror of the vehicle to capture the forward side of the vehicle through a front class
  • The distance data generation unit 13 generates the distance information such as a parallax image stereoscopically obtained by the first image 12 a and the second image 12 b. The image recognition unit 15 recognizes an object body such as a traffic signal, a traffic sign, a preceding vehicle, a pedestrian, and a bicycle using at least one of the first image 12 a and the second image 12 b. As a recognition method, the well-known pattern matching may be used.
  • The light source recognition unit 16 recognizes a light source from at least one of the first image 12 a and the second image 12 b, and recognizes whether the recognized light source is a flickering light source. The exposure timing adjustment unit 17 adjusts the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b based on an image recognition result of the image recognition unit 15 and the external information received by the external information reception unit 18. The first exposure control unit 19 a and the second exposure control unit 19 b control the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b according to, for example, an instruction from the exposure timing adjustment unit 17. In addition, the exposure timings are adjusted to be exposure parameters optimized to the image recognition based on the information of the images acquired from the first imaging unit 11 a and the second imaging unit 11 b. The external information reception unit 18 is connected to a vehicle network 20 and acquires the external information such as vehicle speed information 21, a yaw rate 22, GPS information 23, time information 24, and map data 25 using the external information reception unit 18.
  • When capturing the traffic signal using an LED as the flickering light source and an indicator such as the traffic sign of an electric light display type, the vehicle image recognition device 10 captures the images at exposure timings of the first imaging unit 11 a and the second imaging unit 11 b in synchronization with a flickering period of the LED.
  • Specifically, capturing cycles Tc of the first imaging unit 11 a and the second imaging unit 11 b are set to be Nc times (an integer) of a flickering cycle Tl, and the exposure timing of the first imaging unit 11 a and the exposure timing of the second imaging unit 11 b are set to be deviated by Ns (an integer)+0.5 times the flickering cycle, so that the first image and the second image deviated by a half cycle with respect to the flickering cycle are acquired.
  • Therefore, the light source in the ON state is recorded in at least any one of the first image 12 a and the second image 12 b.
  • The light source recognition unit 16 performs a recognition process in which the light source is extracted from the first image 12 a and the second image 12 b. Then, in a case where there is a difference equal to or more than a predetermined value between pixel values of the light source extracted from the first image 12 a and the second image 12 b, it is determined that the subject light source is a flickering light source. In a case where the light source recognition unit 16 determines that the light source is a flickering light source, the exposure timing adjustment unit 17 changes the exposure timing such that the exposure timing of one imaging unit is changed to be matched to the exposure timing of the other imaging unit in which the flickering light source is brightly captured among the first imaging unit 11 a and the second imaging unit 11 b, and adjusts the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b to be synchronized. The distance data generation unit generates distance data indicating a distance from the vehicle to the light source based on the first image and the second image captured by the first imaging unit and the second imaging unit.
  • Next, an example of adjusting the exposure timing of the invention will be described using FIG. 2. FIG. 2 is a diagram illustrating an example of flickering of the flickering light source and exposure timings of the first imaging unit and the second imaging unit. As illustrated in FIG. 2, in a case where the brightness of the LED is increased or decreased in the flickering cycle Tl and the flickering light source emits the light, the exposure timing adjustment unit 17 sets the capturing cycles Tc of the first imaging unit 11 a and the second imaging unit 11 b to be Nc (an integer) times the flickering cycle Tl. Therefore, even the flickering light source can be captured in the same brightness in every frame, and the brightness of the flickering light source illuminated on each frame can be made constant.
  • Then, the exposure timing adjustment unit 17 sets a deviation time between the exposure timing of the first imaging unit and the exposure timing of the second imaging unit to be Ns (an integer)+0.5 times the flickering cycle Tl (Deviation Time=Tl×(Ns+0.5)). Therefore, the first imaging unit 11 a and the second imaging unit 11 b can acquire brightnesses deviated by the half flickering cycle Tl of the LED. Herein, in a case where the exposure timing of one of the first imaging unit 11 a and the second imaging unit 11 b is a timing when the flickering light source is dark, the flickering light source at the exposure timing of the other imaging unit necessarily comes to be bright since the exposure timing of the other imaging unit is deviated by the half cycle with respect to the flickering cycle Tl.
  • The light source recognition unit 16 can recognize that the light source is flickering by recognizing that there is a light source in the image obtained by an imaging unit and there is no light source in the image obtained by the other imaging unit. Therefore, it is possible to reliably capture the flickering light source from next time by causing the first imaging unit 11 a and the second imaging unit 11 b to perform exposure in synchronization with timing when the flickering light source is bright. Therefore, the flickering light source can be stereoscopically viewed, and accurate distance information up to the flickering light source can be obtained.
  • Even at a timing when the flickering light source is captured in the same brightness by the first imaging unit 11 a and the second imaging unit 11 b, the traffic signal and the indicator such as the traffic sign using the LED are sufficiently bright compared to the preceding vehicle and a road surface in order to secure visibility. Therefore, even in a case where an optimal exposure is performed in order to recognize the preceding vehicle and the road surface, the flickering light source can be captured with a sufficient pixel value, and can be recognized similarly to an unflickering light source (a light source always turned on).
  • In this case, the light source can be recognized in synchronization with the exposure timing of any one of the first imaging unit 11 a and the second imaging unit 11 b. Therefore, the flickering light source can be recognized and the distance information can be obtained by performing the stereoscopic viewing in synchronization with the exposure timing of any one of the imaging units. In other words, the flickering light source can be recognized similarly to a case where the light always turned on is recognized.
  • Further, as a method of recognizing the light source from an image, any method such as a template matching and a classification using color information may be used as long as a feature can be extracted from the image, and a well-known method may be used.
  • FIG. 3 is a flowchart for describing an exemplary method of setting the exposure timing using the exposure timing adjustment unit. The exposure timing adjustment unit 17 determines whether the stereoscopic distance data is generated based on the recognition result of the image recognition unit 15 and the information acquired from the external information reception unit 18.
  • First, in Step S1, recognition result information is received from the image recognition unit 15 and the light source recognition unit 16. Next, in Step S2, the external information such as the vehicle speed information 21, the yaw rate 22, the GPS information 23, the time information 24, and the map data 25 linked to the vehicle network 20 are acquired from the external information reception unit 18.
  • Then, in Step S3, it is determined whether the flickering cycle Tl of the flickering light source is 1/50 second. Since the capturing cycle Tc and a capture deviation time are determined based on the flickering cycle of the flickering light source, the flickering cycle of the flickering light source is necessarily determined.
  • In Step S3, in a case where it is determined that the flickering cycle of the flickering light source is 1/50 second, the flickering cycle of the flickering light source is set to 1/50 second in Step S4, and if not (that is, it is determined that the flickering cycle is not 1/50 second), the flickering cycle of the flickering power source is set to 1/60 second in Step S5.
  • At the first time, since the flickering cycle is not clear, the flickering cycle is set to any one of 1/50 and 1/60 second. For example, in a case where the pixel value of the light source is almost the same value in the image captured by the same imaging unit, it can be determined that the capturing cycle is an integer times an intended flickering cycle and the flickering cycle is correctly set.
  • On the other hand, in a case where the pixel value of the light source is changed by stages in the image captured by the same imaging unit, it is determined that the flickering cycle of the light source is different from the current setting, and thus the setting of the flickering cycle is switched to the other value. The flickering cycle from next time is set to the cycle at the previous time. In a case where it is determined that the flickering cycle is different on the same basis, the setting is switched.
  • The flickering cycle of the flickering light source may be determined based on a power source frequency of the flickering light source. Since the frequency of the commercial power source is determined depending on regions, the flickering cycle may be determined according to GPS information or a running location of a subject vehicle received in cooperation with a car navigation system. For example, in the case of Japan, the power source frequency in the east region of Japan is 50 Hz, and the power source frequency in the west region of Japan is 60 Hz. Therefore, the power source frequency of the flickering power source can be recognized by specifying the current location of the vehicle based on the external information such as the GPS information 23 and the map data 25. For example, in a case where the power source frequency is 50 Hz, and in a case where the flickering cycle of the flickering light source is 1/50 second and the power source frequency is 60 Hz, the flickering cycle of the flickering light source can be set to 1/60 second.
  • Next, in Step S6, it is determined whether a light emission timing of the flickering light source is estimated. In a case where the light emission timing of the flickering light source is estimated, the first imaging unit 11 a and the second imaging unit 11 b are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S8. In this case, the flickering light source can reliably perform the capturing using both the first imaging unit 11 a and the second imaging unit 11 b. Further, the distance information up to the light source can be stereoscopically acquired.
  • On the other hand, in a case where the light emission timing of the flickering light source is not estimated yet, it is determined whether the distance information is necessary in Step S7. While the acquirement of the distance information using the stereo camera is a feature, the distance information may be not necessary for all the frames. Therefore, the stereoscopic viewing (the imaging units are caused to perform the capturing at the same exposure timing) or the deviating of the exposure timings is selected according to a situation based on the recognition result acquired in Step S1 and the external information acquired in Step S2. In a case where it is determined that the distance information is necessary, the first imaging unit and the second imaging unit are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S8. In a case where it is determined that the distance information is not necessary, the capturing is performed by deviating the exposure timings of the first imaging unit and the second imaging unit by a predetermined time in Step S9.
  • For example, in a case where it is a situation that the distance information from the stereo camera is necessary for the other applications such as a pedestrian detection and a preceding vehicle detection, the capturing is performed at the same exposure timing based on the other applications with priority higher than that of the light emission timing of the flickering light source.
  • In the in-vehicle image recognition device 10, the preceding vehicle and the pedestrian are recognized from the image, and the result is acquired from the image recognition unit 15. In a case where the preceding vehicle is within a predetermined distance, or in a case where the pedestrian or the bicycle is near the advancing path of the subject vehicle, it is determined that the distance information is necessary for avoiding a risk. In addition, a three-dimensional object can be effectively detected in the stereoscope viewing. Therefore, a counter is provided in the device, and the capturing may be performed alternately using the stereoscopic viewing and the deviating of the exposure timings in a predetermined cycle even in a case where there is no change in an ambient environment.
  • In a case where the traffic signal or the traffic sign of an electric light display type can be acquired from the map data 25, the exposure timings may be deviated only when the subject vehicle comes close to a predetermined distance or more. It is desirable that a parameter related to the exposure be optimal to recognize the preceding vehicle and the pedestrian. However, for example, in a case where it is determined that the light source is recognized with priority, a shutter opening time may be lengthened.
  • According to the in-vehicle image recognition device 10 described above, the shutter opening time and a pixel transfer amount of the imaging unit are not changed, so that the performance of the image recognition application is not damaged. In addition, since the extraction process of the light source is performed as in the conventional process, a processing load is not increased. Further, there is no increase in hardware, the recognition performance of the flickering light source can be improved at a low cost.
  • Hitherto, the description has been made about embodiments of the invention, but the invention is not limited to the embodiments. Various changes in design can be made within a scope not departing from the spirit of the invention described in the accompanying claims. For example, the embodiments are described in a clearly understandable way for the invention, and thus the invention is not necessarily to include all the configurations described above. In addition, some configurations of a certain embodiment may be replaced with the configurations of another embodiment, and the configuration of the other embodiment may also be added to that of a certain embodiment. Furthermore, various additions, omissions, and substitutions may be partially made in some of the configurations of the respective embodiments.
  • REFERENCE SIGNS LIST
    • 10 in-vehicle image recognition device
    • 11 a first imaging unit
    • 11 b second imaging unit
    • 12 a first image
    • 12 b second image
    • 13 distance data generation unit
    • 14 distance data
    • 15 image recognition unit
    • 16 light source recognition unit
    • 17 exposure timing adjustment unit
    • 18 external information reception unit
    • 19 a first exposure control unit
    • 19 b second exposure control unit
    • 20 vehicle network
    • 21 vehicle speed information
    • 22 yaw rate
    • 23 GPS information
    • 24 time information
    • 25 map data

Claims (8)

1. An in-vehicle image recognition device which processes an image captured by a pair of imaging units,
wherein the same light source is captured by the pair of imaging units in a capturing cycle which is an integer times a predetermined flickering cycle, and a capturing cycle where exposure timings of the pair of imaging units are deviated by a half of the flickering cycle, and
wherein, in a case where a difference between a pixel value of the light source captured by one imaging unit and a pixel value of the light source captured by the other imaging unit is equal to or more than a threshold, the light source is determined as a flickering light source.
2. The in-vehicle image recognition device according to claim 1,
wherein, in a case where the light source is determined as the flickering light source, the exposure timing of the imaging unit is matched to the exposure timing of the other imaging unit of which the captured flickering light source is bright in the pair of imaging units, and distance data up to the light source is generated based on a pair of images captured by the pair of imaging units at the same exposure timing.
3. The in-vehicle image recognition device according to claim 2,
wherein two types of flickering cycles are set in advance,
wherein, in a case where the pixel value of the light source is constant in the image captured by the same imaging unit, it is determined that the setting of the flickering cycle is correct, and
wherein, in a case where the pixel value of the light source is changed by stages, it is determined that the setting of the flickering cycle is wrong, and the setting of the flickering cycle is switched to the other setting.
4. The in-vehicle image recognition device according to claim 1, comprising:
a first imaging unit;
a second imaging unit which is disposed in parallel to the first imaging unit;
a distance data generation unit which generates distance data based on images acquired from the first imaging unit and the second imaging unit; and
a light source recognition unit which recognizes the light source based on the images acquired from the first imaging unit and the second imaging unit,
wherein the first imaging unit and the second imaging unit perform capturing in a capturing cycle based on a flickering cycle of the recognized flickering light source and in a capturing cycle obtained by deviating exposure timings of the first imaging unit and the second imaging unit by a half of the flickering cycle, and
wherein the light source recognition unit determines whether the light source is a flickering light source based on an image captured by the first imaging unit and an image captured by the second imaging unit.
5. The in-vehicle image recognition device according to claim 4, comprising:
an image recognition unit which performs image recognition based on the images acquired from the first imaging unit and the second imaging unit and the distance data;
an external information reception unit which receives information containing at least one of a vehicle speed, a yaw rate, GPS information, time information, and map data; and
an exposure timing adjustment unit which adjusts the exposure timing based on a recognition result of the light source recognition unit, a recognition result of the image recognition unit, and the information of the external information reception unit.
6. The in-vehicle image recognition device according to claim 5, comprising:
a first exposure control unit and a second exposure adjustment unit which adjust the exposure timing to be an exposure parameter optimized to the image recognition based on the information of the images acquired by the first imaging unit and the second imaging unit.
7. The in-vehicle image recognition device according to claim 5,
wherein the exposure timing adjustment unit estimates the flickering cycle of the flickering light source and adjusts the exposure timing based on the recognition result of the light source recognition unit.
8. The in-vehicle image recognition device according to claim 5,
wherein the exposure timing adjustment unit determines whether the distance data is stereoscopically generated based on the recognition result of the image recognition unit and the information acquired from the external information reception unit.
US15/102,282 2013-12-25 2014-11-12 Vehicle-Mounted Image Recognition Device Abandoned US20170041591A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2013-267292 2013-12-25
JP2013267292 2013-12-25
PCT/JP2014/079902 WO2015098325A1 (en) 2013-12-25 2014-11-12 Vehicle-mounted image recognition device

Publications (1)

Publication Number Publication Date
US20170041591A1 true US20170041591A1 (en) 2017-02-09

Family

ID=53478202

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/102,282 Abandoned US20170041591A1 (en) 2013-12-25 2014-11-12 Vehicle-Mounted Image Recognition Device

Country Status (4)

Country Link
US (1) US20170041591A1 (en)
EP (1) EP3089442B1 (en)
JP (1) JP6325000B2 (en)
WO (1) WO2015098325A1 (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160031371A1 (en) * 2014-07-29 2016-02-04 Denso Corporation In-vehicle apparatus
US20160371552A1 (en) * 2014-03-10 2016-12-22 Nissan Motor Co., Ltd. Traffic Light Detecting Device and Traffic Light Detecting Method
US20170017850A1 (en) * 2014-03-10 2017-01-19 Nissan Motor Co., Ltd. Traffic Light Detecting Device and Traffic Light Detecting Method
US20170339326A1 (en) * 2016-05-20 2017-11-23 Gentex Corporation Imager with active exposure correction
JP2018064257A (en) * 2016-10-15 2018-04-19 キヤノン株式会社 Imaging System
US10055655B2 (en) * 2015-04-08 2018-08-21 Nissan Motor Co., Ltd. Traffic light detection device and traffic light detection method
WO2019050654A1 (en) * 2017-09-07 2019-03-14 Qualcomm Incorporated Image capture setting determination in devices having access to multiple cameras
CN111291620A (en) * 2020-01-14 2020-06-16 北京小马智行科技有限公司 Method, device and system for identifying light source
US10715738B1 (en) * 2019-04-30 2020-07-14 Axon Enterprise, Inc. Asymmetrical license plate reading (ALPR) camera system
US10863106B1 (en) * 2019-10-21 2020-12-08 GM Global Technology Operations LLC Systems and methods for LED flickering and banding detection
WO2021043892A1 (en) * 2019-09-03 2021-03-11 Jaguar Land Rover Limited Method and system for mitigating image flicker from strobed lighting systems
US11030472B2 (en) 2019-04-30 2021-06-08 Axon Enterprise, Inc. Asymmetrical license plate reading (ALPR) camera system
US11343441B2 (en) * 2017-08-21 2022-05-24 Sony Semiconductor Solutions Corporation Imaging device and apparatus
US11483505B2 (en) * 2017-11-06 2022-10-25 SK Hynix Inc. Image synchronization device and image information generation apparatus including the same
US11532170B2 (en) 2019-04-30 2022-12-20 Axon Enterprise, Inc. License plate reading system with enhancements
EP4086874A4 (en) * 2020-01-23 2023-05-31 Huawei Technologies Co., Ltd. Control method and apparatus for intelligent vehicle, and related device

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109496187B (en) * 2016-08-08 2022-07-26 金泰克斯公司 System and method for processing video data to detect and eliminate flicker light source through dynamic exposure control
WO2018225516A1 (en) * 2017-06-07 2018-12-13 日立オートモティブシステムズ株式会社 Image processing device

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030133018A1 (en) * 2002-01-16 2003-07-17 Ted Ziemkowski System for near-simultaneous capture of multiple camera images
US20040012692A1 (en) * 2002-03-25 2004-01-22 Shinichi Arazaki Flicker detection apparatus, a flicker correction apparatus, an image-pickup apparatus, a flicker detection program and a flicker correction program
JP2007286943A (en) * 2006-04-18 2007-11-01 Fujifilm Corp Signal light detection apparatus
JP2008134844A (en) * 2006-11-28 2008-06-12 Alpine Electronics Inc Drive recorder, and image acquisition timing control method therefor
WO2008108025A1 (en) * 2007-03-05 2008-09-12 Nec Electronics Corporation Imaging apparatus and flicker detection method
JP2009065586A (en) * 2007-09-10 2009-03-26 Hitachi Ltd On-vehicle camera
US20150035949A1 (en) * 2013-07-31 2015-02-05 Morpho Method for synchronising several cameras with each other in a photographic system, in particular a stereoscopic photographic system, and photographic system for implementing said method
US20150138324A1 (en) * 2012-05-16 2015-05-21 Denso Corporation Apparatus for detecting vehicle light and method thereof

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011234318A (en) * 2010-04-30 2011-11-17 Yamaha Corp Imaging device

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030133018A1 (en) * 2002-01-16 2003-07-17 Ted Ziemkowski System for near-simultaneous capture of multiple camera images
US20040012692A1 (en) * 2002-03-25 2004-01-22 Shinichi Arazaki Flicker detection apparatus, a flicker correction apparatus, an image-pickup apparatus, a flicker detection program and a flicker correction program
JP2007286943A (en) * 2006-04-18 2007-11-01 Fujifilm Corp Signal light detection apparatus
JP2008134844A (en) * 2006-11-28 2008-06-12 Alpine Electronics Inc Drive recorder, and image acquisition timing control method therefor
WO2008108025A1 (en) * 2007-03-05 2008-09-12 Nec Electronics Corporation Imaging apparatus and flicker detection method
JP2009065586A (en) * 2007-09-10 2009-03-26 Hitachi Ltd On-vehicle camera
US20150138324A1 (en) * 2012-05-16 2015-05-21 Denso Corporation Apparatus for detecting vehicle light and method thereof
US20150035949A1 (en) * 2013-07-31 2015-02-05 Morpho Method for synchronising several cameras with each other in a photographic system, in particular a stereoscopic photographic system, and photographic system for implementing said method
FR3009469A1 (en) * 2013-07-31 2015-02-06 Morpho METHOD OF SYNCHRONIZING MULTIPLE CAMERAS OF A SHOOTING SYSTEM, IN PARTICULAR A STEREOSCOPIC VIEWING SYSTEM AND VIEWING SYSTEM FOR IMPLEMENTING SAID METHOD

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160371552A1 (en) * 2014-03-10 2016-12-22 Nissan Motor Co., Ltd. Traffic Light Detecting Device and Traffic Light Detecting Method
US20170017850A1 (en) * 2014-03-10 2017-01-19 Nissan Motor Co., Ltd. Traffic Light Detecting Device and Traffic Light Detecting Method
US9679208B2 (en) * 2014-03-10 2017-06-13 Nissan Motor Co., Ltd. Traffic light detecting device and traffic light detecting method
US9679207B2 (en) * 2014-03-10 2017-06-13 Nissan Motor Co., Ltd. Traffic light detecting device and traffic light detecting method
US20160031371A1 (en) * 2014-07-29 2016-02-04 Denso Corporation In-vehicle apparatus
US10055655B2 (en) * 2015-04-08 2018-08-21 Nissan Motor Co., Ltd. Traffic light detection device and traffic light detection method
US20170339326A1 (en) * 2016-05-20 2017-11-23 Gentex Corporation Imager with active exposure correction
US10154200B2 (en) * 2016-05-20 2018-12-11 Gentex Corporation Imager with active exposure correction
JP2018064257A (en) * 2016-10-15 2018-04-19 キヤノン株式会社 Imaging System
US11343441B2 (en) * 2017-08-21 2022-05-24 Sony Semiconductor Solutions Corporation Imaging device and apparatus
WO2019050654A1 (en) * 2017-09-07 2019-03-14 Qualcomm Incorporated Image capture setting determination in devices having access to multiple cameras
US11483505B2 (en) * 2017-11-06 2022-10-25 SK Hynix Inc. Image synchronization device and image information generation apparatus including the same
US11030472B2 (en) 2019-04-30 2021-06-08 Axon Enterprise, Inc. Asymmetrical license plate reading (ALPR) camera system
US10715738B1 (en) * 2019-04-30 2020-07-14 Axon Enterprise, Inc. Asymmetrical license plate reading (ALPR) camera system
US11532170B2 (en) 2019-04-30 2022-12-20 Axon Enterprise, Inc. License plate reading system with enhancements
US11682219B2 (en) 2019-04-30 2023-06-20 Axon Enterprise, Inc. Asymmetrical license plate reading (ALPR) camera system
US11881039B2 (en) 2019-04-30 2024-01-23 Axon Enterprise, Inc. License plate reading system with enhancements
WO2021043892A1 (en) * 2019-09-03 2021-03-11 Jaguar Land Rover Limited Method and system for mitigating image flicker from strobed lighting systems
CN112766030A (en) * 2019-10-21 2021-05-07 通用汽车环球科技运作有限责任公司 System and method for LED flicker and strip detection
US10863106B1 (en) * 2019-10-21 2020-12-08 GM Global Technology Operations LLC Systems and methods for LED flickering and banding detection
CN111291620A (en) * 2020-01-14 2020-06-16 北京小马智行科技有限公司 Method, device and system for identifying light source
EP4086874A4 (en) * 2020-01-23 2023-05-31 Huawei Technologies Co., Ltd. Control method and apparatus for intelligent vehicle, and related device

Also Published As

Publication number Publication date
WO2015098325A1 (en) 2015-07-02
EP3089442A1 (en) 2016-11-02
JP6325000B2 (en) 2018-05-16
JPWO2015098325A1 (en) 2017-03-23
EP3089442B1 (en) 2022-01-05
EP3089442A4 (en) 2017-08-30

Similar Documents

Publication Publication Date Title
EP3089442B1 (en) Vehicle-mounted image recognition device
CN103213540B (en) Vehicle driving environment recognition apparatus
CN102047166B (en) Device, camera, and method for generating images of the vicinity of a motor vehicle
US11343441B2 (en) Imaging device and apparatus
JP5863536B2 (en) Outside monitoring device
US20140293055A1 (en) Image processing apparatus
JP5772714B2 (en) Light detection device and vehicle control system
US10055655B2 (en) Traffic light detection device and traffic light detection method
CN107431747B (en) Image pickup apparatus
JP6447289B2 (en) Imaging apparatus, imaging method, program, vehicle control system, and vehicle
US20170116488A1 (en) Method for identifying an incoming vehicle and corresponding system
EP3637758B1 (en) Image processing device
JP6259335B2 (en) Outside environment recognition device
EP2709356B1 (en) Method for operating a front camera of a motor vehicle considering the light of the headlight, corresponding device and motor vehicle
KR102155374B1 (en) System and method for forming nighttime images for a motor vehicle
EP3168779A1 (en) Method for identifying an incoming vehicle and corresponding system
US10217006B2 (en) Method and device for detecting objects in the dark using a vehicle camera and a vehicle lighting system
GB2586802A (en) System and method for identifying light emitter flicker
JP6866212B2 (en) Display control device, display control method and camera monitoring system
JP2016038700A (en) Vehicle external environment recognition device
US20220329723A1 (en) Method and system for mitigating image flicker from strobed lighting systems
KR20140030749A (en) Apparatus for detecting inter-vehicle distance using lamp image and method for detecting inter-vehicle distance using the same
GB2586804A (en) Method and system for mitigating image flicker from strobed lighting systems
CN110798675A (en) Camera module
JP2021044667A (en) Exposure control apparatus

Legal Events

Date Code Title Description
AS Assignment

Owner name: HITACHI AUTOMOTIVE SYSTEM, LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOROGI, NAOKI;OTSUKA, YUJI;SIGNING DATES FROM 20160525 TO 20160530;REEL/FRAME:038903/0182

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STCV Information on status: appeal procedure

Free format text: NOTICE OF APPEAL FILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION