US20170041591A1 - Vehicle-Mounted Image Recognition Device - Google Patents
Vehicle-Mounted Image Recognition Device Download PDFInfo
- Publication number
- US20170041591A1 US20170041591A1 US15/102,282 US201415102282A US2017041591A1 US 20170041591 A1 US20170041591 A1 US 20170041591A1 US 201415102282 A US201415102282 A US 201415102282A US 2017041591 A1 US2017041591 A1 US 2017041591A1
- Authority
- US
- United States
- Prior art keywords
- light source
- imaging unit
- unit
- flickering
- cycle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000003384 imaging method Methods 0.000 claims abstract description 115
- 238000000034 method Methods 0.000 claims description 11
- 238000010586 diagram Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000015556 catabolic process Effects 0.000 description 3
- 238000006731 degradation reaction Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000007792 addition Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- H04N13/0296—
-
- G06K9/00825—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/141—Control of illumination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/582—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/584—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
-
- H04N13/0239—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/45—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from two or more image sensors being of different type or operating in different modes, e.g. with a CMOS sensor for moving images in combination with a charge-coupled device [CCD] for still images
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/57—Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/71—Circuitry for evaluating the brightness variation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/73—Circuitry for compensating brightness variation in the scene by influencing the exposure time
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/745—Detection of flicker frequency or suppression of flicker wherein the flicker is caused by illumination, e.g. due to fluorescent tube illumination or pulsed LED illumination
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- H04N5/2258—
-
- H04N5/23229—
-
- H04N5/2351—
-
- H04N5/2353—
-
- H04N5/2357—
Definitions
- the present invention relates to an in-vehicle camera, and particularly to an in-vehicle image recognition device having a function of simultaneously recognizing a plurality of targets such as a preceding vehicle, a white lane, a pedestrian, and a light source.
- the LED used in the traffic signal and the traffic sign is flickered to be repeatedly on and off at a high speed when being turned on.
- a cycle of the flickering is 100 Hz in a region of 50 Hz commercial power source or 120 Hz in a region of 60 Hz commercial power source. The flickering is performed at a high speed, and thus it appears to person as always turned on.
- the emission light of the light source is not recorded in the image obtained from the camera.
- the capturing cycle of the camera is close to an integer times the cycle of the flickering light source, the above problem continuously occurs in a plurality of frames.
- PTL 1 discloses a technique in which a number of imaging units are provided to capture an image while deviating the exposure timings in a time shorter than a flickering cycle of the flickering light source and to observe a change in brightness of the light source in order to recognize the flickering light source.
- an image is acquired in a cycle Tled shorter than the flickering cycle of the light source, and a recognition process is performed on each image, thereby significantly increasing a processing load.
- a transfer amount of a bus may be significantly expanded.
- An in-vehicle camera having an image recognition function is requested for simultaneously recognizing a number of objects such as a preceding vehicle, a pedestrian, and a white lane on a road. Therefore, when the processing load and the bus transfer amount are increased, it becomes difficult to be compatible with the other recognition processing in addition, since more imaging units are required, the cost is increased.
- the invention has been made in view of the above problems, and an object thereof is to provide an in-vehicle imaging device which can improve a performance of recognizing the flickering light source at a low cost without causing a degradation of an image recognition application such as a preceding vehicle recognition and a white lane recognition.
- an in-vehicle image recognition device in order to solve the above issue which processes an image captured by a pair of imaging units, wherein the same light source is captured by the pair of imaging units in a capturing cycle which is an integer times a predetermined flickering cycle, and a capturing cycle where exposure timings of the pair of imaging units are deviated by a half of the flickering cycle, and wherein, in a case where a difference between a pixel value of the light source captured by one imaging unit and a pixel value of the light source captured by the other imaging unit is equal to or more than a threshold, the light source is determined as a flickering light source.
- the visibility of a flickering light source can be improved at a low cost without causing a degradation of the performance of an image recognition application such as a preceding vehicle recognition and a white lane recognition.
- FIG. 1 is a block diagram for describing functions of an in-vehicle image recognition device.
- FIG. 2 is a diagram illustrating an example of flickering of a flickering light source and exposure timings of a first imaging unit and a second imaging unit.
- FIG. 3 is a flowchart illustrating an exemplary operation of an exposure timing adjustment unit.
- FIG. 1 is a functional block diagram illustrating an example of an in-vehicle image recognition device in this embodiment.
- An in-vehicle image recognition device 10 is a device which alternately captures an image at a different timing using a first imaging unit 11 a and a second imaging unit 11 b which are paired in an in-vehicle camera, detects a flickering light source from one or both of a first image 12 a obtained from the first imaging unit 11 a and a second image 12 b obtained from the second imaging unit 11 b, and performs exposure using the first imaging unit 11 a and the second imaging unit 11 b in synchronization with timing of turning on the flickering light source from next time, so that the flickering light source is recognized with high accuracy without causing degradation of the performance of an image recognition function.
- the in-vehicle image recognition device 10 is a stereo camera which is mounted in a vehicle and, as described in the drawing, provided with the first imaging unit 11 a and the second imaging unit 11 b, a distance data generation unit 13 which generates distance information based on the first image 12 a and the second image 12 b acquired by the first imaging unit 11 a and the second imaging unit 11 b, a distance data storage unit 14 which stores the distance information indicating a distance to a capturing target, an image recognition unit 15 which performs image recognition, a light source recognition unit 16 which performs recognition of a light source, an exposure timing adjustment unit 17 which adjusts each timing of exposure, an external information reception unit 18 which receives external information, and a first exposure control unit 19 a and a second exposure control unit 19 b which control the exposure of the first imaging unit 11 a and the second imaging unit 11 b.
- the first imaging unit 11 a and the second imaging unit 11 b are directed in the same direction and capture the same range.
- the imaging units are separately disposed on right and left in the vicinity of a rear-view mirror of the vehicle to capture the forward side of the vehicle through a front class
- the distance data generation unit 13 generates the distance information such as a parallax image stereoscopically obtained by the first image 12 a and the second image 12 b.
- the image recognition unit 15 recognizes an object body such as a traffic signal, a traffic sign, a preceding vehicle, a pedestrian, and a bicycle using at least one of the first image 12 a and the second image 12 b. As a recognition method, the well-known pattern matching may be used.
- the light source recognition unit 16 recognizes a light source from at least one of the first image 12 a and the second image 12 b, and recognizes whether the recognized light source is a flickering light source.
- the exposure timing adjustment unit 17 adjusts the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b based on an image recognition result of the image recognition unit 15 and the external information received by the external information reception unit 18 .
- the first exposure control unit 19 a and the second exposure control unit 19 b control the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b according to, for example, an instruction from the exposure timing adjustment unit 17 .
- the exposure timings are adjusted to be exposure parameters optimized to the image recognition based on the information of the images acquired from the first imaging unit 11 a and the second imaging unit 11 b.
- the external information reception unit 18 is connected to a vehicle network 20 and acquires the external information such as vehicle speed information 21 , a yaw rate 22 , GPS information 23 , time information 24 , and map data 25 using the external information reception unit 18 .
- the vehicle image recognition device 10 When capturing the traffic signal using an LED as the flickering light source and an indicator such as the traffic sign of an electric light display type, the vehicle image recognition device 10 captures the images at exposure timings of the first imaging unit 11 a and the second imaging unit 11 b in synchronization with a flickering period of the LED.
- capturing cycles Tc of the first imaging unit 11 a and the second imaging unit 11 b are set to be Nc times (an integer) of a flickering cycle Tl, and the exposure timing of the first imaging unit 11 a and the exposure timing of the second imaging unit 11 b are set to be deviated by Ns (an integer)+0.5 times the flickering cycle, so that the first image and the second image deviated by a half cycle with respect to the flickering cycle are acquired.
- the light source in the ON state is recorded in at least any one of the first image 12 a and the second image 12 b.
- the light source recognition unit 16 performs a recognition process in which the light source is extracted from the first image 12 a and the second image 12 b. Then, in a case where there is a difference equal to or more than a predetermined value between pixel values of the light source extracted from the first image 12 a and the second image 12 b, it is determined that the subject light source is a flickering light source.
- the exposure timing adjustment unit 17 changes the exposure timing such that the exposure timing of one imaging unit is changed to be matched to the exposure timing of the other imaging unit in which the flickering light source is brightly captured among the first imaging unit 11 a and the second imaging unit 11 b, and adjusts the exposure timings of the first imaging unit 11 a and the second imaging unit 11 b to be synchronized.
- the distance data generation unit generates distance data indicating a distance from the vehicle to the light source based on the first image and the second image captured by the first imaging unit and the second imaging unit.
- FIG. 2 is a diagram illustrating an example of flickering of the flickering light source and exposure timings of the first imaging unit and the second imaging unit.
- the exposure timing adjustment unit 17 sets the capturing cycles Tc of the first imaging unit 11 a and the second imaging unit 11 b to be Nc (an integer) times the flickering cycle Tl. Therefore, even the flickering light source can be captured in the same brightness in every frame, and the brightness of the flickering light source illuminated on each frame can be made constant.
- the flickering light source at the exposure timing of the other imaging unit necessarily comes to be bright since the exposure timing of the other imaging unit is deviated by the half cycle with respect to the flickering cycle Tl.
- the light source recognition unit 16 can recognize that the light source is flickering by recognizing that there is a light source in the image obtained by an imaging unit and there is no light source in the image obtained by the other imaging unit. Therefore, it is possible to reliably capture the flickering light source from next time by causing the first imaging unit 11 a and the second imaging unit 11 b to perform exposure in synchronization with timing when the flickering light source is bright. Therefore, the flickering light source can be stereoscopically viewed, and accurate distance information up to the flickering light source can be obtained.
- the flickering light source can be captured with a sufficient pixel value, and can be recognized similarly to an unflickering light source (a light source always turned on).
- the light source can be recognized in synchronization with the exposure timing of any one of the first imaging unit 11 a and the second imaging unit 11 b. Therefore, the flickering light source can be recognized and the distance information can be obtained by performing the stereoscopic viewing in synchronization with the exposure timing of any one of the imaging units. In other words, the flickering light source can be recognized similarly to a case where the light always turned on is recognized.
- any method such as a template matching and a classification using color information may be used as long as a feature can be extracted from the image, and a well-known method may be used.
- FIG. 3 is a flowchart for describing an exemplary method of setting the exposure timing using the exposure timing adjustment unit.
- the exposure timing adjustment unit 17 determines whether the stereoscopic distance data is generated based on the recognition result of the image recognition unit 15 and the information acquired from the external information reception unit 18 .
- Step S 1 recognition result information is received from the image recognition unit 15 and the light source recognition unit 16 .
- Step S 2 the external information such as the vehicle speed information 21 , the yaw rate 22 , the GPS information 23 , the time information 24 , and the map data 25 linked to the vehicle network 20 are acquired from the external information reception unit 18 .
- Step S 3 it is determined whether the flickering cycle Tl of the flickering light source is 1/50 second. Since the capturing cycle Tc and a capture deviation time are determined based on the flickering cycle of the flickering light source, the flickering cycle of the flickering light source is necessarily determined.
- Step S 3 in a case where it is determined that the flickering cycle of the flickering light source is 1/50 second, the flickering cycle of the flickering light source is set to 1/50 second in Step S 4 , and if not (that is, it is determined that the flickering cycle is not 1/50 second), the flickering cycle of the flickering power source is set to 1/60 second in Step S 5 .
- the flickering cycle is set to any one of 1/50 and 1/60 second. For example, in a case where the pixel value of the light source is almost the same value in the image captured by the same imaging unit, it can be determined that the capturing cycle is an integer times an intended flickering cycle and the flickering cycle is correctly set.
- the flickering cycle of the light source is different from the current setting, and thus the setting of the flickering cycle is switched to the other value.
- the flickering cycle from next time is set to the cycle at the previous time.
- the setting is switched.
- the flickering cycle of the flickering light source may be determined based on a power source frequency of the flickering light source. Since the frequency of the commercial power source is determined depending on regions, the flickering cycle may be determined according to GPS information or a running location of a subject vehicle received in cooperation with a car navigation system. For example, in the case of Japan, the power source frequency in the east region of Japan is 50 Hz, and the power source frequency in the west region of Japan is 60 Hz. Therefore, the power source frequency of the flickering power source can be recognized by specifying the current location of the vehicle based on the external information such as the GPS information 23 and the map data 25 .
- the flickering cycle of the flickering light source can be set to 1/60 second.
- Step S 6 it is determined whether a light emission timing of the flickering light source is estimated.
- the first imaging unit 11 a and the second imaging unit 11 b are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S 8 .
- the flickering light source can reliably perform the capturing using both the first imaging unit 11 a and the second imaging unit 11 b. Further, the distance information up to the light source can be stereoscopically acquired.
- Step S 7 it is determined whether the distance information is necessary in Step S 7 . While the acquirement of the distance information using the stereo camera is a feature, the distance information may be not necessary for all the frames. Therefore, the stereoscopic viewing (the imaging units are caused to perform the capturing at the same exposure timing) or the deviating of the exposure timings is selected according to a situation based on the recognition result acquired in Step S 1 and the external information acquired in Step S 2 . In a case where it is determined that the distance information is necessary, the first imaging unit and the second imaging unit are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S 8 . In a case where it is determined that the distance information is not necessary, the capturing is performed by deviating the exposure timings of the first imaging unit and the second imaging unit by a predetermined time in Step S 9 .
- the capturing is performed at the same exposure timing based on the other applications with priority higher than that of the light emission timing of the flickering light source.
- the preceding vehicle and the pedestrian are recognized from the image, and the result is acquired from the image recognition unit 15 .
- the preceding vehicle is within a predetermined distance, or in a case where the pedestrian or the bicycle is near the advancing path of the subject vehicle, it is determined that the distance information is necessary for avoiding a risk.
- a three-dimensional object can be effectively detected in the stereoscope viewing. Therefore, a counter is provided in the device, and the capturing may be performed alternately using the stereoscopic viewing and the deviating of the exposure timings in a predetermined cycle even in a case where there is no change in an ambient environment.
- the exposure timings may be deviated only when the subject vehicle comes close to a predetermined distance or more. It is desirable that a parameter related to the exposure be optimal to recognize the preceding vehicle and the pedestrian. However, for example, in a case where it is determined that the light source is recognized with priority, a shutter opening time may be lengthened.
- the shutter opening time and a pixel transfer amount of the imaging unit are not changed, so that the performance of the image recognition application is not damaged.
- a processing load is not increased.
- the recognition performance of the flickering light source can be improved at a low cost.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Studio Devices (AREA)
- Traffic Control Systems (AREA)
Abstract
Description
- The present invention relates to an in-vehicle camera, and particularly to an in-vehicle image recognition device having a function of simultaneously recognizing a plurality of targets such as a preceding vehicle, a white lane, a pedestrian, and a light source.
- In recent years, light sources used in a traffic signal and a traffic sign of an electric light display type are replaced from the conventional incandescent lamps to LEDs. The LED has a merit on less power consumption and long life span compared to the incandescent lamp. In addition, there is no pseudo lighting considered as emission caused by the reflection of sunlight, so that the visibility is high and also the stability is improved.
- The LED used in the traffic signal and the traffic sign is flickered to be repeatedly on and off at a high speed when being turned on. A cycle of the flickering is 100 Hz in a region of 50 Hz commercial power source or 120 Hz in a region of 60 Hz commercial power source. The flickering is performed at a high speed, and thus it appears to person as always turned on.
- In a case where the exposure timing of an in-vehicle camera is matched to the OFF state of the flickering light source, the emission light of the light source is not recorded in the image obtained from the camera. In addition, in a case where the capturing cycle of the camera is close to an integer times the cycle of the flickering light source, the above problem continuously occurs in a plurality of frames.
- In order to solve the above problem, for example,
PTL 1 discloses a technique in which a number of imaging units are provided to capture an image while deviating the exposure timings in a time shorter than a flickering cycle of the flickering light source and to observe a change in brightness of the light source in order to recognize the flickering light source. - PTL 1: Japanese Patent Application Laid-Open No. 2007-286943
- However, in the conventional technique described above, an image is acquired in a cycle Tled shorter than the flickering cycle of the light source, and a recognition process is performed on each image, thereby significantly increasing a processing load. In addition, since a number of images are transferred from an imaging unit to a recognition unit, a transfer amount of a bus may be significantly expanded. An in-vehicle camera having an image recognition function is requested for simultaneously recognizing a number of objects such as a preceding vehicle, a pedestrian, and a white lane on a road. Therefore, when the processing load and the bus transfer amount are increased, it becomes difficult to be compatible with the other recognition processing in addition, since more imaging units are required, the cost is increased.
- The invention has been made in view of the above problems, and an object thereof is to provide an in-vehicle imaging device which can improve a performance of recognizing the flickering light source at a low cost without causing a degradation of an image recognition application such as a preceding vehicle recognition and a white lane recognition.
- There is provided an in-vehicle image recognition device according to the present invention in order to solve the above issue which processes an image captured by a pair of imaging units, wherein the same light source is captured by the pair of imaging units in a capturing cycle which is an integer times a predetermined flickering cycle, and a capturing cycle where exposure timings of the pair of imaging units are deviated by a half of the flickering cycle, and wherein, in a case where a difference between a pixel value of the light source captured by one imaging unit and a pixel value of the light source captured by the other imaging unit is equal to or more than a threshold, the light source is determined as a flickering light source.
- According to the invention, the visibility of a flickering light source can be improved at a low cost without causing a degradation of the performance of an image recognition application such as a preceding vehicle recognition and a white lane recognition. Further, other objects, configurations, and effects will become clear through the explanation of the following embodiments.
- [
FIG. 1 ]FIG. 1 is a block diagram for describing functions of an in-vehicle image recognition device. - [
FIG. 2 ]FIG. 2 is a diagram illustrating an example of flickering of a flickering light source and exposure timings of a first imaging unit and a second imaging unit. - [
FIG. 3 ]FIG. 3 is a flowchart illustrating an exemplary operation of an exposure timing adjustment unit. - Next, embodiments of the invention will be described below using the drawings.
FIG. 1 is a functional block diagram illustrating an example of an in-vehicle image recognition device in this embodiment. - An in-vehicle
image recognition device 10 is a device which alternately captures an image at a different timing using a first imaging unit 11 a and asecond imaging unit 11 b which are paired in an in-vehicle camera, detects a flickering light source from one or both of afirst image 12 a obtained from the first imaging unit 11 a and asecond image 12 b obtained from thesecond imaging unit 11 b, and performs exposure using the first imaging unit 11 a and thesecond imaging unit 11 b in synchronization with timing of turning on the flickering light source from next time, so that the flickering light source is recognized with high accuracy without causing degradation of the performance of an image recognition function. - The in-vehicle
image recognition device 10 is a stereo camera which is mounted in a vehicle and, as described in the drawing, provided with the first imaging unit 11 a and thesecond imaging unit 11 b, a distancedata generation unit 13 which generates distance information based on thefirst image 12 a and thesecond image 12 b acquired by the first imaging unit 11 a and thesecond imaging unit 11 b, a distancedata storage unit 14 which stores the distance information indicating a distance to a capturing target, animage recognition unit 15 which performs image recognition, a lightsource recognition unit 16 which performs recognition of a light source, an exposuretiming adjustment unit 17 which adjusts each timing of exposure, an externalinformation reception unit 18 which receives external information, and a firstexposure control unit 19 a and a second exposure control unit 19 b which control the exposure of the first imaging unit 11 a and thesecond imaging unit 11 b. - The first imaging unit 11 a and the
second imaging unit 11 b are directed in the same direction and capture the same range. For example, the imaging units are separately disposed on right and left in the vicinity of a rear-view mirror of the vehicle to capture the forward side of the vehicle through a front class - The distance
data generation unit 13 generates the distance information such as a parallax image stereoscopically obtained by thefirst image 12 a and thesecond image 12 b. Theimage recognition unit 15 recognizes an object body such as a traffic signal, a traffic sign, a preceding vehicle, a pedestrian, and a bicycle using at least one of thefirst image 12 a and thesecond image 12 b. As a recognition method, the well-known pattern matching may be used. - The light
source recognition unit 16 recognizes a light source from at least one of thefirst image 12 a and thesecond image 12 b, and recognizes whether the recognized light source is a flickering light source. The exposuretiming adjustment unit 17 adjusts the exposure timings of the first imaging unit 11 a and thesecond imaging unit 11 b based on an image recognition result of theimage recognition unit 15 and the external information received by the externalinformation reception unit 18. The firstexposure control unit 19 a and the second exposure control unit 19 b control the exposure timings of the first imaging unit 11 a and thesecond imaging unit 11 b according to, for example, an instruction from the exposuretiming adjustment unit 17. In addition, the exposure timings are adjusted to be exposure parameters optimized to the image recognition based on the information of the images acquired from the first imaging unit 11 a and thesecond imaging unit 11 b. The externalinformation reception unit 18 is connected to avehicle network 20 and acquires the external information such asvehicle speed information 21, ayaw rate 22,GPS information 23,time information 24, andmap data 25 using the externalinformation reception unit 18. - When capturing the traffic signal using an LED as the flickering light source and an indicator such as the traffic sign of an electric light display type, the vehicle
image recognition device 10 captures the images at exposure timings of the first imaging unit 11 a and thesecond imaging unit 11 b in synchronization with a flickering period of the LED. - Specifically, capturing cycles Tc of the first imaging unit 11 a and the
second imaging unit 11 b are set to be Nc times (an integer) of a flickering cycle Tl, and the exposure timing of the first imaging unit 11 a and the exposure timing of thesecond imaging unit 11 b are set to be deviated by Ns (an integer)+0.5 times the flickering cycle, so that the first image and the second image deviated by a half cycle with respect to the flickering cycle are acquired. - Therefore, the light source in the ON state is recorded in at least any one of the
first image 12 a and thesecond image 12 b. - The light
source recognition unit 16 performs a recognition process in which the light source is extracted from thefirst image 12 a and thesecond image 12 b. Then, in a case where there is a difference equal to or more than a predetermined value between pixel values of the light source extracted from thefirst image 12 a and thesecond image 12 b, it is determined that the subject light source is a flickering light source. In a case where the lightsource recognition unit 16 determines that the light source is a flickering light source, the exposuretiming adjustment unit 17 changes the exposure timing such that the exposure timing of one imaging unit is changed to be matched to the exposure timing of the other imaging unit in which the flickering light source is brightly captured among the first imaging unit 11 a and thesecond imaging unit 11 b, and adjusts the exposure timings of the first imaging unit 11 a and thesecond imaging unit 11 b to be synchronized. The distance data generation unit generates distance data indicating a distance from the vehicle to the light source based on the first image and the second image captured by the first imaging unit and the second imaging unit. - Next, an example of adjusting the exposure timing of the invention will be described using
FIG. 2 .FIG. 2 is a diagram illustrating an example of flickering of the flickering light source and exposure timings of the first imaging unit and the second imaging unit. As illustrated inFIG. 2 , in a case where the brightness of the LED is increased or decreased in the flickering cycle Tl and the flickering light source emits the light, the exposuretiming adjustment unit 17 sets the capturing cycles Tc of the first imaging unit 11 a and thesecond imaging unit 11 b to be Nc (an integer) times the flickering cycle Tl. Therefore, even the flickering light source can be captured in the same brightness in every frame, and the brightness of the flickering light source illuminated on each frame can be made constant. - Then, the exposure
timing adjustment unit 17 sets a deviation time between the exposure timing of the first imaging unit and the exposure timing of the second imaging unit to be Ns (an integer)+0.5 times the flickering cycle Tl (Deviation Time=Tl×(Ns+0.5)). Therefore, the first imaging unit 11 a and thesecond imaging unit 11 b can acquire brightnesses deviated by the half flickering cycle Tl of the LED. Herein, in a case where the exposure timing of one of the first imaging unit 11 a and thesecond imaging unit 11 b is a timing when the flickering light source is dark, the flickering light source at the exposure timing of the other imaging unit necessarily comes to be bright since the exposure timing of the other imaging unit is deviated by the half cycle with respect to the flickering cycle Tl. - The light
source recognition unit 16 can recognize that the light source is flickering by recognizing that there is a light source in the image obtained by an imaging unit and there is no light source in the image obtained by the other imaging unit. Therefore, it is possible to reliably capture the flickering light source from next time by causing the first imaging unit 11 a and thesecond imaging unit 11 b to perform exposure in synchronization with timing when the flickering light source is bright. Therefore, the flickering light source can be stereoscopically viewed, and accurate distance information up to the flickering light source can be obtained. - Even at a timing when the flickering light source is captured in the same brightness by the first imaging unit 11 a and the
second imaging unit 11 b, the traffic signal and the indicator such as the traffic sign using the LED are sufficiently bright compared to the preceding vehicle and a road surface in order to secure visibility. Therefore, even in a case where an optimal exposure is performed in order to recognize the preceding vehicle and the road surface, the flickering light source can be captured with a sufficient pixel value, and can be recognized similarly to an unflickering light source (a light source always turned on). - In this case, the light source can be recognized in synchronization with the exposure timing of any one of the first imaging unit 11 a and the
second imaging unit 11 b. Therefore, the flickering light source can be recognized and the distance information can be obtained by performing the stereoscopic viewing in synchronization with the exposure timing of any one of the imaging units. In other words, the flickering light source can be recognized similarly to a case where the light always turned on is recognized. - Further, as a method of recognizing the light source from an image, any method such as a template matching and a classification using color information may be used as long as a feature can be extracted from the image, and a well-known method may be used.
-
FIG. 3 is a flowchart for describing an exemplary method of setting the exposure timing using the exposure timing adjustment unit. The exposuretiming adjustment unit 17 determines whether the stereoscopic distance data is generated based on the recognition result of theimage recognition unit 15 and the information acquired from the externalinformation reception unit 18. - First, in Step S1, recognition result information is received from the
image recognition unit 15 and the lightsource recognition unit 16. Next, in Step S2, the external information such as thevehicle speed information 21, theyaw rate 22, theGPS information 23, thetime information 24, and themap data 25 linked to thevehicle network 20 are acquired from the externalinformation reception unit 18. - Then, in Step S3, it is determined whether the flickering cycle Tl of the flickering light source is 1/50 second. Since the capturing cycle Tc and a capture deviation time are determined based on the flickering cycle of the flickering light source, the flickering cycle of the flickering light source is necessarily determined.
- In Step S3, in a case where it is determined that the flickering cycle of the flickering light source is 1/50 second, the flickering cycle of the flickering light source is set to 1/50 second in Step S4, and if not (that is, it is determined that the flickering cycle is not 1/50 second), the flickering cycle of the flickering power source is set to 1/60 second in Step S5.
- At the first time, since the flickering cycle is not clear, the flickering cycle is set to any one of 1/50 and 1/60 second. For example, in a case where the pixel value of the light source is almost the same value in the image captured by the same imaging unit, it can be determined that the capturing cycle is an integer times an intended flickering cycle and the flickering cycle is correctly set.
- On the other hand, in a case where the pixel value of the light source is changed by stages in the image captured by the same imaging unit, it is determined that the flickering cycle of the light source is different from the current setting, and thus the setting of the flickering cycle is switched to the other value. The flickering cycle from next time is set to the cycle at the previous time. In a case where it is determined that the flickering cycle is different on the same basis, the setting is switched.
- The flickering cycle of the flickering light source may be determined based on a power source frequency of the flickering light source. Since the frequency of the commercial power source is determined depending on regions, the flickering cycle may be determined according to GPS information or a running location of a subject vehicle received in cooperation with a car navigation system. For example, in the case of Japan, the power source frequency in the east region of Japan is 50 Hz, and the power source frequency in the west region of Japan is 60 Hz. Therefore, the power source frequency of the flickering power source can be recognized by specifying the current location of the vehicle based on the external information such as the
GPS information 23 and themap data 25. For example, in a case where the power source frequency is 50 Hz, and in a case where the flickering cycle of the flickering light source is 1/50 second and the power source frequency is 60 Hz, the flickering cycle of the flickering light source can be set to 1/60 second. - Next, in Step S6, it is determined whether a light emission timing of the flickering light source is estimated. In a case where the light emission timing of the flickering light source is estimated, the first imaging unit 11 a and the
second imaging unit 11 b are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S8. In this case, the flickering light source can reliably perform the capturing using both the first imaging unit 11 a and thesecond imaging unit 11 b. Further, the distance information up to the light source can be stereoscopically acquired. - On the other hand, in a case where the light emission timing of the flickering light source is not estimated yet, it is determined whether the distance information is necessary in Step S7. While the acquirement of the distance information using the stereo camera is a feature, the distance information may be not necessary for all the frames. Therefore, the stereoscopic viewing (the imaging units are caused to perform the capturing at the same exposure timing) or the deviating of the exposure timings is selected according to a situation based on the recognition result acquired in Step S1 and the external information acquired in Step S2. In a case where it is determined that the distance information is necessary, the first imaging unit and the second imaging unit are caused to simultaneously perform the capturing in synchronization with the light emission timing of the flickering light source in Step S8. In a case where it is determined that the distance information is not necessary, the capturing is performed by deviating the exposure timings of the first imaging unit and the second imaging unit by a predetermined time in Step S9.
- For example, in a case where it is a situation that the distance information from the stereo camera is necessary for the other applications such as a pedestrian detection and a preceding vehicle detection, the capturing is performed at the same exposure timing based on the other applications with priority higher than that of the light emission timing of the flickering light source.
- In the in-vehicle
image recognition device 10, the preceding vehicle and the pedestrian are recognized from the image, and the result is acquired from theimage recognition unit 15. In a case where the preceding vehicle is within a predetermined distance, or in a case where the pedestrian or the bicycle is near the advancing path of the subject vehicle, it is determined that the distance information is necessary for avoiding a risk. In addition, a three-dimensional object can be effectively detected in the stereoscope viewing. Therefore, a counter is provided in the device, and the capturing may be performed alternately using the stereoscopic viewing and the deviating of the exposure timings in a predetermined cycle even in a case where there is no change in an ambient environment. - In a case where the traffic signal or the traffic sign of an electric light display type can be acquired from the
map data 25, the exposure timings may be deviated only when the subject vehicle comes close to a predetermined distance or more. It is desirable that a parameter related to the exposure be optimal to recognize the preceding vehicle and the pedestrian. However, for example, in a case where it is determined that the light source is recognized with priority, a shutter opening time may be lengthened. - According to the in-vehicle
image recognition device 10 described above, the shutter opening time and a pixel transfer amount of the imaging unit are not changed, so that the performance of the image recognition application is not damaged. In addition, since the extraction process of the light source is performed as in the conventional process, a processing load is not increased. Further, there is no increase in hardware, the recognition performance of the flickering light source can be improved at a low cost. - Hitherto, the description has been made about embodiments of the invention, but the invention is not limited to the embodiments. Various changes in design can be made within a scope not departing from the spirit of the invention described in the accompanying claims. For example, the embodiments are described in a clearly understandable way for the invention, and thus the invention is not necessarily to include all the configurations described above. In addition, some configurations of a certain embodiment may be replaced with the configurations of another embodiment, and the configuration of the other embodiment may also be added to that of a certain embodiment. Furthermore, various additions, omissions, and substitutions may be partially made in some of the configurations of the respective embodiments.
-
- 10 in-vehicle image recognition device
- 11 a first imaging unit
- 11 b second imaging unit
- 12 a first image
- 12 b second image
- 13 distance data generation unit
- 14 distance data
- 15 image recognition unit
- 16 light source recognition unit
- 17 exposure timing adjustment unit
- 18 external information reception unit
- 19 a first exposure control unit
- 19 b second exposure control unit
- 20 vehicle network
- 21 vehicle speed information
- 22 yaw rate
- 23 GPS information
- 24 time information
- 25 map data
Claims (8)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2013-267292 | 2013-12-25 | ||
JP2013267292 | 2013-12-25 | ||
PCT/JP2014/079902 WO2015098325A1 (en) | 2013-12-25 | 2014-11-12 | Vehicle-mounted image recognition device |
Publications (1)
Publication Number | Publication Date |
---|---|
US20170041591A1 true US20170041591A1 (en) | 2017-02-09 |
Family
ID=53478202
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/102,282 Abandoned US20170041591A1 (en) | 2013-12-25 | 2014-11-12 | Vehicle-Mounted Image Recognition Device |
Country Status (4)
Country | Link |
---|---|
US (1) | US20170041591A1 (en) |
EP (1) | EP3089442B1 (en) |
JP (1) | JP6325000B2 (en) |
WO (1) | WO2015098325A1 (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160031371A1 (en) * | 2014-07-29 | 2016-02-04 | Denso Corporation | In-vehicle apparatus |
US20160371552A1 (en) * | 2014-03-10 | 2016-12-22 | Nissan Motor Co., Ltd. | Traffic Light Detecting Device and Traffic Light Detecting Method |
US20170017850A1 (en) * | 2014-03-10 | 2017-01-19 | Nissan Motor Co., Ltd. | Traffic Light Detecting Device and Traffic Light Detecting Method |
US20170339326A1 (en) * | 2016-05-20 | 2017-11-23 | Gentex Corporation | Imager with active exposure correction |
JP2018064257A (en) * | 2016-10-15 | 2018-04-19 | キヤノン株式会社 | Imaging System |
US10055655B2 (en) * | 2015-04-08 | 2018-08-21 | Nissan Motor Co., Ltd. | Traffic light detection device and traffic light detection method |
WO2019050654A1 (en) * | 2017-09-07 | 2019-03-14 | Qualcomm Incorporated | Image capture setting determination in devices having access to multiple cameras |
CN111291620A (en) * | 2020-01-14 | 2020-06-16 | 北京小马智行科技有限公司 | Method, device and system for identifying light source |
US10715738B1 (en) * | 2019-04-30 | 2020-07-14 | Axon Enterprise, Inc. | Asymmetrical license plate reading (ALPR) camera system |
US10863106B1 (en) * | 2019-10-21 | 2020-12-08 | GM Global Technology Operations LLC | Systems and methods for LED flickering and banding detection |
WO2021043892A1 (en) * | 2019-09-03 | 2021-03-11 | Jaguar Land Rover Limited | Method and system for mitigating image flicker from strobed lighting systems |
US11030472B2 (en) | 2019-04-30 | 2021-06-08 | Axon Enterprise, Inc. | Asymmetrical license plate reading (ALPR) camera system |
US11343441B2 (en) * | 2017-08-21 | 2022-05-24 | Sony Semiconductor Solutions Corporation | Imaging device and apparatus |
US11483505B2 (en) * | 2017-11-06 | 2022-10-25 | SK Hynix Inc. | Image synchronization device and image information generation apparatus including the same |
US11532170B2 (en) | 2019-04-30 | 2022-12-20 | Axon Enterprise, Inc. | License plate reading system with enhancements |
EP4086874A4 (en) * | 2020-01-23 | 2023-05-31 | Huawei Technologies Co., Ltd. | Control method and apparatus for intelligent vehicle, and related device |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109496187B (en) * | 2016-08-08 | 2022-07-26 | 金泰克斯公司 | System and method for processing video data to detect and eliminate flicker light source through dynamic exposure control |
WO2018225516A1 (en) * | 2017-06-07 | 2018-12-13 | 日立オートモティブシステムズ株式会社 | Image processing device |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030133018A1 (en) * | 2002-01-16 | 2003-07-17 | Ted Ziemkowski | System for near-simultaneous capture of multiple camera images |
US20040012692A1 (en) * | 2002-03-25 | 2004-01-22 | Shinichi Arazaki | Flicker detection apparatus, a flicker correction apparatus, an image-pickup apparatus, a flicker detection program and a flicker correction program |
JP2007286943A (en) * | 2006-04-18 | 2007-11-01 | Fujifilm Corp | Signal light detection apparatus |
JP2008134844A (en) * | 2006-11-28 | 2008-06-12 | Alpine Electronics Inc | Drive recorder, and image acquisition timing control method therefor |
WO2008108025A1 (en) * | 2007-03-05 | 2008-09-12 | Nec Electronics Corporation | Imaging apparatus and flicker detection method |
JP2009065586A (en) * | 2007-09-10 | 2009-03-26 | Hitachi Ltd | On-vehicle camera |
US20150035949A1 (en) * | 2013-07-31 | 2015-02-05 | Morpho | Method for synchronising several cameras with each other in a photographic system, in particular a stereoscopic photographic system, and photographic system for implementing said method |
US20150138324A1 (en) * | 2012-05-16 | 2015-05-21 | Denso Corporation | Apparatus for detecting vehicle light and method thereof |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011234318A (en) * | 2010-04-30 | 2011-11-17 | Yamaha Corp | Imaging device |
-
2014
- 2014-11-12 EP EP14873199.5A patent/EP3089442B1/en active Active
- 2014-11-12 JP JP2015554663A patent/JP6325000B2/en active Active
- 2014-11-12 US US15/102,282 patent/US20170041591A1/en not_active Abandoned
- 2014-11-12 WO PCT/JP2014/079902 patent/WO2015098325A1/en active Application Filing
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030133018A1 (en) * | 2002-01-16 | 2003-07-17 | Ted Ziemkowski | System for near-simultaneous capture of multiple camera images |
US20040012692A1 (en) * | 2002-03-25 | 2004-01-22 | Shinichi Arazaki | Flicker detection apparatus, a flicker correction apparatus, an image-pickup apparatus, a flicker detection program and a flicker correction program |
JP2007286943A (en) * | 2006-04-18 | 2007-11-01 | Fujifilm Corp | Signal light detection apparatus |
JP2008134844A (en) * | 2006-11-28 | 2008-06-12 | Alpine Electronics Inc | Drive recorder, and image acquisition timing control method therefor |
WO2008108025A1 (en) * | 2007-03-05 | 2008-09-12 | Nec Electronics Corporation | Imaging apparatus and flicker detection method |
JP2009065586A (en) * | 2007-09-10 | 2009-03-26 | Hitachi Ltd | On-vehicle camera |
US20150138324A1 (en) * | 2012-05-16 | 2015-05-21 | Denso Corporation | Apparatus for detecting vehicle light and method thereof |
US20150035949A1 (en) * | 2013-07-31 | 2015-02-05 | Morpho | Method for synchronising several cameras with each other in a photographic system, in particular a stereoscopic photographic system, and photographic system for implementing said method |
FR3009469A1 (en) * | 2013-07-31 | 2015-02-06 | Morpho | METHOD OF SYNCHRONIZING MULTIPLE CAMERAS OF A SHOOTING SYSTEM, IN PARTICULAR A STEREOSCOPIC VIEWING SYSTEM AND VIEWING SYSTEM FOR IMPLEMENTING SAID METHOD |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160371552A1 (en) * | 2014-03-10 | 2016-12-22 | Nissan Motor Co., Ltd. | Traffic Light Detecting Device and Traffic Light Detecting Method |
US20170017850A1 (en) * | 2014-03-10 | 2017-01-19 | Nissan Motor Co., Ltd. | Traffic Light Detecting Device and Traffic Light Detecting Method |
US9679208B2 (en) * | 2014-03-10 | 2017-06-13 | Nissan Motor Co., Ltd. | Traffic light detecting device and traffic light detecting method |
US9679207B2 (en) * | 2014-03-10 | 2017-06-13 | Nissan Motor Co., Ltd. | Traffic light detecting device and traffic light detecting method |
US20160031371A1 (en) * | 2014-07-29 | 2016-02-04 | Denso Corporation | In-vehicle apparatus |
US10055655B2 (en) * | 2015-04-08 | 2018-08-21 | Nissan Motor Co., Ltd. | Traffic light detection device and traffic light detection method |
US20170339326A1 (en) * | 2016-05-20 | 2017-11-23 | Gentex Corporation | Imager with active exposure correction |
US10154200B2 (en) * | 2016-05-20 | 2018-12-11 | Gentex Corporation | Imager with active exposure correction |
JP2018064257A (en) * | 2016-10-15 | 2018-04-19 | キヤノン株式会社 | Imaging System |
US11343441B2 (en) * | 2017-08-21 | 2022-05-24 | Sony Semiconductor Solutions Corporation | Imaging device and apparatus |
WO2019050654A1 (en) * | 2017-09-07 | 2019-03-14 | Qualcomm Incorporated | Image capture setting determination in devices having access to multiple cameras |
US11483505B2 (en) * | 2017-11-06 | 2022-10-25 | SK Hynix Inc. | Image synchronization device and image information generation apparatus including the same |
US11030472B2 (en) | 2019-04-30 | 2021-06-08 | Axon Enterprise, Inc. | Asymmetrical license plate reading (ALPR) camera system |
US10715738B1 (en) * | 2019-04-30 | 2020-07-14 | Axon Enterprise, Inc. | Asymmetrical license plate reading (ALPR) camera system |
US11532170B2 (en) | 2019-04-30 | 2022-12-20 | Axon Enterprise, Inc. | License plate reading system with enhancements |
US11682219B2 (en) | 2019-04-30 | 2023-06-20 | Axon Enterprise, Inc. | Asymmetrical license plate reading (ALPR) camera system |
US11881039B2 (en) | 2019-04-30 | 2024-01-23 | Axon Enterprise, Inc. | License plate reading system with enhancements |
WO2021043892A1 (en) * | 2019-09-03 | 2021-03-11 | Jaguar Land Rover Limited | Method and system for mitigating image flicker from strobed lighting systems |
CN112766030A (en) * | 2019-10-21 | 2021-05-07 | 通用汽车环球科技运作有限责任公司 | System and method for LED flicker and strip detection |
US10863106B1 (en) * | 2019-10-21 | 2020-12-08 | GM Global Technology Operations LLC | Systems and methods for LED flickering and banding detection |
CN111291620A (en) * | 2020-01-14 | 2020-06-16 | 北京小马智行科技有限公司 | Method, device and system for identifying light source |
EP4086874A4 (en) * | 2020-01-23 | 2023-05-31 | Huawei Technologies Co., Ltd. | Control method and apparatus for intelligent vehicle, and related device |
Also Published As
Publication number | Publication date |
---|---|
WO2015098325A1 (en) | 2015-07-02 |
EP3089442A1 (en) | 2016-11-02 |
JP6325000B2 (en) | 2018-05-16 |
JPWO2015098325A1 (en) | 2017-03-23 |
EP3089442B1 (en) | 2022-01-05 |
EP3089442A4 (en) | 2017-08-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3089442B1 (en) | Vehicle-mounted image recognition device | |
CN103213540B (en) | Vehicle driving environment recognition apparatus | |
CN102047166B (en) | Device, camera, and method for generating images of the vicinity of a motor vehicle | |
US11343441B2 (en) | Imaging device and apparatus | |
JP5863536B2 (en) | Outside monitoring device | |
US20140293055A1 (en) | Image processing apparatus | |
JP5772714B2 (en) | Light detection device and vehicle control system | |
US10055655B2 (en) | Traffic light detection device and traffic light detection method | |
CN107431747B (en) | Image pickup apparatus | |
JP6447289B2 (en) | Imaging apparatus, imaging method, program, vehicle control system, and vehicle | |
US20170116488A1 (en) | Method for identifying an incoming vehicle and corresponding system | |
EP3637758B1 (en) | Image processing device | |
JP6259335B2 (en) | Outside environment recognition device | |
EP2709356B1 (en) | Method for operating a front camera of a motor vehicle considering the light of the headlight, corresponding device and motor vehicle | |
KR102155374B1 (en) | System and method for forming nighttime images for a motor vehicle | |
EP3168779A1 (en) | Method for identifying an incoming vehicle and corresponding system | |
US10217006B2 (en) | Method and device for detecting objects in the dark using a vehicle camera and a vehicle lighting system | |
GB2586802A (en) | System and method for identifying light emitter flicker | |
JP6866212B2 (en) | Display control device, display control method and camera monitoring system | |
JP2016038700A (en) | Vehicle external environment recognition device | |
US20220329723A1 (en) | Method and system for mitigating image flicker from strobed lighting systems | |
KR20140030749A (en) | Apparatus for detecting inter-vehicle distance using lamp image and method for detecting inter-vehicle distance using the same | |
GB2586804A (en) | Method and system for mitigating image flicker from strobed lighting systems | |
CN110798675A (en) | Camera module | |
JP2021044667A (en) | Exposure control apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HITACHI AUTOMOTIVE SYSTEM, LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOROGI, NAOKI;OTSUKA, YUJI;SIGNING DATES FROM 20160525 TO 20160530;REEL/FRAME:038903/0182 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STCV | Information on status: appeal procedure |
Free format text: NOTICE OF APPEAL FILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |