US10026178B2 - Method of determining normality of imaging direction, and evaluation device of evaluating attachment state of imaging device - Google Patents
Method of determining normality of imaging direction, and evaluation device of evaluating attachment state of imaging device Download PDFInfo
- Publication number
- US10026178B2 US10026178B2 US14/672,565 US201514672565A US10026178B2 US 10026178 B2 US10026178 B2 US 10026178B2 US 201514672565 A US201514672565 A US 201514672565A US 10026178 B2 US10026178 B2 US 10026178B2
- Authority
- US
- United States
- Prior art keywords
- distribution
- imaging device
- image data
- reference direction
- attachment state
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G06T7/0042—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/95—Computational photography systems, e.g. light-field imaging systems
-
- H04N5/23229—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20068—Projection on vertical or horizontal image axis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20224—Image subtraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30244—Camera pose
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
Definitions
- the embodiment discussed herein is related to a method of determining normality of imaging direction.
- a method of determining normality of an imaging direction executed by a computer including: acquiring captured images from an imaging device mounted on a vehicle; and determining the normality of the imaging direction of the imaging device in accordance with whether an appearance pattern of a combination of a portion in which a variation is detected and a portion in which a variation is not detected among the acquired captured images corresponds to a predetermined pattern.
- FIG. 1 is a functional block diagram illustrating the configuration of an image processing apparatus according to an embodiment of the present invention
- FIG. 2 is a schematic diagram illustrating an example of the data structure of drive recorder information
- FIG. 3 is a flowchart illustrating the flow of a process of acquiring image data on an image captured by a camera during driving on a straight line;
- FIG. 4 is a flowchart illustrating the flow of a process of adding a difference image
- FIG. 5 is a flowchart illustrating the flow of a process of calculating a horizontal projection value of a mean difference image
- FIG. 6 is a schematic diagram (No. 1) for giving a supplementary explanation of a process performed by an evaluating unit
- FIG. 7 is a schematic diagram (No. 2) for giving a supplementary explanation of a process performed by the evaluating unit
- FIG. 8 is a flowchart illustrating the flow of a process performed by the image processing apparatus according to the embodiment.
- FIG. 9 is a block diagram illustrating an example of a computer that executes an imaging device attachment state evaluation program.
- the conventional technologies are considered on the basis of the assumption that image data corresponding to an object to be analyzed is image data captured by an imaging device that is normally arranged in a vehicle. Consequently, if an analysis is performed on image data captured by an imaging device that is not normally arranged, an effective analysis result is not obtained.
- FIG. 1 is a functional block diagram illustrating the configuration of an image processing apparatus according to an embodiment of the present invention.
- An image processing apparatus is an example of an imaging device attachment state evaluation device.
- an image processing apparatus 100 includes a communication unit 110 , an input unit 120 , a displaying unit 130 , a storing unit 140 , and a control unit 150 .
- the communication unit 110 is a processing unit that performs data communication with another device via a network.
- the communication unit 110 corresponds to a communication device or the like.
- the input unit 120 is an input device that inputs various kinds of data to the image processing apparatus 100 .
- the input unit 120 corresponds to a keyboard, a mouse, a touch panel, or the like.
- the displaying unit 130 is a display device that displays data output from the control unit 150 .
- the displaying unit 130 corresponds to a liquid crystal display, a touch panel, or the like.
- the storing unit 140 is a storing unit that stores therein drive recorder information 141 .
- the storing unit 140 corresponds to a storage device, such as a semiconductor memory device including, for example, a random access memory (RAM), a read only memory (ROM), a flash memory, or the like.
- RAM random access memory
- ROM read only memory
- flash memory or the like.
- the drive recorder information 141 includes various kinds of data recorded by a drive recorder.
- FIG. 2 is a schematic diagram illustrating an example of the data structure of drive recorder information. As illustrated in FIG. 2 , the drive recorder information 141 stores therein, in an associated manner, a frame number, the date and time, the speed, the acceleration, the position coordinates, and an image.
- the frame number is the number that is used to uniquely identify a frame.
- the date and time is date and time at which a subject frame is captured.
- the speed is a speed of a vehicle having mounted thereon a drive recorder at the time of capturing the subject frame.
- the acceleration is an acceleration of a vehicle having mounted thereon a drive recorder at the time of capturing the subject frame.
- the position coordinates are the position coordinates of a vehicle having mounted thereon a drive recorder at the time of capturing the subject frame.
- an image is image data in a subject frame.
- the control unit 150 includes an acquiring unit 151 , a calculating unit 152 , and an evaluating unit 153 .
- the control unit 150 corresponds to, for example, an integrated circuit, such as an application specific integrated circuit (ASIC), a field programmable gate array (FPGA), or the like.
- the control unit 150 corresponds to, for example, an electronic circuit, such as a central processing unit (CPU), a micro processing unit (MPU), or the like.
- the acquiring unit 151 is a processing unit that acquires drive recorder information from a drive recorder or the like.
- the acquiring unit 151 registers, in the storing unit 140 , drive recorder information acquired from the drive recorder or the like as the drive recorder information 141 .
- the acquiring unit 151 may also sequentially acquire image data from an imaging device, such as a camera or the like, and then sequentially register the acquired image data in the drive recorder information 141 .
- the calculating unit 152 is a processing unit that creates, on the basis of the drive recorder information 141 , a mean difference image that is the average of difference images of adjacent frames.
- the calculating unit 152 outputs information on the calculated mean difference image to the evaluating unit 153 .
- the calculating unit 152 performs a process of acquiring image data captured by a camera when a vehicle is driving on a straight line, performs a process of adding difference images, and performs a process of creating a mean difference image.
- the calculating unit 152 accesses the drive recorder information 141 , specifies image data in a frame in which the speed is equal to or greater than a first threshold and the acceleration is less than a second threshold, and acquires the specified image data.
- the image data in the frame specified by the calculating unit 152 corresponds to the image data captured by a camera during driving on a straight line.
- FIG. 3 is a flowchart illustrating the flow of a process of acquiring image data on an image captured by a camera during driving on a straight line.
- the calculating unit 152 sets the value of i to zero (Step S 10 ).
- the calculating unit 152 acquires the speed and the acceleration of the frame number of [i] (Step S 11 ).
- the calculating unit 152 adds image data with the frame number of [i] to data A (Step S 12 ).
- the calculating unit 152 determines whether the value of i is less than the maximum frame number (Step S 13 ). If the value of i is not less than the maximum frame number (No at Step S 13 ), the calculating unit 152 ends the process of acquiring the image data captured by the camera during the driving.
- Step S 13 the calculating unit 152 updates the value of i to the value that is obtained by adding 1 to the value of i (Step S 14 ) and then proceeds to Step S 11 .
- the calculating unit 152 calculates a difference image by calculating a difference between two pieces of consecutive image data stored in the data A by using the process described in FIG. 3 and then adds each of the difference images.
- FIG. 4 is a flowchart illustrating the flow of a process of adding a difference image.
- the calculating unit 152 sets the value of j to zero, sets the value of “count” to zero, and sets the value of IMG_SUM to zero (Step S 20 ).
- the calculating unit 152 determines whether the frame number of data A [j] is the same as the value that is obtained by subtracting one from the frame number of data A [j+1] (Step S 21 ). If the frame number of data A [j] does not match the value obtained by subtracting one from the frame number of data A [j+1] (No at Step S 21 ), the calculating unit 152 proceeds to Step S 25 .
- Step S 21 the calculating unit 152 proceeds to Step S 22 .
- the calculating unit 152 acquires image data (IMG 1 ) in the frame number of data A[j] and image data (IMG 2 ) in the frame number of data A [j+1] (Step S 22 ).
- the calculating unit 152 adds the absolute value “abs (IMG 1 -IMG 2 )” of the difference value between IMG 1 and IMG 2 to the total value “IMG_SUM” of the difference images, thereby the calculating unit 152 updates the total value “IMG_SUM” of the difference images (Step S 23 ).
- the calculating unit 152 updates the value of a “count” to the value that is obtained by adding one to the value of the “count” (Step S 24 ).
- the calculating unit 152 determines whether j is less than the number of pieces of data in the data A (Step S 25 ). If j is not less than the number of pieces of data in the data A (No at Step S 25 ), the calculating unit 152 ends the process of adding the difference images.
- Step S 25 If the value of j is less than the number of pieces of data in the data A (Yes at Step S 25 ), the calculating unit 152 updates the value of j to the value obtained by adding one to the value of j (Step S 26 ) and proceeds to Step S 21 .
- the calculating unit 152 creates a mean difference image by dividing total value “IMG_SUM” of difference images described with reference to FIG. 4 by the value of “count” that is the number of pieces of data A. The number of pieces of the data A corresponds to the number of difference images.
- the evaluating unit 153 is a processing unit that calculates a horizontal projection value of a mean difference image and that evaluates, on the basis of distribution of horizontal projection values, the attachment state of the camera.
- the evaluating unit 153 corresponds to a determining unit and an evaluating unit.
- the evaluating unit 153 calculates a horizontal projection value of a mean difference image. For the location of each of the pixels in a mean difference image in the y coordinate, the evaluating unit 153 calculates a horizontal projection value by integrating the luminance in the x direction.
- FIG. 5 is a flowchart illustrating the flow of a process of calculating a horizontal projection value of a mean difference image.
- the evaluating unit 153 sets the value of k to zero (Step S 30 ).
- the symbol “k” represents the location of a pixel in a mean difference image in the y coordinate and varies from 0 to H ⁇ 1.
- the symbol “H” indicates the height of a mean difference image.
- Hist [k] indicates the result obtained, at “k” in the mean difference image in the y coordinate, by adding luminance of mean difference images in the x-axis direction.
- W indicates the width of a mean difference image.
- the evaluating unit 153 determines whether k is less than H (Step S 32 ). If k is less than H (Yes at Step S 32 ), the evaluating unit 153 updates the value of k to the value that is obtained by adding one to the value of k (Step S 33 ) and proceeds to Step S 31 .
- the evaluating unit 153 ends the process of calculating a horizontal projection value of the mean difference image.
- the evaluating unit 153 performs the process illustrated in FIG. 5 , thereby a horizontal projection value in the mean difference image is calculated.
- the evaluating unit 153 evaluates the attachment state of a camera on the basis of a horizontal projection value.
- the evaluating unit 153 performs a first evaluation process and a second evaluation process.
- the evaluating unit 153 averages the horizontal projection values for the y coordinate and determines whether the average value of the horizontal projection values is less than a third threshold. If the average value of the horizontal projection values is less than the third threshold, the evaluating unit 153 determines that the attachment state of the camera is abnormal and then outputs the determination result.
- the evaluating unit 153 determines that the attachment state of the camera is abnormal. Average (Hist[0],Hist[1], . . . ,Hist[ H ⁇ 1]) ⁇ third threshold (2)
- FIG. 6 is a schematic diagram (No. 1) for giving a supplementary explanation of a process performed by an evaluating unit.
- reference numeral 10 a represents image data of a certain frame number.
- Reference numeral 10 b represents a mean difference image that is calculated from multiple pieces of image data including the image data 10 a .
- Reference numeral 11 represents horizontal projection values calculated on the basis of the mean difference image 10 b .
- the vertical axis of the mean difference image 10 b is defined to be the y axis and the horizontal axis is defined to be the X-axis.
- the evaluating unit 153 calculates the horizontal projection values 11 by performing additions of luminance in the x-axis direction. If the average value of the horizontal projection values 11 is less than the third threshold, the evaluating unit 153 determines that the attachment state of the camera is abnormal.
- the evaluating unit 153 calculates the average value of the horizontal projection values 11 , the evaluating unit 153 calculates the average value of the values between Hist [0] and Hist [H ⁇ 1]; however, the calculation is not limited thereto.
- the evaluating unit 153 may also calculate the average value of the horizontal projection values near the center and then compare the value with the third threshold. In this way, by limiting the area that is used to calculate the average value of the horizontal projection values, an amount of calculation can be reduced.
- the evaluating unit 153 compares a value at the lower portion of the horizontal projection value with a value at the upper portion of the horizontal projection value. If the value at the upper portion of the horizontal projection value is greater than the value at the lower portion of the horizontal projection value, the evaluating unit 153 determines that the attachment state of the camera is abnormal and outputs the determination result.
- the evaluating unit 153 determines that the attachment state of the camera is abnormal.
- a value smaller than that of “H/2” is set for “N” in Equations (3) and (4).
- Average (Hist [0], . . . , Hist [N]) is the average value of the values at the lower portion of the horizontal projection value.
- Average (Hist [H ⁇ 1 ⁇ N], . . . , Hist [H ⁇ 1]) is the average value of the values at the upper portion of the horizontal projection value.
- Equations (3) and (4) are satisfied, this indicates that a variation is great in the upper portion of the screen data captured by a camera and a variation is small in the lower portion of the screen data.
- This feature is an inverse of the feature of the image data captured by a camera that is normally attached. In the image data captured by a camera that is normally attached, because a variation in screen data in the lower portion in which another vehicle or the like is included becomes great, for a horizontal projection value, the value of the lower portion is greater than that of the upper portion.
- FIG. 7 is a schematic diagram (No. 2) for giving a supplementary explanation of a process performed by the evaluating unit.
- reference numeral 20 a represents image data with a certain frame number.
- Reference numeral 20 b represents a mean difference image calculated from multiple pieces of image data including the image data 20 a .
- Reference numeral 21 represents the horizontal projection values calculated on the basis of the mean difference image 20 b .
- the vertical axis of the mean difference image 20 b is defined to be the y-axis and the horizontal axis of the mean difference image 20 b is defined to be the X-axis.
- the evaluating unit 153 calculates the horizontal projection values 21 by performing additions of luminance in the x-axis direction.
- reference numeral 21 a is defined to be the lower portion of the horizontal projection values and reference numeral 21 b is defined to be the upper portion of the horizontal projection values.
- the evaluating unit 153 determines that the attachment state of a camera is abnormal.
- FIG. 8 is a flowchart illustrating the flow of a process performed by the image processing apparatus according to the embodiment.
- the image processing apparatus 100 acquires image data in which the speed is equal to or greater than the first threshold and the acceleration is less than the second threshold (Step S 101 ).
- the calculating unit 152 in the image processing apparatus 100 creates a difference image (Step S 102 ).
- the calculating unit 152 creates a mean difference image (Step S 103 ).
- the evaluating unit 153 in the image processing apparatus 100 calculates the horizontal projection values by adding the luminance of the mean difference images in the x direction (Step S 104 ).
- the evaluating unit 153 determines whether the average value of the horizontal projection values is less than the third threshold (Step S 105 ).
- Step S 106 If the average value of the horizontal projection values is less than the third threshold (Yes at Step S 105 ), the evaluating unit 153 determines that the attachment is abnormal (Step S 106 ).
- the evaluating unit 153 determines whether the horizontal projection values in the upper portion are greater than the horizontal projection values in the lower portion (Step S 107 ). If the horizontal projection values in the upper portion are greater than the horizontal projection values in the lower portion (Yes at Step S 107 ), the evaluating unit 153 proceeds to Step S 106 .
- the evaluating unit 153 determines that the attachment is not abnormal (Step S 108 ).
- the image processing apparatus 100 determines the normality of the imaging direction of the camera in accordance with whether the appearance pattern of a combination of a portion in which a variation is detected and a portion in which a variation is not detected correspond to a predetermined pattern. Therefore, it is possible to accurately determine whether the imaging direction of a camera is normal.
- the image processing apparatus 100 calculates horizontal projection values and evaluates the attachment state of the camera on the basis of the distribution of these horizontal projection values. Therefore, it is possible to accurately determine whether the imaging direction of a camera is normal.
- the image processing apparatus 100 evaluates the attachment state of a camera on the basis of the distribution of the horizontal projection values in an area at the end side. Therefore, by narrowing down the area in which integrated values are evaluated, it is possible to reduce the calculation targets and thus the processing load can be reduced. Furthermore, the distribution in the vicinity of the center of the horizontal projection values varies when compared with the area at the end side regardless of whether the camera is normally attached. Therefore, by using the portion on the end sides, it is possible to more accurately determine whether the imaging direction of the camera is appropriate.
- the image processing apparatus 100 calculates the average value of the horizontal projection values near the center, compares the average value with the third threshold, and evaluates the normality of the imaging direction of the camera, the image processing apparatus 100 can limit the area that is used to obtain the average value of the horizontal projection values and thus the calculation amount can be reduced.
- the attachment state of a camera is analyzed by acquiring the drive recorder information 141 from a drive recorder and then analyzing the drive recorder information 141 ; however, the method is not limited thereto.
- the image processing apparatus 100 may also be embedded in a drive recorder, image data may also be acquired from a camera in real time, the normality of the camera in an imaging direction may also be determined, and the determination result may also be sent to a driver as a notification.
- FIG. 9 is a block diagram illustrating an example of a computer that executes an imaging device attachment state evaluation program.
- a computer 200 includes a CPU 201 that executes various arithmetic processing, an input device 202 that receives an input of data from a user, and a display 203 . Furthermore, the computer 200 includes a reading device 204 that reads a program or the like from a storage medium and an interface device 205 that exchanges data with another computer via a network. Furthermore, the computer 200 includes a RAM 206 and a hard disk device 207 that temporarily store therein various kinds of information. Then, each of the devices 201 to 207 is connected to a bus 208 .
- the hard disk device 207 includes an acquisition program 207 a , a calculating program 207 b , and an evaluation program 207 c .
- the CPU 201 reads the acquisition program 207 a , the calculating program 207 b , and the evaluation program 207 c and loads the programs in the RAM 206 .
- the acquisition program 207 a functions as an acquisition process 206 a .
- the calculating program 207 b functions as a calculating process 206 b .
- the evaluation program 207 c functions as an evaluation process 206 c.
- the acquisition process 206 a corresponds to the acquiring unit 151 .
- the calculating process 206 b corresponds to the calculating unit 152 .
- the evaluation process 206 c corresponds to the evaluating unit 153 .
- the acquisition program 207 a , the calculating program 207 b , and the evaluation program 207 c do not need to be stored in the hard disk device 207 from the beginning.
- the programs are stored in a “portable physical medium”, such as a flexible disk (FD), a CD-ROM, a DVD disk, a magneto-optic disk, an IC CARD, or the like that is to be inserted into the computer 200 .
- the computer 200 may also read and execute the acquisition program 207 a , the calculating program 207 b , and the evaluation program 207 c from the portable physical medium.
- an advantage is provided in that the normality of an imaging direction of an imaging device can be evaluated.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computing Systems (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- Image Processing (AREA)
- Closed-Circuit Television Systems (AREA)
- Traffic Control Systems (AREA)
- Image Analysis (AREA)
Abstract
Description
IMG_AVE=IMG_SUM/count (1)
Average (Hist[0],Hist[1], . . . ,Hist[H−1])<third threshold (2)
Average (Hist[0], . . . ,Hist[N])<fourth threshold (3)
Average (Hist[0], . . . ,Hist[N])+fifth threshold<Average (Hist[H−1−N], . . . ,Hist[H−1]) (4)
Claims (8)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2014086682A JP6264173B2 (en) | 2014-04-18 | 2014-04-18 | Normality determination method for imaging direction, imaging device attachment state evaluation program, and imaging device attachment state evaluation device |
| JP2014-086682 | 2014-04-18 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20150302588A1 US20150302588A1 (en) | 2015-10-22 |
| US10026178B2 true US10026178B2 (en) | 2018-07-17 |
Family
ID=54322440
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/672,565 Active 2036-02-22 US10026178B2 (en) | 2014-04-18 | 2015-03-30 | Method of determining normality of imaging direction, and evaluation device of evaluating attachment state of imaging device |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US10026178B2 (en) |
| JP (1) | JP6264173B2 (en) |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP6299373B2 (en) | 2014-04-18 | 2018-03-28 | 富士通株式会社 | Imaging direction normality determination method, imaging direction normality determination program, and imaging direction normality determination apparatus |
| JP6299371B2 (en) | 2014-04-18 | 2018-03-28 | 富士通株式会社 | Imaging direction inclination detection method, imaging direction inclination detection program, and imaging direction inclination detection apparatus |
Citations (19)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH08285534A (en) | 1995-04-12 | 1996-11-01 | Suzuki Motor Corp | In-vehicle image processing device |
| US20030202694A1 (en) * | 2002-04-25 | 2003-10-30 | Fujitsu Ten Limited | Diagnostic apparatus for abnormalities in image recognition system |
| JP2004239613A (en) | 2003-02-03 | 2004-08-26 | Pioneer Electronic Corp | Installation angle detector |
| US20040201483A1 (en) * | 2003-02-21 | 2004-10-14 | Stam Joseph S. | Automatic vehicle exterior light control systems |
| US6963661B1 (en) | 1999-09-09 | 2005-11-08 | Kabushiki Kaisha Toshiba | Obstacle detection system and method therefor |
| US6993159B1 (en) | 1999-09-20 | 2006-01-31 | Matsushita Electric Industrial Co., Ltd. | Driving support system |
| JP2008123036A (en) | 2006-11-08 | 2008-05-29 | Nec Corp | Vanishing point detection system, vanishing point detection method and program for vanishing point detection |
| WO2009001510A1 (en) | 2007-06-28 | 2008-12-31 | Panasonic Corporation | Image processing device, image processing method, and program |
| US20100054542A1 (en) * | 2008-09-03 | 2010-03-04 | Texas Instruments Incorporated | Processing video frames with the same content but with luminance variations across frames |
| JP2010183294A (en) | 2009-02-04 | 2010-08-19 | Aisin Seiki Co Ltd | Image processing apparatus, method of processing image, and program |
| US20110311156A1 (en) * | 2010-06-22 | 2011-12-22 | Hynix Semiconductor Inc. | Method and system for correcting lens shading |
| US8594370B2 (en) | 2004-07-26 | 2013-11-26 | Automotive Systems Laboratory, Inc. | Vulnerable road user protection system |
| US20130321630A1 (en) | 2012-06-05 | 2013-12-05 | Hyundai Mobis Co., Ltd. | System and method for lane departure warning |
| WO2014033936A1 (en) | 2012-08-31 | 2014-03-06 | 富士通株式会社 | Image processing device, image processing method, and image processing program |
| WO2014045344A1 (en) | 2012-09-18 | 2014-03-27 | トヨタ自動車株式会社 | Foe setting device and foe setting method for on-vehicle camera |
| US8687070B2 (en) | 2009-12-22 | 2014-04-01 | Apple Inc. | Image capture device having tilt and/or perspective correction |
| US20150169970A1 (en) * | 2012-08-30 | 2015-06-18 | Fujitsu Limited | Image processing apparatus and image processing method |
| US20150304650A1 (en) | 2014-04-18 | 2015-10-22 | Fujitsu Limited | Determination method and determination apparatus |
| US20150304651A1 (en) | 2014-04-18 | 2015-10-22 | Fujitsu Limited | Method of detecting tilt of image capturing direction and apparatus of detecting tilt of image capturing direction |
-
2014
- 2014-04-18 JP JP2014086682A patent/JP6264173B2/en active Active
-
2015
- 2015-03-30 US US14/672,565 patent/US10026178B2/en active Active
Patent Citations (23)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH08285534A (en) | 1995-04-12 | 1996-11-01 | Suzuki Motor Corp | In-vehicle image processing device |
| US6963661B1 (en) | 1999-09-09 | 2005-11-08 | Kabushiki Kaisha Toshiba | Obstacle detection system and method therefor |
| US6993159B1 (en) | 1999-09-20 | 2006-01-31 | Matsushita Electric Industrial Co., Ltd. | Driving support system |
| US20030202694A1 (en) * | 2002-04-25 | 2003-10-30 | Fujitsu Ten Limited | Diagnostic apparatus for abnormalities in image recognition system |
| JP2004239613A (en) | 2003-02-03 | 2004-08-26 | Pioneer Electronic Corp | Installation angle detector |
| US20040172173A1 (en) | 2003-02-03 | 2004-09-02 | Pioneer Corporation | Mounting angle detection device |
| US20040201483A1 (en) * | 2003-02-21 | 2004-10-14 | Stam Joseph S. | Automatic vehicle exterior light control systems |
| US8594370B2 (en) | 2004-07-26 | 2013-11-26 | Automotive Systems Laboratory, Inc. | Vulnerable road user protection system |
| JP2008123036A (en) | 2006-11-08 | 2008-05-29 | Nec Corp | Vanishing point detection system, vanishing point detection method and program for vanishing point detection |
| US20100080467A1 (en) | 2006-11-08 | 2010-04-01 | Kazuyuki Sakurai | Vanishing point detecting system, vanishing point detecting method, and vanishing point detecting program |
| WO2009001510A1 (en) | 2007-06-28 | 2008-12-31 | Panasonic Corporation | Image processing device, image processing method, and program |
| US20090290809A1 (en) | 2007-06-28 | 2009-11-26 | Hitoshi Yamada | Image processing device, image processing method, and program |
| US20100054542A1 (en) * | 2008-09-03 | 2010-03-04 | Texas Instruments Incorporated | Processing video frames with the same content but with luminance variations across frames |
| JP2010183294A (en) | 2009-02-04 | 2010-08-19 | Aisin Seiki Co Ltd | Image processing apparatus, method of processing image, and program |
| US8687070B2 (en) | 2009-12-22 | 2014-04-01 | Apple Inc. | Image capture device having tilt and/or perspective correction |
| US20110311156A1 (en) * | 2010-06-22 | 2011-12-22 | Hynix Semiconductor Inc. | Method and system for correcting lens shading |
| US20130321630A1 (en) | 2012-06-05 | 2013-12-05 | Hyundai Mobis Co., Ltd. | System and method for lane departure warning |
| US20150169970A1 (en) * | 2012-08-30 | 2015-06-18 | Fujitsu Limited | Image processing apparatus and image processing method |
| WO2014033936A1 (en) | 2012-08-31 | 2014-03-06 | 富士通株式会社 | Image processing device, image processing method, and image processing program |
| US20150154460A1 (en) | 2012-08-31 | 2015-06-04 | Fujitsu Limited | Image processing apparatus and image processing method |
| WO2014045344A1 (en) | 2012-09-18 | 2014-03-27 | トヨタ自動車株式会社 | Foe setting device and foe setting method for on-vehicle camera |
| US20150304650A1 (en) | 2014-04-18 | 2015-10-22 | Fujitsu Limited | Determination method and determination apparatus |
| US20150304651A1 (en) | 2014-04-18 | 2015-10-22 | Fujitsu Limited | Method of detecting tilt of image capturing direction and apparatus of detecting tilt of image capturing direction |
Non-Patent Citations (12)
| Title |
|---|
| Japanese Office Action dated Nov. 21, 2017 for corresponding Japanese Application No. 2014-086526, with English Translation, 5 pages. |
| Japanese Office Action dated Sep. 12, 2017 for corresponding Japanese Patent Application No. 2014-086682, with English Translation, 7 pages. |
| Japanese Office Action dated Sep. 12, 2017 for corresponding Japanese Patent Application No. 2014-086771, with English Translation, 6 pages. |
| Office Action dated Jun. 12, 2017 for copending U.S. Appl. No. 14/666,552, 31 pages. |
| Restriction Requirement dated Apr. 4, 2017 for copending U.S. Appl. No. 14/662,777, 6 pages. |
| Restriction Requirement dated Jan. 17, 2017 for copending U.S. Appl. No. 14/666,552, 5 pages. |
| U.S. Non-Final Office Action dated May 19, 2016 for copending U.S. Appl. No. 14/666,552 , 25 pages. |
| U.S. Notice of Allowance dated Apr. 12, 2018 for copending U.S. Appl. No. 14/662,777, 25 pages. |
| U.S. Office Action dated May 3, 2018 for copending U.S. Appl. No. 14/666,552, 39 pages. |
| U.S. Office Action dated Sep. 15, 2017 for co-pending U.S. Appl. No. 14/662,777, 32 pages. |
| U.S. Office Action dated Sep. 28, 2017 for co-pending U.S. Appl. No. 14/662,552, 25 pages. |
| WO 2014033885 A1, Baba et al, Mar. 2014, Image Processing device, image processing method, and image processing program. * |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2015207879A (en) | 2015-11-19 |
| US20150302588A1 (en) | 2015-10-22 |
| JP6264173B2 (en) | 2018-01-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10616570B2 (en) | Determination method and determination apparatus | |
| US9454704B2 (en) | Apparatus and method for determining monitoring object region in image | |
| US20150146006A1 (en) | Display control apparatus and display control method | |
| EP3379509A1 (en) | Apparatus, method, and image processing device for smoke detection | |
| US10909365B2 (en) | Image processing apparatus, control method thereof, and non-transitory computer-readable storage medium | |
| US11036974B2 (en) | Image processing apparatus, image processing method, and storage medium | |
| US20170287445A1 (en) | Information display apparatus, information display system, control method thereof, and program | |
| CN110706229B (en) | Method and device for determining laser stripe center, electronic equipment and storage medium | |
| US10230946B2 (en) | Method of detecting tilt of image capturing direction and apparatus of detecting tilt of image capturing direction | |
| CN113362238B (en) | Test image processing method, device, electronic equipment and storage medium | |
| US20220180529A1 (en) | Image processing apparatus, image processing method, and non-transitory computer-readable medium | |
| US10026178B2 (en) | Method of determining normality of imaging direction, and evaluation device of evaluating attachment state of imaging device | |
| US9639763B2 (en) | Image target detecting apparatus and method | |
| JP2021043141A (en) | Object distance estimation device and object distance estimation method | |
| US20140126827A1 (en) | Motion image region identification device and method thereof | |
| KR20210116247A (en) | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium | |
| JP6806066B2 (en) | Object counting device, object counting method, object counting program, and object counting system | |
| KR101117289B1 (en) | Method and device for estimating touch pressure using image, recording medium for the same, and mobile device comprising the same | |
| JP6677711B2 (en) | Image processing apparatus, image processing method, and program | |
| CN105243062B (en) | Method and device for detecting webpage feature area | |
| CN113286086A (en) | Camera use control method and device, electronic equipment and storage medium | |
| JPWO2018179119A1 (en) | Video analysis device, video analysis method, and program | |
| JP6897448B2 (en) | Line width estimation programs, devices, and methods | |
| JP2024091918A (en) | Line detection device, line detection method, program, and storage medium | |
| JP2024032396A (en) | Information processing device, information processing method and program |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: FUJITSU LIMITED, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BABA, KOZO;HASHIGUCHI, NORIO;SIGNING DATES FROM 20150321 TO 20150323;REEL/FRAME:035289/0760 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |