WO2012124055A1 - 撮像装置、撮像方法及び撮像プログラム - Google Patents
撮像装置、撮像方法及び撮像プログラム Download PDFInfo
- Publication number
- WO2012124055A1 WO2012124055A1 PCT/JP2011/056031 JP2011056031W WO2012124055A1 WO 2012124055 A1 WO2012124055 A1 WO 2012124055A1 JP 2011056031 W JP2011056031 W JP 2011056031W WO 2012124055 A1 WO2012124055 A1 WO 2012124055A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- imaging
- unit
- distance
- image
- imaging target
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/14—Vascular patterns
- G06V40/145—Sensors therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10141—Special mode during image acquisition
- G06T2207/10152—Varying illumination
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20224—Image subtraction
Definitions
- the present invention relates to an imaging apparatus, an imaging method, and an imaging program.
- an imaging device that performs biometric authentication reads a palm vein that is imaged in a non-contact manner.
- Such an imaging device has an imaging target at a predetermined distance from the imaging camera of the imaging device. When located in the range, it is possible to perform imaging normally. For this reason, the imaging apparatus needs to detect the distance to the imaging target.
- the imaging target is imaged while spot light is projected onto the imaging target, and imaging is performed according to the spot light of the captured image.
- a method for detecting a distance of an imaging target with respect to a camera see, for example, Patent Document 1).
- the present invention has been made in view of such a point, and an object thereof is to provide an imaging apparatus, an imaging method, and an imaging program capable of determining that an imaging target is close to the imaging apparatus.
- a first light emitting unit that emits light when measuring the distance of the imaging target and illuminates the side of the imaging target that is imaged by the imaging unit, and light emission when the imaging target is imaged Then, a second light emitting unit that illuminates the image capturing side of the imaging target of the imaging target, a distance measuring unit that measures a distance from the imaging target by light emission of the first light emitting unit, and the distance measuring unit
- the imaging target is close to the imaging unit based on the bright area of the captured image of the imaging target captured by the imaging unit with the second light emitting unit emitting light.
- An image pickup apparatus is provided that includes a determination unit that determines whether the image is being processed.
- FIG. 1 is a conceptual diagram for explaining the imaging apparatus according to the first embodiment.
- FIG. 1A is a side view of the imaging device 1 that images the imaging target 5 located at heights (a) to (d), and FIG. 1B is an upper plan view of the imaging device 1. ing.
- the imaging device 1 can determine that the imaging target 5 is located in the vicinity of the imaging surface 1a of the imaging device 1.
- such an imaging apparatus 1 includes an imaging unit 2, light emitting units 3 a and 3 b, and a determination unit 4.
- the imaging unit 2 images the imaging target 5.
- the imaging unit 2 includes an imaging camera 2a and an optical system 2b, and captures an image of the imaging target 5 with the imaging camera 2a.
- the light emitting unit 3a is parallel (or substantially parallel) to the optical axis of the imaging camera 2a of the imaging unit 2 when the distance measuring unit 6 described later measures the distance from the imaging surface 1a of the imaging device 1 to the imaging target 5. Then, the spot light, which is light concentrated at one point, is projected onto the imaging target 5.
- the imaging device 1 compares the interval between the center of the captured image of the imaging target 5 imaged by the imaging unit 2 and the spot light from the light emitting unit 3b and the interval between the center of the imaging unit 2 and the light emitting unit 3a. By doing so, the distance from the imaging surface 1a to the imaging target 5 can be measured. In addition, the distance from the imaging surface 1a of the imaging device 1 to the imaging target 5 will be described later.
- the light emitting unit 3b emits light to illuminate the side of the imaging target 5 that is imaged by the imaging unit 2. In addition, the light emission part 3b illuminates the imaging object 5 as needed, and does not always illuminate.
- the distance measuring unit 6 measures the distance from the imaging target 5 by the light emission of the light emitting unit 3a.
- the determination unit 4 determines that the imaging target 5 is close to the imaging unit 2 based on the bright area of the captured image of the imaging target 5 captured by the imaging unit 2 with the light emitting unit 3b emitting light.
- the bright region is a region having a predetermined lightness (or luminance) or more, such as a set of lightness (or luminance) of 69 or more of 256 gradations.
- FIG. 2 is a diagram illustrating an example of a captured image corresponding to the distance to the imaging device to be imaged.
- 2A to 2C respectively represent captured images g obtained by capturing the imaging target 5 located at the distances (a) to (c) in FIG. 1A.
- the captured image g when the imaging target 5 is located at the distance (a) in FIG. 1A is, together with the imaging target 5 at the center, as shown in FIG. Spot lights 3a1s to 3a4s projected from 3a1 to 3a4 are reflected.
- the captured image g corresponding to each distance expands the imaging target 5, and increases the interval between the spot lights 3a1s to 3a4s, so that the position moves outward.
- FIG. 3 is a diagram for explaining the detection of the distance from the imaging apparatus to the imaging target.
- FIG. 3A shows a side view of the imaging camera 2a and the optical system 2b constituting the imaging unit 2 and the imaging target 5, and FIG. 3B shows a captured image g of the imaging target 5. ing.
- FIG. 3 only the spot light 3a2s by the light emitting unit 3a2 and the light emitting unit 3a2 is taken as an example.
- the imaging unit 2 captures the captured image 5 in which the spot light 3a2s is projected from the light emitting unit 3a2 to the imaging target 5.
- the imaging range from the position O where the optical axis of the imaging unit 2 and the imaging target 5 are orthogonal is represented by w.
- the distance from the imaging unit 2 to the light emitting unit 3b2 is d
- the position of the spot light 3a2s of the imaging target 5 from the position O is d + x.
- x represents the amount of variation caused by the optical axis shift ⁇ 1 of the spot light from the light emitting portion 3a2.
- the captured image g captured by the imaging unit 2 has a distance a from the center (position O) of the imaging target 5 to the spot light 3a2s, and the moving direction of the spot light 3a2s.
- the distance origin when determining the known distance L that is, the imaging camera of the imaging unit 2 for obtaining the known distance L
- tan ⁇ 1 is calculated using two calibration measurement points.
- tan ⁇ 1 is not calculated based only on the distance L from the lens of the imaging camera 2a to the imaging target 5, but is calculated based on the distance (difference) between two calibration measurement points.
- the imaging object 5 is placed at each of two calibration measurement points for which the distance difference ⁇ has been determined strictly in advance, the captured image g is captured by the imaging camera 2a, and the distance from the captured captured image g to the imaging object 5 is captured.
- the optical axis deviation ⁇ 1 can be calibrated.
- the position a1 ′ of the spot light in the captured image captured at the calibration measurement point a1 of the known distance L is measured, and further, the image captured at the calibration measurement point a2 that is separated from the calibration measurement point a1 by the distance ⁇ .
- the following formula (5) can be obtained by measuring the position a2 ′ of the spot light in the image and substituting each into formula (4).
- FIG. 4 is a diagram for explaining detection of the tilt state of the imaging target with respect to the imaging device.
- the coordinate in the Y direction is defined in the depth direction of the paper surface, where the horizontal direction of the paper surface is the X direction and the vertical direction of the paper surface is the Z direction.
- the positions of the light emitting units 3a1 to 3a4 and the spot lights 3a1s to 3a4s are represented by dots with respect to the coordinates.
- the light emitting unit 3b3 is positioned at the origin, the light emitting unit 3a4 is set to (n, 0, 0).
- the light emitting part 3a1 is located at (0, n, 0) and the light emitting part 3a2 is located at (n, n, 0).
- the distance measuring unit 6 can calculate the distances L1 to L4 to the spot lights 3a1s to 3a4s using, for example, the spot lights 3a1s to 3a4s projected from the light emitting units 3a1 to 3a4. .
- the surface of the imaging target 5 that faces the imaging camera 2a of the imaging unit 2 can be regarded as a plane or a plane
- the coefficients a, b, c, and d can be calculated, and the normal vector is determined from these coefficients, and the imaging target 5 tilt and posture information can be detected.
- the plane is divided by a straight line passing through the spot lights 3a4s and 3a1s, and the plane passing through the spot lights 3a4s, 3a3s and 3a1s and the plane passing through the spot lights 3a2s, 3a4s and 3a1s are divided.
- a plane can be obtained.
- the imaging object 5 is within an appropriate imaging distance (for example, 5 to 100 mm) with respect to the imaging surface 1a of the imaging device 1 so that the spot light is reflected in the captured image g. 1 (corresponding to the positions (a) to (c) in FIG. 1A), and the spot light is also reflected in the captured image g. Therefore, the distance measuring unit 6 can calculate the distance L.
- an appropriate imaging distance for example, 5 to 100 mm
- the imaging target 5 moves from the distance (c) to (d) in FIG. 1A along the optical axis of the imaging camera 2a of the imaging unit 2, the spot light 3a2s moves outward and is photographed.
- the spot light 3a2s does not appear in the captured image g.
- the distance measuring unit 6 reaches the imaging target 5.
- the distance L cannot be calculated, and it is not clear whether the imaging target 5 is within the appropriate imaging distance range for the imaging device 1 or whether the imaging target 5 is not present.
- the imaging target 5 for example, the palm
- the imaging device 1 is connected to the light emitting unit 3 a.
- the following processing is executed. The reason why the distance from the imaging device 1 to the imaging target 5 cannot be detected will be described in the second embodiment.
- the imaging unit 2 images the imaging target 5 in a state where the light emitting unit 3b emits light.
- the captured image of the imaging target 5 captured in this way is emitted from the imaging target 5 by the light emission of the light emitting unit 3b when the imaging target 5 is located within an inappropriate imaging distance with respect to the imaging device 1. Since the reflected light becomes stronger than the appropriate distance, the ratio of the bright area is higher than the ratio of the bright area in the captured image at the appropriate distance.
- the imaging target 5 is not arranged on the imaging surface of the imaging device 1 or is separated from the appropriate distance for imaging, the captured image of the imaging target 5 is a bright region in the captured image at the appropriate distance. Since the ratio of the bright area is lower than the ratio of
- the imaging device 1 is configured to capture an image with the imaging unit 2 while emitting light from the light emitting unit 3b. As a result, if the bright area occupies a large amount with respect to the captured image of the captured image 5, it is determined that the imaging object 5 is close to the imaging device 1, and there are many dark areas with respect to the captured image. , It is determined that the imaging target 5 is not set in the imaging apparatus 1 or is separated from an appropriate distance for imaging.
- the imaging target 5 is the palm of the user, the palm is too close to the imaging device 1 for the user, and the palm is appropriate for imaging from the imaging device 1. As a result, the user can be prompted to move away from the distance, and the imaging of the imaging object 5 can be performed smoothly.
- FIG. 5 is a diagram illustrating a perspective view of the imaging apparatus according to the second embodiment.
- the imaging device 10 shown in FIG. 5 is covered with a filter 11 that is held up by the palm of the hand, and the outer case 12 covers the periphery. Further, the exterior case 12 is provided with an external connection port 12a to which a cable or the like for connection with an external device is connected.
- Such an imaging apparatus 10 is used for user authentication, for example, connected to a PC (Personal Computer) or incorporated in an automatic teller machine such as a financial institution.
- FIG. 6 is a diagram for explaining an internal configuration of the imaging apparatus according to the second embodiment.
- 6A is a side view of the imaging device 10
- FIG. 6B is a top view of the imaging device 10 taken along one-dot chain line XX in FIG. 6A.
- the imaging apparatus 10 is first provided with an imaging camera 14 such as a CMOS (Complementary Metal Oxide Semiconductor) image sensor in the center of the camera circuit board 13. Furthermore, an optical unit 15 including a lens optical system such as a condenser lens is provided for the imaging camera 14. Thereby, the imaging camera 14 can image the palm facing the filter 11.
- an imaging camera 14 such as a CMOS (Complementary Metal Oxide Semiconductor) image sensor in the center of the camera circuit board 13.
- an optical unit 15 including a lens optical system such as a condenser lens
- a plurality of light emitting elements 16a for example, emitting near infrared rays, and a light receiving element 16b such as a photodiode are mounted around the imaging camera 14 in a circular shape.
- a ring-shaped light guide 17 is provided above the light emitting element 16a and the light receiving element 16b.
- the light guide 17 is made of, for example, resin, and can guide the light of the light emitting element 16a of the camera circuit board 13 upward and irradiate the object with uniform light.
- each distance measuring light emitting element 18 for measuring the distance from the imaging device 10 to the palm are provided on the diagonal lines of the four corners of the camera circuit board 13. ing.
- the light emitting element 18 projects a spot light on the palm facing the filter 11, and the distance from the filter 11 to the palm of the imaging target using the spot light as described in the first embodiment. Can be measured at four points.
- interval of each light emitting element 18 is arrange
- the imaging device 10 is electrically connected to the circuit unit 21 on the back surface of the camera circuit board 13 and the circuit unit 21 that is electrically connected to each component such as the imaging camera 14 on the camera circuit board 13. And an external connector portion 22.
- the circuit unit 21 includes a central processing unit (CPU) 21a, a read only memory (ROM) (not shown), a random access memory (RAM) 21c, and an input / output interface 21d.
- CPU central processing unit
- ROM read only memory
- RAM random access memory
- the information of the captured image captured by the imaging camera 14 is notified by being electrically connected to each component such as 14, and image processing or the like is performed on the captured image. Further, the circuit unit 21 uses the captured image notified from the imaging camera 14 to measure the distance to the palm and the like.
- the external connector unit 22 is electrically connected to a cable or the like connected to an external device via the external connection port 12a, transmits the processing result of the circuit unit 21 to the external device, and receives a signal from the external device. Can be received.
- FIG. 7 is a diagram illustrating a hardware configuration example of the imaging apparatus according to the second embodiment.
- the circuit unit 21 of the imaging device 10 includes a CPU 21a, a ROM 21b, a RAM 21c, and an input / output interface 21d.
- the CPU 21a includes a ROM 21b, a RAM 21c, an input / output interface 21d, an imaging camera 14, and a light emitting device. Elements 15 and 16 are connected.
- the CPU 21a comprehensively controls the entire imaging apparatus 10 by executing various programs stored in a storage medium such as the ROM 21b.
- the ROM 21b stores an OS (Operating System) and application programs on the imaging apparatus 10.
- the RAM 21c temporarily stores at least a part of the OS and programs to be executed by the CPU 21a.
- the RAM 21c stores various data necessary for processing by the CPU 21a.
- the input / output interface 21d is electrically connected to the external connector unit 22d, notifies the CPU 21a of a signal transmitted from the external device via the external connector unit 22, and transmits a signal transmitted from the CPU 21a to the external device. To do.
- the palm is imaged in a state where spot light is projected from the light emitting element 16a onto the palm, and is reflected in the captured image.
- the distance from the imaging device 10 (filter 11) to the palm can be detected according to the distance between the spot light, the center of the captured image, and the imaging range of the imaging camera 14. Specifically, since spot light is projected from four locations on the palm, the distance from the filter 11 to the four portions of the palm can be detected.
- the appropriate imaging distance that can measure from the imaging device 10 to the palm is in the range of 5 to 100 mm from the imaging device 10 (filter 11), and the optimum distance is around 35 to 75 mm.
- the distance to the palm held over an inappropriate imaging distance range of 0 to 5 mm and 100 mm or more cannot be measured, and measurement is impossible.
- FIG. 8 is a block diagram illustrating functions provided in the imaging apparatus according to the second embodiment.
- the imaging device 10 measures the distance from the imaging device 10 to the palm, and determines whether the palm is located within the range of 0 to 5 mm from the imaging device 10 when measurement is impossible. It is something that can be done.
- Such an imaging apparatus 10 includes an image information holding unit 211, an imaging execution unit 212, an image information reception unit 213, a distance measurement unit 214, an image processing unit 215, an image determination unit 216, and an output request unit 217.
- the image information holding unit 211 holds information such as a captured image captured by the imaging camera 14.
- the imaging execution unit 212 causes the imaging camera 14 to perform imaging with or without causing the light emitting elements 15 and 16 to emit light depending on whether the distance is measured or the proximity of the palm is determined.
- the image information receiving unit 213 receives the captured image information of the imaging camera 14 and causes the image information holding unit 211 to hold it.
- the distance measuring unit 214 uses the captured image that reflects the spot light projected from the four light emitting elements 18 held in the image information holding unit 211, as described above, from the imaging device 10 to the palm. Measure the distance.
- the distance measuring unit 214 determines whether the palm is tilted from four distances from the imaging device 10 to the palm. For example, when the four distances from the imaging device 10 to the palm of the captured image in which each of the four spot lights is reflected are not aligned by the method described in the first embodiment, it is determined that the palm is tilted. Is done.
- the image processing unit 215 takes the difference between the captured images of the palms captured when the light emitting element 15 held by the image information holding unit 211 emits light and when it does not emit light. In addition, the image processing unit 215 binarizes the captured image obtained from the difference.
- the palm held over the imaging device 10 does not necessarily cover the entire filter 11, but the palm may cover a part of the filter 11 or the filter 11 may be covered with the palm tilted.
- the portion covering the filter 11 is bright (white), and the portion not covering the filter 11 Will become brighter, and the boundary between the palm area and the other areas will not be clear. Therefore, by taking the difference between the captured images obtained by capturing the palm with the imaging camera 14 when the light emitting element 16a emits light and without emitting light, the area other than the palm becomes dark (black), and the palm area and the non-palm area It is possible to project separately.
- the image processing unit 215 obtains the brightness (or the brightness of the captured image obtained by taking the difference in this way for each region of the captured image divided into four in response to the spot light being projected from four locations.
- a pixel whose luminance is equal to or higher than a predetermined gradation is binarized as a white pixel and a pixel whose luminance is lower than the predetermined gradation is a black pixel.
- the image processing unit 215 causes the image information holding unit 211 to hold information obtained by performing such processing.
- the image discriminating unit 216 performs binarization by the image processing unit 215 so that the number of white pixels (bright region) in the captured image region where the distance cannot be measured is smaller than the area of the selected captured image region. To determine whether it occupies a predetermined ratio. As a result of the determination, if the bright area occupies a predetermined ratio, it is confirmed that the palm corresponding to the area of the captured image for which the distance could not be measured is located within the range of 0 to 5 mm from the imaging apparatus 10. Determine.
- the output request unit 217 outputs the measurement result and the discrimination result in the distance measurement unit 214 and the image discrimination unit 216 to the external device. Next, palm proximity determination processing performed by the imaging apparatus 10 having such a function will be described.
- FIG. 9 is a flowchart illustrating a processing procedure of proximity determination of the imaging apparatus according to the second embodiment.
- the imaging execution unit 212 projects light from the four light emitting elements 18 for distance measurement onto the palm and causes the imaging camera 14 to capture an image of the palm on which the spot light is projected.
- the spot light received from the imaging camera 14 by the image information receiving unit 213 receives the information of the captured image and causes the image information holding unit 211 to hold it.
- the distance measurement unit 214 measures the distance from the imaging device 10 to the palm using spot light reflected in the captured image.
- Step S12 The distance measurement unit 214 determines, for each region, whether or not each distance from the imaging device 10 to the four regions of the captured image is not measurable in the measurement result of the distance to the palm. . The next process proceeds to step S13 if there is no measurement possible, and if there is no measurement impossible, the measurement of the distance to the palm is completed and the proximity determination process ends.
- Step S13 Upon receiving an imaging execution request from the distance measurement unit 214, the imaging execution unit 212 causes the imaging camera 14 to perform imaging of the palm without causing the light emitting element 16a to emit light.
- the image information receiving unit 213 receives information of a captured image without light emission received from the imaging camera 14 and causes the image information holding unit 211 to hold it. [Step S14] When the imaging execution request is notified from the distance measurement unit 214, the imaging execution unit 212 causes the imaging camera 14 to perform imaging of the palm while the light emitting element 16a is caused to emit light.
- the image information receiving unit 213 receives the information of the captured image with light emission received from the imaging camera 14 and causes the image information holding unit 211 to hold it. Note that the processing of steps S13 and S14 may be performed in the order of steps S14 and S13.
- the image processing unit 215 calculates the difference between the captured images of the palm according to the presence or absence of light emission of the light emitting element 15 held by the image information holding unit 211. Further, the image processing unit 215 equally divides the captured image obtained by the difference into four.
- the image discriminating unit 216 takes the difference notified from the image processing unit 215 and locates the palm within the range of 0 to 5 mm with respect to the imaging device 10 for each area of the captured image divided into four. It is determined whether or not.
- step S16 If the determination result in step S16 is that the palm is not located within the range of 0 to 5 mm, the palm is 100 mm or more away from the imaging device 10 or the palm is held over the imaging device 10. Will not be.
- FIG. 10 and FIG. 11 are flowcharts illustrating the pixel discrimination processing procedure of the captured image of the imaging apparatus according to the second embodiment.
- Step S16b The image processing unit 215 first binarizes the region of the captured image divided into four by taking the difference held by the image information holding unit 211 selected in step S16a.
- the binarization for example, 69 or more of 256 gradations are set as white pixels, and 68 or less are set as black pixels.
- Step S16c The image processing unit 215 counts the number of white pixels (bright region) with respect to an area of a half of the binarized captured image region selected in step S16b.
- Step S16d The image determination unit 216 determines whether or not the bright region measured by the image processing unit 215 in step S16c is 10% or more of the entire region of the captured image selected by dividing into four.
- step S16e when the bright area is 10% or more of the entire imaging target area selected, and proceeds to step S16j when it is less than 10%.
- Step S ⁇ b> 16 e The image processing unit 215 counts the number of white pixels (bright region) with respect to an area of three quarters of the region of the selected captured image.
- Step S16f The image discriminating unit 216 discriminates whether or not the bright area measured by the image processing unit 215 in step S16e is 60% or more of the entire area of the picked-up image selected.
- step S16i the bright area is 60% or more of the entire selected area to be imaged, and proceeds to step S16g if it is less than 60%.
- Step S16g The image processing unit 215 counts the number of white pixels (bright region) with respect to the area of the entire region of the selected captured image.
- Step S16h The image determination unit 216 determines whether or not the bright region measured by the image processing unit 215 in step S16g is 60% or more of the entire region of the captured image selected by dividing into four.
- step S16i the bright area is 60% or more of the entire area of the selected captured image, and proceeds to step S16j if it is less than 60%.
- Step S16k The image processing unit 215 selects an area of the next number (i + 1) that has not been selected from the captured images divided into four in step S15, and proceeds to step S16c again.
- the lightness (or luminance) threshold value for binarizing the captured image is set to 69 of 256 gradations, but the user can also set a desired threshold value. . Further, it is desirable that the threshold value of the bright region for reliably determining that the palm is located within the range of 0 to 5 mm from the imaging device 10 is 50% or more of the captured image. In the flowchart, it is 60% or more. Further, in this case, in steps S16c and S16d, if the light area of the half of the selected area is less than 10%, the threshold is 60% even if the remaining half is the light area. Do not meet. Therefore, by steps S16c and S16d, a region of the captured image including a bright region that does not satisfy the threshold can be removed from the discrimination target at an early stage, and the processing burden of FIGS. 10 and 11 can be reduced.
- FIG. 12 is a diagram illustrating a hardware configuration example of the information processing apparatus to which the imaging apparatus according to the second embodiment is connected.
- the information processing apparatus 30 is, for example, a PC, and includes a CPU 30a, a RAM 30b, an HDD (Hard Disk Drive) 30c, a graphic processing unit 30d, and an input / output interface 30e.
- a PC main processing unit
- HDD Hard Disk Drive
- the CPU 30a performs overall control of the information processing apparatus 30 by executing various programs stored in a storage medium such as the HDD 30c.
- the RAM 30b temporarily stores at least a part of the OS and programs executed by the CPU 30a.
- the RAM 30b stores various data necessary for processing by the CPU 30a.
- the HDD 30c stores the OS and application programs on the information processing apparatus 30.
- the HDD 30c stores various data necessary for processing by the CPU 30a.
- the graphic processing unit 30d is a video card or a graphic board, and transmits image data to a monitor 30f that displays an image in accordance with a command from the CPU 30a.
- the input / output interface 30e is connected to a keyboard 30g and a mouse 30h.
- the input / output interface 30e outputs signals sent from the keyboard 30g and the mouse 30h to the CPU 30a.
- the imaging device 10 described above is connected to the input interface 30e.
- the imaging device 10 is electrically connected to the input interface 30e of the information processing device 30 from the external connection port 12a by, for example, a USB (Universal Serial Bus) cable.
- USB Universal Serial Bus
- the vein pattern of the palm of the user associated with the login password is stored in advance.
- the information processing device 30 is notified of the vein pattern information of the palm of the user read by the imaging device 10
- the login is completed when the vein pattern information associated with the login password is authenticated and matched.
- the imaging apparatus 10 first processes the output request for the display that the output request unit 217 displays “Please bring your palm close to the filter 11 of the imaging apparatus 10” on the monitor 30f. Notify device 30.
- the user brings the palm close to the filter 11 of the imaging device 10 in accordance with such display.
- the distance measurement unit 214 measures the palm distance using the distance measuring light emitting element 18 that emits light. Then, since at least one of the four locations of the palm cannot be measured, the imaging execution unit 212 causes the imaging camera 14 to execute imaging of the palm when the light emitting element 15 emits light and when it does not emit light (steps S11 to S14). ).
- the image processing unit 215 takes the difference between the captured images when the light emitting element 16a emits light and when it does not emit light, and divides it into four (step S15).
- the image discriminating unit 216 selects one area from among the areas of the captured image obtained by dividing the difference into four and binarizes it. The number of white pixels (bright area) in the binarized captured image area is counted, and the bright area is 60% or more of the entire captured image area selected by dividing into four areas. It is determined that the palm corresponding to the image area is located within the range of 0 to 5 mm from the filter 11 of the imaging device 10 (steps S16a to S16f, S16i).
- the output request unit 217 notifies the information processing apparatus 30 of an output request for displaying “Please keep your palm away from the imaging apparatus 10” on the monitor 30f.
- the user moves the palm away from the imaging device 10 in accordance with the display output from the information processing device 30.
- the distance measuring unit 214 measures the distance of the palm using the light emitting element 18 for distance measurement that emits light.
- the output requesting unit 217 reads “Please stop palm movement” on the monitor 30f. Is notified to the information processing apparatus 30.
- the distance measuring unit 214 confirms that the palm of the user stops within an appropriate distance from the imaging device 10 and that the palm is horizontal with respect to the filter 11 of the imaging device 10 based on the distances of the four locations of the palm. Upon detection, the imaging device 10 reads the vein pattern of the user's palm and notifies the information processing device 30 of information on the read vein pattern. When the distance measuring unit 214 detects that the palm is not level with respect to the filter 11, the output requesting unit 217 displays on the monitor 30f that "please level the palm with respect to the filter 11". Is sent to the information processing apparatus 30.
- the information processing apparatus 30 verifies the vein pattern information notified from the imaging apparatus 10 and the vein pattern associated with the login password held in advance, and if the vein patterns match, the information processing apparatus 30 The login of the device 30 is completed.
- the image discrimination unit 216 has the palm in the range of 0 to 5 mm from the imaging device 10 through the processing of steps S16a to S16k.
- the output request unit 217 determines that it is not located within the position, the output request unit 217 processes an output request for display of “Please bring your hand close to the imaging device 10 or hold your palm over the imaging device 10” on the monitor 30f. Notify device 30. Thereafter, the distance of the palm is again measured using the distance measuring light emitting element 18 emitted from the distance measuring unit 214, and the above processing is executed.
- the imaging camera 14 captures an image when the light emitting element 16a emits light and does not emit light, and obtains a difference between the captured images. I did it. Thereby, in the captured image of the palm, if the bright area is equal to or larger than the predetermined area with respect to the captured image, it is determined that the palm is located within the inappropriate imaging distance with respect to the imaging device 10. If it is less than the predetermined area, it is determined that the palm is not set in the imaging device 10 or is separated from a proper distance for imaging.
- the above processing functions can be realized by a computer.
- a program describing the processing contents of the functions that the imaging apparatuses 1 and 10 should have is provided.
- the program describing the processing contents can be recorded on a computer-readable recording medium.
- the computer-readable recording medium include a magnetic storage device, an optical disk, a magneto-optical recording medium, and a semiconductor memory.
- Magnetic storage devices include hard disk devices (HDD), flexible disks (FD), magnetic tapes, and the like.
- Optical discs include DVD (Digital Versatile Disc), DVD-RAM, CD (Compact Disc) -ROM / RW (Re-Writable), and the like.
- Magneto-optical recording media include MO (Magneto-Optical disk).
- the computer that executes the program stores, for example, the program recorded on the portable recording medium or the program transferred from the server computer in its own storage device. Then, the computer reads the program from its own storage device and executes processing according to the program. The computer can also read the program directly from the portable recording medium and execute processing according to the program. In addition, each time a program is transferred from a server computer connected via a network, the computer can sequentially execute processing according to the received program.
- processing functions described above can be realized by an electronic circuit such as a DSP (Digital Signal Processor), an ASIC (Application Specific Integrated Circuit), or a PLD (Programmable Logic Device).
- DSP Digital Signal Processor
- ASIC Application Specific Integrated Circuit
- PLD Programmable Logic Device
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Computer Networks & Wireless Communication (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Human Computer Interaction (AREA)
- Vascular Medicine (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Image Input (AREA)
- Optical Radar Systems And Details Thereof (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
本発明の上記および他の目的、特徴および利点は本発明の例として好ましい実施の形態を表す添付の図面と関連した以下の説明により明らかになるであろう。
[第1の実施の形態]
図1は、第1の実施の形態に係る撮像装置を説明するための概念図である。
撮像装置1は、撮像対象5が当該撮像装置1の撮像面1aの近傍に位置していることを判別することができるものである。
撮像部2は、撮像対象5を撮像する。撮像部2は、例えば、後述する図3に示されるように、撮像カメラ2a及び光学系2bを備えており、撮像カメラ2aで撮像対象5の画像を撮像する。
測距部6は、発光部3aの発光により撮像対象5との距離を測定する。
まず、撮像対象5の撮像面1aに対する高さに応じた発光部3aから投光されたスポット光の撮像画像について説明する。
なお、図2(A)~(C)は、図1(A)の距離(a)~(c)に位置する撮像対象5を撮像した撮像画像gをそれぞれ表している。
図3は、撮像装置から撮像対象までの距離の検出を説明するための図である。
この時、図3(A)に示されるように、撮像部2から撮像対象5までが距離Lの時、撮像部2の光軸と撮像対象5とが直交する位置Oからの撮像範囲をw、撮像部2から発光部3b2までの距離をd、位置Oから撮像対象5のスポット光3a2sの位置をd+xとする。なお、xは、発光部3a2からのスポット光の光軸のずれθ1により生じた変異量を表す。
(a/W)={(d+w)/w} ・・・(1)
また、撮像範囲wは撮像部2の画角の配置α1と距離Lとで表すことができ、光軸ずれθ1によるずれxも距離Lと光軸ずれθ1とで表すことができることから、式(1)を以下の式(2)のように表すことができる。
式(2)から、距離Lは次のように表すことができる。
L=d/{(a/W)*tanα1-tanθ1} ・・・(3)
tanθ1=(a/W)*tanα1-d/L ・・・(4)
ここで、光軸のずれθ1を算出する場合には、撮像装置1が予め測定した距離Lにおいてスポット光の距離aを計測して、式(4)に適用することでtanθ1を算出(校正)することができる。なお、この処理は、発光部3a1~3a4のそれぞれについて実行する。
当該式(5)から距離Lは次式で表される。
但し、k=Δ*d/{(a1’-a2’)*tanα1/W} ・・・(7)
したがって、スポット光の位置を利用して式(6)及び(7)により距離Lを算出することが可能となる。
図4は、撮像装置に対する撮像対象の傾き状態の検出を説明するための図である。
ax+by+cz+d=0 ・・・(8)
式(8)に対して、スポット光3a1s~3a4sの位置の座標を代入することにより、係数a,b,c,dを算出することができ、これらの係数から法線ベクトルが定まり、撮像対象5の傾き及び姿勢の情報を検出することができる。
また、
tanθ2=(L4-L3)/n ・・・(10)
と表すこともできる。
撮像対象5(例えば、手のひら)を、図1に示されるように、撮像装置1の撮像面1aを覆うように、撮像面1aの直上に配置してしまい、撮像装置1が発光部3aで当該撮像装置1から撮像対象5までの距離を検出できない場合に以下の処理が実行される。なお、撮像装置1から撮像対象5までの距離を検出できない理由については第2の実施の形態で説明する。
第2の実施の形態では第1の実施の形態の撮像装置についてより具体的に説明する。
図5は、第2の実施の形態に係る撮像装置の斜視図を示す図である。
図6は、第2の実施の形態に係る撮像装置の内部構成を説明するための図である。
なお、図6(A)は撮像装置10の側面図、図6(B)は図6(A)の一点鎖線X-Xにおける撮像装置10の上面図である。
図7は、第2の実施の形態に係る撮像装置のハードウェア構成例を示す図である。
ROM21bには、撮像装置10上のOS(Operating System)及びアプリケーションのプログラムが格納される。
図8は、第2の実施の形態に係る撮像装置が備える機能を示すブロック図である。
撮像装置10は、撮像装置10から手のひらまでの距離を計測して、計測不可となった場合に、当該手のひらが撮像装置10から0~5mmの範囲内に位置しているか否かを判別することができるものである。
撮像実行部212は、距離を計測する場合、また、手のひらの近接判別を行う場合に応じて、発光素子15,16を発光させて、または発光させずに、撮像カメラ14による撮像を実行させる。
距離計測部214は、画像情報保持部211に保持されている4つ発光素子18から投光されたスポット光が映った撮像画像を用いて、既述のようにして、撮像装置10から手のひらまでの距離を計測する。また、距離計測部214は、撮像装置10から手のひらまでの4つの距離から手のひらが傾いているか否かを判定する。例えば、第1の実施の形態で説明した方法により、4つのスポット光がそれぞれ映った撮像画像から、撮像装置10から手のひらまでの4つの距離が揃わない場合には、手のひらは傾いていると判定される。
次にこのような機能を備える撮像装置10で行われる手のひらの近接判別の処理について説明する。
[ステップS11] 撮像実行部212が距離計測用の4つの発光素子18から手のひらに光を投光させてスポット光が投光された手のひらの画像を撮像カメラ14に撮像させる。
距離計測部214は撮像画像に映っているスポット光を利用して撮像装置10から手のひらまでの距離を計測する。
[ステップS14] 撮像実行部212は、距離計測部214から撮像実行要求が通知されると、発光素子16aを発光させた状態で撮像カメラ14により手のひらの撮像を実行させる。
なお、ステップS13,S14の処理は、ステップS14,S13の順に行っても構わない。
図10及び図11は、第2の実施の形態に係る撮像装置の撮像画像の画素判別処理手順を示すフローチャートである。
[ステップS16d] 画像判別部216は、ステップS16cで画像処理部215が計測した明領域が4分割して選択した撮像画像の領域全体の10%以上であるか否かを判別する。
[ステップS16e] 画像処理部215は、選択した撮像画像の領域の4分の3の面積に対する白画素の画素数(明領域)を計数する。
[ステップS16h] 画像判別部216は、ステップS16gで画像処理部215が計測した明領域が4分割して選択した撮像画像の領域全体の60%以上であるか否かを判別する。
[ステップS16j] 画像判別部216は、選択した撮像画像の領域がi=4であるか否かの判別を行う。
[ステップS16k] 画像処理部215は、ステップS15で4分割した撮像画像から選択していない次の番号(i+1)の領域を選択して、再びステップS16cに進められる。
[実施例]
本実施例では、撮像装置10を、PC等の情報処理装置に対して、撮像装置10を利用して、ユーザ認証を行って、ログイン等を行う場合について説明する。
情報処理装置30は、例えば、PCであって、CPU30a、RAM30b、HDD(Hard Disk Drive)30c、グラフィック処理部30d及び入出力インタフェース30eを備える。
RAM30bには、CPU30aに実行させるOS及びプログラムの少なくとも一部が一時的に格納される。また、RAM30bには、CPU30aによる処理に必要な各種データが格納される。
撮像装置10は、外部接続口12aから、例えば、USB(Universal Serial Bus)ケーブルにより、情報処理装置30の入力インタフェース30eと電気的に接続されている。
情報処理装置30に電源が投入されると、撮像装置10は、まず、出力要求部217がモニタ30fに「撮像装置10のフィルタ11に手のひらを近づけてください」との表示の出力要求を情報処理装置30に通知する。
画像判別部216は、差分を取って4分割した撮像画像の領域中から1つの領域を選択して、二値化する。二値化した撮像画像の領域の白画素の画素数(明領域)を計数し、当該明領域が、4分割して選択した撮像画像の領域全体の60%以上であるために、選択した撮像画像の領域に対応する手のひらの部分が撮像装置10のフィルタ11から0~5mmの範囲内に位置していることを判別する(ステップS16a~S16f,S16i)。
情報処理装置30が出力する上記の表示に応じて、ユーザが手のひらを撮像装置10から遠ざける。この間も、距離計測部214は発光する距離測定用の発光素子18を利用して手のひらの距離を計測する。距離計測部214がユーザの手のひらが撮像装置10から適切な距離の範囲(35~75mm)に入ったことを検知すると、出力要求部217は、モニタ30fに「手のひらの動きを止めてください」との表示の出力要求を情報処理装置30に通知する。
1a 撮像面
2 撮像部
2a,14 撮像カメラ
2b 光学系
3a,3a1~3a4,3b 発光部
3a1s~3a4s スポット光
4 判別部
5 撮像対象
6 測距部
11 フィルタ
12 外装ケース
12a 外部接続口
13 カメラ回路基板
15 光学ユニット
16a,18 発光素子
16b 受光素子
17 導光体
19 ビーム照射機構
21 回路部
21a,30a CPU
21b ROM
21c,30b RAM
21d,30e 入出力インタフェース
22d 外部コネクタ部
30 情報処理装置
30c HDD
30d グラフィック処理部
30f モニタ
30g キーボード
30h マウス
211 画像情報保持部
212 撮像実行部
213 画像情報受信部
214 距離計測部
215 画像処理部
216 画像判別部
217 出力要求部
Claims (6)
- 撮像対象を撮像する撮像部と、
前記撮像対象の距離を計測する際に発光して前記撮像対象の前記撮像部が撮像する側を照らす第1の発光部と、
前記撮像対象を撮像する際に発光して前記撮像対象の前記撮像部が撮像する側を照らす第2の発光部と、
前記第1の発光部の発光により前記撮像対象との距離を測定する測距部と、
前記測距部が前記撮像対象との距離を測定できない場合に、前記第2の発光部が発光した状態で前記撮像部が撮像した前記撮像対象の撮像画像の明領域に基づき、前記撮像対象が前記撮像部に近接していることを判別する判別部と、
を有することを特徴とする撮像装置。 - 前記判別部は、前記発光部が発光していない状態で前記撮像部が撮像した前記撮像対象の撮像画像と、前記発光した状態で撮像された前記撮像画像との差分を取った差分画像の明領域に基づき、判別する、
ことを特徴とする請求の範囲第1項記載の撮像装置。 - 前記差分画像を二値化する画像処理部を更に有し、
前記判別部は、前記画像処理部が二値化した前記差分画像の白画素の前記差分画像の面積に対して占める明領域の割合に基づき、判別する、
ことを特徴とする請求の範囲第2項記載の撮像装置。 - 前記判別部は、前記二値化された前記差分画像を均等に4分割したうちのいずれかの差分画像の白画素の当該差分画像の面積に対して占める明領域の割合に基づき、判別する、
ことを特徴とする請求の範囲第3項記載の撮像装置。 - 撮影対象との距離を計測する際に第1の発光部が発光して前記撮像対象の撮像部が撮像する側を照らし、前記撮像対象との距離を測定し、
前記撮像対象との距離が測定できない場合に、前記撮像対象を撮影する際に第2の発光部が発光して前記撮像対象の撮像する側を照らした状態で前記撮像した前記撮像対象の撮像画像の明領域に基づき、前記撮像対象が近接していることを判別する、
ことを特徴とする撮像方法。 - コンピュータに、
撮影対象との距離を計測する際に第1の発光部に発光させて前記撮像対象の撮像部が撮像する側を照らし、前記撮像対象との距離を測定させ、
前記撮像対象との距離が測定できない場合に、前記撮像対象を撮影する際に第2の発光部が発光させて前記撮像対象の撮像する側を照らした状態で前記撮像した前記撮像対象の撮像画像の明領域に基づき、前記撮像対象が近接していることを判別させる、
処理を実行させることを特徴とする撮像プログラム。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR112013023154A BR112013023154A2 (pt) | 2011-03-15 | 2011-03-15 | aparelho de imageamento, método de imageamento e programa de imageamento |
EP11861178.9A EP2688278A4 (en) | 2011-03-15 | 2011-03-15 | IMAGING DEVICE, IMAGING METHOD, AND IMAGING PROGRAM |
JP2013504443A JP5866337B2 (ja) | 2011-03-15 | 2011-03-15 | 撮像装置、撮像方法及び撮像プログラム |
CN201180069191.1A CN103444165B (zh) | 2011-03-15 | 2011-03-15 | 摄像装置和摄像方法 |
PCT/JP2011/056031 WO2012124055A1 (ja) | 2011-03-15 | 2011-03-15 | 撮像装置、撮像方法及び撮像プログラム |
US13/974,897 US9807348B2 (en) | 2011-03-15 | 2013-08-23 | Imaging apparatus, imaging method, and imaging program |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2011/056031 WO2012124055A1 (ja) | 2011-03-15 | 2011-03-15 | 撮像装置、撮像方法及び撮像プログラム |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/974,897 Continuation US9807348B2 (en) | 2011-03-15 | 2013-08-23 | Imaging apparatus, imaging method, and imaging program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012124055A1 true WO2012124055A1 (ja) | 2012-09-20 |
Family
ID=46830186
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/056031 WO2012124055A1 (ja) | 2011-03-15 | 2011-03-15 | 撮像装置、撮像方法及び撮像プログラム |
Country Status (6)
Country | Link |
---|---|
US (1) | US9807348B2 (ja) |
EP (1) | EP2688278A4 (ja) |
JP (1) | JP5866337B2 (ja) |
CN (1) | CN103444165B (ja) |
BR (1) | BR112013023154A2 (ja) |
WO (1) | WO2012124055A1 (ja) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9420149B2 (en) * | 2014-06-13 | 2016-08-16 | Lips Corporation | Integrated depth camera |
JP6586729B2 (ja) * | 2015-01-16 | 2019-10-09 | 富士通株式会社 | 照明装置及び生体認証装置 |
JP2016225814A (ja) * | 2015-05-29 | 2016-12-28 | セイコーエプソン株式会社 | 撮像装置、撮像装置の制御方法 |
JP6939501B2 (ja) * | 2017-12-15 | 2021-09-22 | オムロン株式会社 | 画像処理システム、画像処理プログラム、および画像処理方法 |
JP6835020B2 (ja) * | 2018-03-15 | 2021-02-24 | オムロン株式会社 | 画像処理システム、画像処理装置、画像処理プログラム |
JP7106668B2 (ja) * | 2018-09-27 | 2022-07-26 | 芝浦メカトロニクス株式会社 | 錠剤印刷装置及び錠剤印刷方法 |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS60135712U (ja) * | 1984-02-20 | 1985-09-09 | 三洋電機株式会社 | 自動焦点カメラ |
JPH1194542A (ja) * | 1997-09-16 | 1999-04-09 | Hamamatsu Photonics Kk | 測距モジュール |
JP2002150440A (ja) * | 2000-11-13 | 2002-05-24 | Fujitsu General Ltd | 監視対象物検出装置 |
JP2002232779A (ja) * | 2001-01-29 | 2002-08-16 | Yoshiaki Masuno | 証明写真撮影カメラおよびその使用方法 |
JP2006243189A (ja) * | 2005-03-01 | 2006-09-14 | Fujinon Corp | カメラ |
JP2007025796A (ja) * | 2005-07-12 | 2007-02-01 | Sony Corp | 表示装置及び表示方法 |
JP2008128815A (ja) * | 2006-11-21 | 2008-06-05 | Stanley Electric Co Ltd | 遠近判定方法およびその装置 |
JP2010107722A (ja) * | 2008-10-30 | 2010-05-13 | Hoya Corp | 焦点検出方法及び焦点検出装置 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3894038B2 (ja) | 2002-05-08 | 2007-03-14 | 日本電気株式会社 | 画像処理装置、画像処理装置のはみ出し検知方法およびプログラム |
JP4644540B2 (ja) | 2005-06-28 | 2011-03-02 | 富士通株式会社 | 撮像装置 |
JP4799216B2 (ja) | 2006-03-03 | 2011-10-26 | 富士通株式会社 | 距離測定機能を有する撮像装置 |
JP4745084B2 (ja) | 2006-03-03 | 2011-08-10 | 富士通株式会社 | 撮像装置 |
JP5312774B2 (ja) | 2006-12-15 | 2013-10-09 | 日立オートモティブシステムズ株式会社 | 薄型認証センサ |
JP5047773B2 (ja) | 2007-12-21 | 2012-10-10 | 株式会社アルファ | 指静脈撮影装置及び個人認証装置 |
JP4853507B2 (ja) * | 2008-10-30 | 2012-01-11 | ソニー株式会社 | 情報処理装置、情報処理方法およびプログラム |
JP4771183B2 (ja) * | 2009-01-30 | 2011-09-14 | 株式会社デンソー | 操作装置 |
JP5237394B2 (ja) | 2009-01-30 | 2013-07-17 | 富士通フロンテック株式会社 | 認証装置、撮像装置、認証方法および認証プログラム |
US20100245289A1 (en) * | 2009-03-31 | 2010-09-30 | Miroslav Svajda | Apparatus and method for optical proximity sensing and touch input control |
JP5297997B2 (ja) * | 2009-12-21 | 2013-09-25 | 株式会社日立メディアエレクトロニクス | 個人認証装置及び移動通信端末 |
TW201241682A (en) * | 2011-04-01 | 2012-10-16 | Besdon Technology Corp | Multi-functional position sensing device |
-
2011
- 2011-03-15 BR BR112013023154A patent/BR112013023154A2/pt not_active IP Right Cessation
- 2011-03-15 CN CN201180069191.1A patent/CN103444165B/zh active Active
- 2011-03-15 WO PCT/JP2011/056031 patent/WO2012124055A1/ja unknown
- 2011-03-15 EP EP11861178.9A patent/EP2688278A4/en not_active Withdrawn
- 2011-03-15 JP JP2013504443A patent/JP5866337B2/ja active Active
-
2013
- 2013-08-23 US US13/974,897 patent/US9807348B2/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS60135712U (ja) * | 1984-02-20 | 1985-09-09 | 三洋電機株式会社 | 自動焦点カメラ |
JPH1194542A (ja) * | 1997-09-16 | 1999-04-09 | Hamamatsu Photonics Kk | 測距モジュール |
JP2002150440A (ja) * | 2000-11-13 | 2002-05-24 | Fujitsu General Ltd | 監視対象物検出装置 |
JP2002232779A (ja) * | 2001-01-29 | 2002-08-16 | Yoshiaki Masuno | 証明写真撮影カメラおよびその使用方法 |
JP2006243189A (ja) * | 2005-03-01 | 2006-09-14 | Fujinon Corp | カメラ |
JP2007025796A (ja) * | 2005-07-12 | 2007-02-01 | Sony Corp | 表示装置及び表示方法 |
JP2008128815A (ja) * | 2006-11-21 | 2008-06-05 | Stanley Electric Co Ltd | 遠近判定方法およびその装置 |
JP2010107722A (ja) * | 2008-10-30 | 2010-05-13 | Hoya Corp | 焦点検出方法及び焦点検出装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2688278A4 * |
Also Published As
Publication number | Publication date |
---|---|
JP5866337B2 (ja) | 2016-02-17 |
US20130335560A1 (en) | 2013-12-19 |
US9807348B2 (en) | 2017-10-31 |
JPWO2012124055A1 (ja) | 2014-07-17 |
EP2688278A1 (en) | 2014-01-22 |
CN103444165B (zh) | 2017-09-08 |
CN103444165A (zh) | 2013-12-11 |
EP2688278A4 (en) | 2014-08-13 |
BR112013023154A2 (pt) | 2016-12-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5866337B2 (ja) | 撮像装置、撮像方法及び撮像プログラム | |
US10310675B2 (en) | User interface apparatus and control method | |
JP4927021B2 (ja) | 画像表示装置のカーソル制御装置及び制御方法、ならびに画像システム | |
US8957864B2 (en) | Coordinate input apparatus and method | |
JP6369078B2 (ja) | 画像補正装置、画像補正方法及び画像補正用コンピュータプログラム | |
WO2013145164A1 (ja) | 撮像装置 | |
US20160300356A1 (en) | Measurement device that measures shape of object to be measured, measurement method, system, and article production method | |
US8121814B2 (en) | Three-dimensional processor and method for controlling display of three-dimensional data in the three-dimensional processor | |
JP6869002B2 (ja) | 計測装置 | |
TWI575406B (zh) | 用以產生顯示區之角落的方法、相關電腦系統及機器可讀儲存媒體 | |
JP6217373B2 (ja) | 動作判定方法、動作判定装置および動作判定プログラム | |
US20170228092A1 (en) | Touch region projection onto touch-sensitive surface | |
JP6314688B2 (ja) | 入力装置 | |
JP4908523B2 (ja) | 位置情報検出装置、位置情報検出方法及び位置情報検出プログラム | |
CN113379837A (zh) | 检测设备的角度校正方法、设备及计算机可读存储介质 | |
JP6904624B1 (ja) | クラブヘッド計測用マークおよび画像処理装置 | |
CN107077196B (zh) | 识别触敏表面上的对象 | |
TWI408611B (zh) | 利用物件特徵相異性進行多個物件之動態影像辨識方法及使用該方法之系統 | |
JP6031751B2 (ja) | ガラス基板検査装置及びガラス基板製造方法 | |
WO2015145977A1 (ja) | 情報処理装置、画像処理方法と記録媒体、ならびに、pos端末装置 | |
JP4496149B2 (ja) | 寸法測定装置 | |
US20170228100A1 (en) | Misalignment detection | |
JP4229964B2 (ja) | エリアイメージセンサを用いた位置検出装置 | |
JP3898706B2 (ja) | エリアイメージセンサを用いた位置検出装置 | |
JPH10288514A (ja) | 対象物の回転角度検出方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11861178 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2013504443 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112013023154 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112013023154 Country of ref document: BR Kind code of ref document: A2 Effective date: 20130910 |