US20180121724A1 - Biometric camera - Google Patents
Biometric camera Download PDFInfo
- Publication number
- US20180121724A1 US20180121724A1 US14/500,674 US201414500674A US2018121724A1 US 20180121724 A1 US20180121724 A1 US 20180121724A1 US 201414500674 A US201414500674 A US 201414500674A US 2018121724 A1 US2018121724 A1 US 2018121724A1
- Authority
- US
- United States
- Prior art keywords
- shutter
- image
- user
- iris
- mobile device
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G06K9/00604—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- G06K9/00033—
-
- G06K9/00617—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/13—Sensors therefor
- G06V40/1312—Sensors therefor direct reading, e.g. contactless acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/19—Sensors therefor
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/197—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
- G06T2207/20201—Motion blur correction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/586—Depth or shape recovery from multiple images from multiple light sources, e.g. photometric stereo
-
- G—PHYSICS
- G07—CHECKING-DEVICES
- G07C—TIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
- G07C9/00—Individual registration on entry or exit
- G07C9/30—Individual registration on entry or exit not involving the use of a pass
- G07C9/32—Individual registration on entry or exit not involving the use of a pass in combination with an identity check
- G07C9/37—Individual registration on entry or exit not involving the use of a pass in combination with an identity check using biometric data, e.g. fingerprints, iris scans or voice recognition
Definitions
- biometric devices such as laptops—and most recently smartphones—have included biometric devices with the purpose of user identification.
- the biometric devices for example include fingerprint scanners, iris scanners and camera systems capable of face and/or voice recognition.
- fingerprint identification systems suffer from the fact that users routinely leave fingerprints on various objects—and those fingerprints can be “lifted” and reproduced to circumvent fingerprint identification systems.
- face and voice recognition systems may not be sufficiently accurate and can often be circumvented with relative ease.
- iris scanning systems may be of particular interest because they are relatively more difficult to circumvent, while being very accurate and easy to use.
- biometric camera systems exist to capture images of the iris.
- a mobile device such as a smartphone or tablet.
- Such biometric camera systems face a number of challenges during their operation, including 1) ambient illumination (e.g. be able to operate in full sunlight); 2) motion blur (e.g. user's hand shaking while holding the smartphone and capturing iris); 3) depth of field at close distances (e.g. while taking pictures at short (“macro”) distances, the object being further or closer than the focused distance makes the image blurry); 4) additional cost and additional space for a biometric sensor; and 5) limited field of view (e.g., the camera must capture a high-resolution image of the iris for reliable identification.
- ambient illumination e.g. be able to operate in full sunlight
- motion blur e.g. user's hand shaking while holding the smartphone and capturing iris
- depth of field at close distances e.g. while taking pictures at short (“macro”) distances, the object being further or closer than the focused distance makes the image
- the iris is small—e.g. 12 mm in diameter—while the distance between the camera and the iris is considerable—e.g. arm's length, ⁇ 25 cm—the camera must be considerably “zoomed in” on the face and iris area, rather than image the user and his/her surroundings as in a wide-angled shot).
- Exemplary embodiments for a biometric camera system for a mobile device comprise: a near infrared (NIR) light source on the mobile device that flashes a user of the mobile device with near infrared light during image capture; a biometric camera located on the mobile device offset from the NIR light source, the biometric camera comprising: an extended depth of field (EDOF) imaging lens; a bandpass filter located adjacent to the EDOF imaging lens to reject ambient light during image capture; and an imaging sensor located adjacent the bandpass filter that converts an optical image of an object into an electronic signal for image processing; and a processor configured to receive video images of an iris of a user from the image sensor, and attempt to match the video images of the iris with previously registered images stored in an iris database, wherein if a match is found, the user is authenticated.
- NIR near infrared
- EDOF extended depth of field
- FIG. 1 is a block diagram illustrating an exemplary embodiment of a mobile device biometric camera system
- FIG. 2 is a block diagram illustrating components of the biometric camera for a mobile device according to an exemplary embodiment
- FIG. 3 a diagram illustrating the captured image bounded by a frame in which a window of interest containing an iris has been identified
- FIGS. 4A, 4B and 4C are diagrams showing one embodiment in which a shutter is patterned into left and right halves.
- FIGS. 5A and 5B are diagrams showing an example of the shutter subdivided into more independently addressable areas so the biometric camera can be used as a plenoptic camera.
- a component or module means, but is not limited to, a software or hardware component, such as a field programmable gate array (FPGA) or an application specific integrated circuit (ASIC), which performs certain tasks.
- a component or module may advantageously be configured to reside in the addressable storage medium and configured to execute on one or more processors.
- a component or module may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.
- the functionality provided for the components and components or modules may be combined into fewer components and components or modules or further separated into additional components and components or modules.
- the exemplary embodiments provide a mobile device with a biometric camera system for reliable iris identification and user authentication.
- the camera is capable of working at high levels of ambient illumination, capturing images practically without motion blur, provides considerable latitude in distances at which a clear image can be captured and can be realized in a small form factor.
- the camera can optionally perform functions including proximity sensing, night-vision camera, 3D time-of-flight sensor, eye position and gaze tracking camera, and structured light for 3D sensing.
- the system can automatically identify whether the left and/or right eye is being imaged.
- FIG. 1 is a block diagram illustrating an exemplary embodiment of a mobile device biometric camera system.
- the system includes a mobile device 10 having conventional components including a memory 12 , at least one processor 14 , input output devices (I/O) 16 , an front-facing camera 18 (e.g., RGB camera), and a display 22 .
- I/O input output devices
- front-facing camera 18 e.g., RGB camera
- the memory 12 , the processor 14 , the I/O 16 , the front-facing camera 18 and the display 22 may be coupled together via one or more system buses (not shown).
- the memory 12 may comprise one or more memories comprising different memory types, including RAM, ROM, cache, virtual memory and flash memory, for example.
- the processor 14 may include a single processor having one or more cores, or multiple processors having one or more cores.
- the I/O 16 is a collection of components that input information and output information. Example components comprising the I/O 16 may include a microphone, speaker, and a wireless network interface controller (or similar component) for communication over the network.
- the processor 14 may execute an operating system (OS) that manages hardware resources and performs basic tasks. Examples of the OS may include SymbianTM, BlackBerry OSTM, iOSTM, WindowsTM, and AndroidTM.
- the display 22 may be integrated with the mobile device 10 , while in another embodiment, the display 22 may be external from the mobile device 10 .
- the mobile device 10 may comprise any type of mobile device form factor, including but not limited to, a cell or smart-phone, a tablet, a notebook or laptop computer, a television, and a wearable computer, for example.
- the mobile device 10 may be implemented with the display 22 and the front-facing camera 18 located on the same side of the mobile device 10 , such that the front-facing camera 18 is pointed at the user has the user holds the device to view the display 22 .
- the front-facing camera is typically housed within a lid of the laptop.
- the mobile device 10 is provided with a biometric camera system 24 that is used to capture images of the human iris for user identification and authentication.
- the biometric camera system 24 may include a near infrared (NIR) light source 20 , a biometric camera 28 , an iris recognition component 30 and an iris database 32 .
- the iris recognition component 30 and the iris database 32 may be software components stored in the memory 12 and executed by the processor 14 .
- the biometric camera 28 may be located in one corner of the mobile device 10 (although other locations are also possible), while the NIR light source 20 may be located in an opposite corner to offset the NIR light source 20 from the biometric camera 28 within the body of the mobile device 10 .
- the NIR light source 20 may be implemented using micro light emitting diodes (LEDs) or a laser diode, and the NIR light source 20 may be slightly angled upwards to point into the user's eye located at an expected distance when held normally by the user.
- the NIR light source 20 flashes a user of the mobile device 10 with near infrared light during image capture of the user's iris by the biometric camera 28 .
- Video images output from the biometric camera 28 are received by the iris recognition component 30 , which attempts to match the image of the iris with previously registered images stored in the iris database 32 . If a match is found, then the user may be authenticated.
- FIG. 2 is a block diagram illustrating components of the biometric camera 28 according to an exemplary embodiment.
- the biometric camera 28 may operate to gather reflected near infrared light and form an image of a human iris using a combination of an imaging lens 200 , a bandpass filter 202 , and an imaging sensor 204 .
- the imaging lens 200 may comprise an extended depth of field (EDOF) imaging lens.
- EEOF extended depth of field
- the bandpass filter 202 may be located adjacent to the imaging lens 200 and functions to reject ambient light.
- the bandpass filter 202 may be combined with a solid-state shutter.
- the band pass filter may comprise a near-infrared bandpass filter.
- the biometric camera 28 may also include an optional visible light filter 206 located adjacent to the bandpass filter 202 and optional solid-state shutter that may be used to further reject visible light.
- the imaging sensor 204 may be located adjacent to the bandpass filter 202 and converts an optical image of an object (e.g., an iris) into an electronic signal for subsequent image processing.
- the image sensor 204 may include a built-in rolling shutter or a freeze-frame shutter.
- the image sensor 204 may comprise a digital charge-coupled device (CCD) or complementary metal—oxide—semiconductor (CMOS) active pixel sensors.
- CCD digital charge-coupled device
- CMOS complementary metal—oxide—semiconductor
- the biometric camera system 28 is designed to have a very small size to fit into a mobile device 10 , such as a smartphone, and to also be low in cost.
- the image sensor 204 may comprise a CMOS sensor with a small pixel pitch, e.g., 1.1 ⁇ m, and a rolling shutter (not freeze-frame).
- CMOS complementary metal-oxide-semiconductor
- a rolling shutter not freeze-frame
- Such a sensor type is commercially available and used in smartphones for front and back-facing cameras.
- the preferred embodiment utilizes an off-the-shelf black-and-white CMOS sensor with rolling shutter and small pixels.
- the bandpass filter 202 may function to reject all wavelengths except near-infrared, e.g., 800 nm+ ⁇ 5 nm.
- the bandpass filter 202 may be combined a solid-state shutter.
- the process of capturing an image comprises:
- One goal may be to make the exposure time (when the solid-state shutter is open and the NIR light source is enabled) to be as short as practically possible.
- the NIR light source power may be made as high as practically possible without damaging the human eye.
- the solid-state shutter closed at all times except during capture the ambient light received from elsewhere in the scene may be significantly reduced. This achieves the goal of rejecting ambient illumination. Also, due to the very short exposure time, the goal of practically motion blur-free image capture may be achieved.
- the solid-state shutter may comprise gallium arsenide (GaAs).
- GaAs gallium arsenide
- This type of shutter can change its state from opaque to transparent in a very short time (several nanoseconds) by applying voltage externally.
- Using a GaAs-based shutter as a filter is particularly desirable in this application since this type of shutter has a high index of refraction.
- Typical dichroic (“interference”) optical bandpass filters often have a relatively wide bandpass width due to the dependence of the central wavelength on the angle of incidence. In camera systems, light rays can pass through the optical filter at various angles.
- the bandpass wavelength shifts as a function of incoming light ray angle, it is typical to make the bandpass wide, such that regardless of the incoming ray angle, light energy of the desirable wavelength in that ray can still pass through the optical filter.
- the amount of shift of the center wavelength is determined as
- ⁇ c ⁇ 0 ⁇ 1 - sin 2 ⁇ ⁇ n * 2
- ⁇ c is the central wavelength
- ⁇ 0 is the central wavelength at normal incidence
- n* is the filter effective index of refraction. Since GaAs has a high n*, the shift effect is reduced, the filter can be designed to have a narrower bandwidth—and thus reject even more ambient light.
- CMOS sensors with a built-in freeze-frame shutter and NIR bandpass shutter may be used instead of using a solid state shutter.
- CMOS sensors with a freeze-frame shutter typically have pixels of larger size compared to CMOS sensors utilizing rolling shutter. Larger pixel size reduces image resolution—or increases sensor size and cost. Larger pixel size also reduces the depth of field, as explained below. This is why the use of GaAs solid shutter is preferred.
- a solid-state shutter such one based on GaAs, may still leak some light in its closed state. If the image sensor 204 continues sensing an image while the image is read out, which is typical in image sensors utilizing rolling shutter, the leaking ambient light can register in the captured image and adversely affect the image quality, e.g., by saturating some image areas or adding noise to the image.
- the processor may be configured to reduce exposure from ambient light by reducing exposure and readout time as follows:
- the NIR light source may be synchronously pulsed with the readout of the rolling shutter.
- the rolling shutter may be addressed in a row-column manner where only the pixels in the window of interest 302 are read out.
- the NIR light source may turned on at or slightly before the time where the upper left of the window of interest 302 is read out and then turned off after the lower right pixel of the window of interest 302 has been read.
- the rejection of ambient light can be made greater by pulsing at a higher light intensity. The total and instantaneous amounts of NIR light must be below certain thresholds to avoid eye damage.
- the intensity of the light may be increased during this period than would be allowed for eye safety if it were used while exposing and reading the entire frame 300 .
- the increased intensity of light enables a shorter exposure and readout time, leading to less ambient light readout.
- a rolling shutter is used to read entire lines rather than a small window 302 .
- the shutter resets the pixels and reads out at the top of the window 302 and finishes reading at the bottom of the window.
- the NIR light may be turned on and off to synchronize with the beginning and end of this readout. As this slice of the frame 300 is smaller than the full frame, the intensity of the pulsed NIR may be higher than if exposing the full frame.
- the NIR light may be further pulsed to synchronize with each line readout.
- the full lines are being read out, the NIR is only turning on while exposing the window of interest 302 . The instantaneous intensity may be made even higher in this embodiment than in the previous one, even though the total NIR flux may be reduced.
- the biometric camera 28 should typically capture an iris image at distances of around 25 cm. At such short distances depth of field is typically limited and it is difficult to obtain a sharp image. To increase the depth of field, the exemplary embodiments may include:
- EDOF extended depth of field
- an auto-focus actuator Using an auto-focus actuator.
- the main purpose of the auto-focus actuator is to focus speedily on objects at “macro” range, e.g. around 10-50 cm, for iris scanning, the actuator may also focus at longer distances up to infinity for applications described below.
- EDOF systems were historically utilized for color capture and experienced some issues capturing and correcting images across all visible wavelengths, resulting in degradation of image quality. Since the present embodiments are capturing a monochromatic NIR image, the issues related to operation across a wide range of wavelengths are eliminated and system performance can improve. Additionally, unlike color sensors typically utilizing Bayer pattern where only one of three colors is sampled at each pixel location, the monochrome sensor samples the image at every pixel location, thus providing EDOF with full information for improved system performance.
- EDOF systems increase the depth of focus to improve the range where iris pictures may be taken, they also tend to decrease the MTF of the picture.
- the image quality may be improved by using known super-resolution techniques to capture several images of the iris and process them to achieve a sharper picture. To achieve this, several pictures may be taken in succession with the sub-window rolling shutter to quickly capture the window of interest 302 multiple times. If there is movement between the frames, image registration can be done before combining.
- the combining step may use a super-resolution technique, or could instead do a simple signal averaging to reduce noise. In a particular embodiment, this may be combined by synchronizing the NIR pulse with reading out of the window of interest 302 to achieve faster readouts of the window and ameliorating effects of movement within the frame.
- the architecture of the biometric camera system 24 enables the biometric camera system 24 to perform additional functions in addition to iris scanning, including proximity sensing, night-vision, 3D time-of-flight sensing, eye position and gaze tracking, and providing a structured light for 3D sensing, as further described below.
- Proximity sensors utilize an imaging sensor and a NIR light source.
- the image sensor typically comprises a single pixel (not array of multiple pixels) and the proximity operation may comprise:
- proximity sensors are typically designed to have a high dynamic range, such that they work outdoors in high ambient light.
- the shutter can open for a very short time (synchronized with the NIR light source pulsing) and thus greatly reduce the exposure from the ambient light. Therefore, the high dynamic functionality may not be necessary, thus potentially reducing imager cost and increasing image quality.
- every smartphone is equipped with a proximity sensor to turn off the touch screen while the user is holding the phone against his/her ear to prevent accidental touch input.
- the biometric camera system 24 may perform the function of the proximity sensor, the proximity sensor may be removed from the phone, and replaced with the biometric camera system 24 without sacrificing the functionality and without unnecessarily growing smartphone size and cost.
- Exemplary embodiments also may reuse an already-existing NIR light source for both proximity sensing and biometric scanning.
- the biometric camera 28 can capture images in darkness.
- the system 24 can also detect a user in total darkness, scan an iris, capture user gestures and so on. Utilizing an auto-focus actuator instead of a fixed-focus (preferably EDOF) lens may be beneficial to be able to clearly capture objects distances further than 25 cm.
- EDOF fixed-focus
- CMOS camera equipped with a fast GaAs shutter and NIR light source is capable of capturing distance images.
- a time-of-flight camera is a class of scannerless LIDAR, in which the entire scene is captured with each laser or light pulse. It should be noted that the distance measured to an object (e.g., a face) by the time-of-flight camera can be used for fast focusing using the auto-focus actuator.
- NIR light source 20 and NIR biometric camera 20 allows capturing iris and eyes continuously to track their location and gaze direction using existing techniques.
- the NIR biometric camera 28 can be used to detect motion, including motion in complete darkness, by utilizing imaging techniques known in the field (e.g., capturing frames and comparing their aggregated sharpness).
- a patterned slide can be inserted in front of the NIR light source 20 to implement 3D sensing.
- the pattern will be reflected off the object being imaged and then captured by the biometric camera 28 .
- the absolute distance of points on the object can be determined.
- structured light capture can only be used for imaging static scenes at high resolution because moving objects change position too rapidly to update the depth map without blur. Furthermore, in bright ambient light conditions, structured lighting depth capture systems suffer from wash out.
- the fast shutter and a strong NIR light source 20 can mitigate against both of these issues.
- the problem of ambient wash out may be reduced if another NIR light source is added with a patterned slide that has a complementary pattern to the pattern on the first slide. For example, if the new slide has a pattern inverse to the first pattern, then an object can be rapidly illuminated with the first NIR light source and then the second, synchronized with the fast shutter, and then create a higher resolution pattern by using standard computational imaging operations on the two captured patterns. This would enable, e.g., 3D video capture, or the ability to track small changes in facial expression of a person holding the camera in front-facing mode.
- the system can use the fast shutter and a structured light pattern over the IR illuminator to capture high precision, low blur, 3D depth maps of a person's face, paired with the RGB images of the face, captured from the standard front-facing camera.
- the field of view of the NIR camera is designed to be wide enough to capture the entirety of the user's face at the target distance range for the application (typically 10 cm-25 cm).
- These 3D depth maps and RGB image pairs can be used to create a 3D model of a person's face, which can then be used to create a biometric signature, unique to that user.
- a benefit of this embodiment is that the same system setup can be used to simultaneously capture the person's iris image and 3D facial image, allowing for multi-factor authentication. Capture of the iris and 3D face may be done with the same IR image, or there may be possibly two IR illuminators, one with the structured light pattern on top of it and the other with no pattern on top. This can potentially simplify the processing of the image which would be needed to remove the structured light pattern from the iris image. The same camera could be used in rapid alternating sequence to capture iris and 3D face images.
- the solid-state shutter of the biometric camera 28 is addressable with a single set of electrodes and covers the entire field of view of the shutter.
- electrodes may be patterned on the shutter so that there are two or more separately addressable areas of the shutter, that is, one portion of the shutter could be left open to the wavelengths of light measured while the other portion or portions could remain closed.
- FIGS. 4A, 4B and 4C are diagrams showing one embodiment in which a shutter 400 is patterned into left and right halves.
- FIG. 4B is a diagram showing the left half 402
- FIG. 4C is a diagram showing the right half 404 .
- alternate pictures can be taken with each half successively blocked while the other side remains open.
- a 3D stereo image may be captured through single lens technology with time-multiplex.
- a left stereo pair image may be captured with the left half 402 of the shutter 400 transparent and the right half 404 opaque.
- a right stereo pair image may be captured with the right half 404 of the shutter 400 transparent and the left half 402 opaque.
- the left and right captured images can then be processed using known techniques to provide stereo pair images using a single lens camera.
- strong ambient lighting might be used to capture the images.
- the NIR light source may also be used to supplement or replace ambient light.
- the NIR pulses may be synchronized to the timing of the shutter halves while increasing exposure lighting without exceeding eye safety limits.
- the shutter of the biometric camera 28 may be further subdivided into more independently addressable areas to create a patterned shutter with time-multiplex for the capture of light-field photos.
- FIGS. 5A and 5B are diagrams showing an example of the shutter 500 subdivided into more independently addressable areas 502 so the biometric camera can be used as a plenoptic camera. In this example, successive frames may be taken, each with on of the areas 502 of the shutter 500 opened during its frame capture. The series of captured frames can then be processed as a light field image set.
- FIG. 5B shows an example of only a single addressable area 502 open at one time for a frame capture, in another embodiment, multiple areas 502 of the shutter could be open at the same time in various patterns.
- the biometric camera 28 may have a relatively narrow field of view ( ⁇ 30 degrees horizontal), it may not always capture the image of the entire user face, but only an image of one eye. To decide which eye this is—left or right—the output of the front-facing camera 18 may be combined to determine which of the eyes is being imaged by the biometric camera 28 .
- the image from the front-facing camera 18 may provide information regarding the position of the face with respect to the field of view imaged by the biometric camera 28 .
- the time-of-flight operation can be used to assist the front-facing camera 18 for focusing;
- the front-facing camera 18 can also automatically capture user's face during the iris scan, e.g. for reference purposes;
- the front-facing camera 18 can capture the user's face to concurrently with the iris scan to additionally verify that the user is a living human (e.g. by tracking pulse using known methods based on spatio-temporal filtering and amplification of color);
- the front-facing camera 18 may confirm the presence and determine the position of the user's head and eyes during iris scan to assist recognition in case of problems with head positioning, such that the system can guide the user to position his/her head properly (and also e.g. by showing the user a target to look at on the display or e.g. flash LED on the phone); and
- the front-facing camera can detect the presence of glasses, hair and identify other problems preventing successful iris scan.
- the system e.g., smartphone
- the system can then direct the user to address these problems, e.g. remove glasses, adjust hair and track and verify that the problems are addressed using the front-facing camera.
- a method and system for a biometric camera system has been disclosed.
- the present invention has been described in accordance with the embodiments shown, and there could be variations to the embodiments, and any variations would be within the spirit and scope of the present invention.
- the exemplary embodiment can be implemented using hardware, software, a computer readable medium containing program instructions, or a combination thereof.
- Software written according to the present invention is to be either stored in some form of computer-readable medium such as a memory, a hard disk, or a CD/DVD-ROM and is to be executed by a processor. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Security & Cryptography (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Ophthalmology & Optometry (AREA)
- General Health & Medical Sciences (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Input (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Exemplary embodiments for a biometric camera system for a mobile device, comprise: a near infrared (NIR) light source on the mobile device that flashes a user of the mobile device with near infrared light during image capture; a biometric camera located on the mobile device offset from the NIR light source, the biometric camera comprising: an extended depth of field (EDOF) imaging lens; a bandpass filter located adjacent to the EDOF imaging lens to reject ambient light during image capture; and an imaging sensor located adjacent the bandpass filter that converts an optical image of an object into an electronic signal for image processing; and a processor configured to receive video images of an iris of a user from the image sensor, and attempt to match the video images of the iris with previously registered images stored in an iris database, wherein if a match is found, the user is authenticated.
Description
- This application claims the benefit of Provisional Application Ser. No. 61/884,778, filed Sep. 30, 2013; Provisional Application Ser. No. 61/885,291, filed Oct. 1, 2013; and Provisional Application Ser. No. 61/913,620, filed Dec. 9, 2013, each assigned to the assignee of the present application, and incorporated herein by reference.
- Mobile devices such as laptops—and most recently smartphones—have included biometric devices with the purpose of user identification. The biometric devices for example include fingerprint scanners, iris scanners and camera systems capable of face and/or voice recognition. However, fingerprint identification systems suffer from the fact that users routinely leave fingerprints on various objects—and those fingerprints can be “lifted” and reproduced to circumvent fingerprint identification systems. Also, face and voice recognition systems may not be sufficiently accurate and can often be circumvented with relative ease. In this situation, iris scanning systems may be of particular interest because they are relatively more difficult to circumvent, while being very accurate and easy to use.
- Various camera systems exist to capture images of the iris. In particular, there is interest in providing a biometric camera system that can be realized in a mobile device such as a smartphone or tablet. Such biometric camera systems face a number of challenges during their operation, including 1) ambient illumination (e.g. be able to operate in full sunlight); 2) motion blur (e.g. user's hand shaking while holding the smartphone and capturing iris); 3) depth of field at close distances (e.g. while taking pictures at short (“macro”) distances, the object being further or closer than the focused distance makes the image blurry); 4) additional cost and additional space for a biometric sensor; and 5) limited field of view (e.g., the camera must capture a high-resolution image of the iris for reliable identification. Since the iris is small—e.g. 12 mm in diameter—while the distance between the camera and the iris is considerable—e.g. arm's length, −25 cm— the camera must be considerably “zoomed in” on the face and iris area, rather than image the user and his/her surroundings as in a wide-angled shot).
- Therefore, a need exists to provide a biometric camera system that is capable of addressing the above issues.
- Exemplary embodiments for a biometric camera system for a mobile device, comprise: a near infrared (NIR) light source on the mobile device that flashes a user of the mobile device with near infrared light during image capture; a biometric camera located on the mobile device offset from the NIR light source, the biometric camera comprising: an extended depth of field (EDOF) imaging lens; a bandpass filter located adjacent to the EDOF imaging lens to reject ambient light during image capture; and an imaging sensor located adjacent the bandpass filter that converts an optical image of an object into an electronic signal for image processing; and a processor configured to receive video images of an iris of a user from the image sensor, and attempt to match the video images of the iris with previously registered images stored in an iris database, wherein if a match is found, the user is authenticated.
- These and/or other features and utilities of the present general inventive concept will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
-
FIG. 1 is a block diagram illustrating an exemplary embodiment of a mobile device biometric camera system; -
FIG. 2 is a block diagram illustrating components of the biometric camera for a mobile device according to an exemplary embodiment; -
FIG. 3 a diagram illustrating the captured image bounded by a frame in which a window of interest containing an iris has been identified; -
FIGS. 4A, 4B and 4C are diagrams showing one embodiment in which a shutter is patterned into left and right halves; and -
FIGS. 5A and 5B are diagrams showing an example of the shutter subdivided into more independently addressable areas so the biometric camera can be used as a plenoptic camera. - Reference will now be made in detail to the embodiments of the present general inventive concept, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. The embodiments are described below in order to explain the present general inventive concept while referring to the figures.
- Advantages and features of the present invention and methods of accomplishing the same may be understood more readily by reference to the following detailed description of embodiments and the accompanying drawings. The present general inventive concept may, however, be embodied in many different forms and should not be construed as being limited to the embodiments set forth herein. Rather, these embodiments are provided so that this disclosure will be thorough and complete and will fully convey the concept of the general inventive concept to those skilled in the art, and the present general inventive concept will only be defined by the appended claims. In the drawings, the thickness of layers and regions are exaggerated for clarity.
- The use of the terms “a” and “an” and “the” and similar referents in the context of describing the invention (especially in the context of the following claims) are to be construed to cover both the singular and the plural, unless otherwise indicated herein or clearly contradicted by context. The terms “comprising,” “having,” “including,” and “containing” are to be construed as open-ended terms (i.e., meaning “including, but not limited to,”) unless otherwise noted.
- The term “component” or “module”, as used herein, means, but is not limited to, a software or hardware component, such as a field programmable gate array (FPGA) or an application specific integrated circuit (ASIC), which performs certain tasks. A component or module may advantageously be configured to reside in the addressable storage medium and configured to execute on one or more processors. Thus, a component or module may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables. The functionality provided for the components and components or modules may be combined into fewer components and components or modules or further separated into additional components and components or modules.
- Unless defined otherwise, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It is noted that the use of any and all examples, or exemplary terms provided herein is intended merely to better illuminate the invention and is not a limitation on the scope of the invention unless otherwise specified. Further, unless defined otherwise, all terms defined in generally used dictionaries may not be overly interpreted.
- According to the method and system disclosed herein, the exemplary embodiments provide a mobile device with a biometric camera system for reliable iris identification and user authentication. The camera is capable of working at high levels of ambient illumination, capturing images practically without motion blur, provides considerable latitude in distances at which a clear image can be captured and can be realized in a small form factor. Additionally, the camera can optionally perform functions including proximity sensing, night-vision camera, 3D time-of-flight sensor, eye position and gaze tracking camera, and structured light for 3D sensing. Combined with a user-facing front-face camera, the system can automatically identify whether the left and/or right eye is being imaged.
-
FIG. 1 is a block diagram illustrating an exemplary embodiment of a mobile device biometric camera system. The system includes amobile device 10 having conventional components including amemory 12, at least oneprocessor 14, input output devices (I/O) 16, an front-facing camera 18 (e.g., RGB camera), and adisplay 22. - The
memory 12, theprocessor 14, the I/O 16, the front-facingcamera 18 and thedisplay 22 may be coupled together via one or more system buses (not shown). Thememory 12 may comprise one or more memories comprising different memory types, including RAM, ROM, cache, virtual memory and flash memory, for example. Theprocessor 14 may include a single processor having one or more cores, or multiple processors having one or more cores. The I/O 16 is a collection of components that input information and output information. Example components comprising the I/O 16 may include a microphone, speaker, and a wireless network interface controller (or similar component) for communication over the network. Theprocessor 14 may execute an operating system (OS) that manages hardware resources and performs basic tasks. Examples of the OS may include Symbian™, BlackBerry OS™, iOS™, Windows™, and Android™. In one embodiment, thedisplay 22 may be integrated with themobile device 10, while in another embodiment, thedisplay 22 may be external from themobile device 10. - In one embodiment, the
mobile device 10 may comprise any type of mobile device form factor, including but not limited to, a cell or smart-phone, a tablet, a notebook or laptop computer, a television, and a wearable computer, for example. In one embodiment, themobile device 10 may be implemented with thedisplay 22 and the front-facingcamera 18 located on the same side of themobile device 10, such that the front-facingcamera 18 is pointed at the user has the user holds the device to view thedisplay 22. In the embodiment where themobile device 10 comprises a laptop or notebook, the front-facing camera is typically housed within a lid of the laptop. - According to the exemplary embodiment, the
mobile device 10 is provided with abiometric camera system 24 that is used to capture images of the human iris for user identification and authentication. In one embodiment, thebiometric camera system 24 may include a near infrared (NIR)light source 20, abiometric camera 28, aniris recognition component 30 and aniris database 32. In one embodiment, theiris recognition component 30 and theiris database 32 may be software components stored in thememory 12 and executed by theprocessor 14. - As shown, in one embodiment the
biometric camera 28 may be located in one corner of the mobile device 10 (although other locations are also possible), while theNIR light source 20 may be located in an opposite corner to offset theNIR light source 20 from thebiometric camera 28 within the body of themobile device 10. In one embodiment, theNIR light source 20 may be implemented using micro light emitting diodes (LEDs) or a laser diode, and theNIR light source 20 may be slightly angled upwards to point into the user's eye located at an expected distance when held normally by the user. - In operation, the NIR
light source 20 flashes a user of themobile device 10 with near infrared light during image capture of the user's iris by thebiometric camera 28. Video images output from thebiometric camera 28 are received by theiris recognition component 30, which attempts to match the image of the iris with previously registered images stored in theiris database 32. If a match is found, then the user may be authenticated. -
FIG. 2 is a block diagram illustrating components of thebiometric camera 28 according to an exemplary embodiment. In one embodiment, thebiometric camera 28 may operate to gather reflected near infrared light and form an image of a human iris using a combination of animaging lens 200, a bandpass filter 202, and animaging sensor 204. According to the exemplary embodiment, theimaging lens 200 may comprise an extended depth of field (EDOF) imaging lens. - The bandpass filter 202 may be located adjacent to the
imaging lens 200 and functions to reject ambient light. In a preferred embodiment, the bandpass filter 202 may be combined with a solid-state shutter. In one embodiment, the band pass filter may comprise a near-infrared bandpass filter. As shown with the dashed lines, thebiometric camera 28 may also include an optional visiblelight filter 206 located adjacent to the bandpass filter 202 and optional solid-state shutter that may be used to further reject visible light. - The
imaging sensor 204 may be located adjacent to the bandpass filter 202 and converts an optical image of an object (e.g., an iris) into an electronic signal for subsequent image processing. In one embodiment, theimage sensor 204 may include a built-in rolling shutter or a freeze-frame shutter. In one embodiment, theimage sensor 204 may comprise a digital charge-coupled device (CCD) or complementary metal—oxide—semiconductor (CMOS) active pixel sensors. - In one embodiment, the
biometric camera system 28 is designed to have a very small size to fit into amobile device 10, such as a smartphone, and to also be low in cost. To meet these design goals, in one embodiment theimage sensor 204 may comprise a CMOS sensor with a small pixel pitch, e.g., 1.1 μm, and a rolling shutter (not freeze-frame). Such a sensor type is commercially available and used in smartphones for front and back-facing cameras. Thus, the preferred embodiment utilizes an off-the-shelf black-and-white CMOS sensor with rolling shutter and small pixels. - The bandpass filter 202 may function to reject all wavelengths except near-infrared, e.g., 800 nm+−5 nm. In a preferred embodiment, the bandpass filter 202 may be combined a solid-state shutter. The process of capturing an image comprises:
-
- 1) Starting the image sensor in a mode where the image sensor is exposing the entire frame simultaneously;
- 2) Opening the solid-state shutter simultaneously while pulsing the NIR light source;
- 3) Closing the solid-state shutter; and
- 4) Reading out the captured image from the image sensor.
- One goal may be to make the exposure time (when the solid-state shutter is open and the NIR light source is enabled) to be as short as practically possible. To compensate for the short exposure time, the NIR light source power may be made as high as practically possible without damaging the human eye. By keeping the solid-state shutter closed at all times except during capture, the ambient light received from elsewhere in the scene may be significantly reduced. This achieves the goal of rejecting ambient illumination. Also, due to the very short exposure time, the goal of practically motion blur-free image capture may be achieved.
- In one embodiment, the solid-state shutter may comprise gallium arsenide (GaAs). This type of shutter can change its state from opaque to transparent in a very short time (several nanoseconds) by applying voltage externally. Using a GaAs-based shutter as a filter is particularly desirable in this application since this type of shutter has a high index of refraction. Typical dichroic (“interference”) optical bandpass filters often have a relatively wide bandpass width due to the dependence of the central wavelength on the angle of incidence. In camera systems, light rays can pass through the optical filter at various angles. Since the bandpass wavelength shifts as a function of incoming light ray angle, it is typical to make the bandpass wide, such that regardless of the incoming ray angle, light energy of the desirable wavelength in that ray can still pass through the optical filter. The amount of shift of the center wavelength is determined as
-
- Where λ c is the central wavelength, λ0 is the central wavelength at normal incidence, and n* is the filter effective index of refraction. Since GaAs has a high n*, the shift effect is reduced, the filter can be designed to have a narrower bandwidth—and thus reject even more ambient light.
- Alternatively, instead of using a solid state shutter, a CMOS sensor with a built-in freeze-frame shutter and NIR bandpass shutter (possibly also GaAs) may be used. However, CMOS sensors with a freeze-frame shutter typically have pixels of larger size compared to CMOS sensors utilizing rolling shutter. Larger pixel size reduces image resolution—or increases sensor size and cost. Larger pixel size also reduces the depth of field, as explained below. This is why the use of GaAs solid shutter is preferred.
- A solid-state shutter, such one based on GaAs, may still leak some light in its closed state. If the
image sensor 204 continues sensing an image while the image is read out, which is typical in image sensors utilizing rolling shutter, the leaking ambient light can register in the captured image and adversely affect the image quality, e.g., by saturating some image areas or adding noise to the image. In one embodiment, the processor may be configured to reduce exposure from ambient light by reducing exposure and readout time as follows: -
- 1) Start exposure of the
image sensor 204, in particular in sensors with a rolling shutter, using a global reset to clear charge from all pixels simultaneously, thereby reducing time during which ambient light can be sensed; and - 2) Reduce the readout time by identifying iris position from a previously captured image and defining a window of interest around the iris position, and subsequently capturing and reading out from a portion of the image sensor defined by the window of interest. For example, as shown in
FIG. 3 , a previously captured image is shown bounded byframe 300, and a window ofinterest 302 is identified containing the iris, which is smaller than the entire frame.
- 1) Start exposure of the
- In one embodiment, the NIR light source may be synchronously pulsed with the readout of the rolling shutter. In this embodiment, the rolling shutter may be addressed in a row-column manner where only the pixels in the window of
interest 302 are read out. For example, the NIR light source may turned on at or slightly before the time where the upper left of the window ofinterest 302 is read out and then turned off after the lower right pixel of the window ofinterest 302 has been read. In a particular embodiment, the rejection of ambient light can be made greater by pulsing at a higher light intensity. The total and instantaneous amounts of NIR light must be below certain thresholds to avoid eye damage. Because the NIR light is only turned on while exposing and reading thewindow 302, the intensity of the light may be increased during this period than would be allowed for eye safety if it were used while exposing and reading theentire frame 300. The increased intensity of light enables a shorter exposure and readout time, leading to less ambient light readout. - In one embodiment, a rolling shutter is used to read entire lines rather than a
small window 302. In this case, the shutter resets the pixels and reads out at the top of thewindow 302 and finishes reading at the bottom of the window. In one embodiment, the NIR light may be turned on and off to synchronize with the beginning and end of this readout. As this slice of theframe 300 is smaller than the full frame, the intensity of the pulsed NIR may be higher than if exposing the full frame. In a particular embodiment, the NIR light may be further pulsed to synchronize with each line readout. Here, although the full lines are being read out, the NIR is only turning on while exposing the window ofinterest 302. The instantaneous intensity may be made even higher in this embodiment than in the previous one, even though the total NIR flux may be reduced. - Depth of Field
- The
biometric camera 28 should typically capture an iris image at distances of around 25 cm. At such short distances depth of field is typically limited and it is difficult to obtain a sharp image. To increase the depth of field, the exemplary embodiments may include: - 1) Using EDOF (extended depth of field) techniques, in particular wavefront coding type of EDOF, preferably combined with a fixed-focus lens;
- 2) Reducing lens aperture, while boosting light source power to compensate;
- 3) Reducing pixel size, while boosting light source power to compensate; and
- 4) Using an auto-focus actuator. Although the main purpose of the auto-focus actuator is to focus speedily on objects at “macro” range, e.g. around 10-50 cm, for iris scanning, the actuator may also focus at longer distances up to infinity for applications described below.
- It should be noted, EDOF systems were historically utilized for color capture and experienced some issues capturing and correcting images across all visible wavelengths, resulting in degradation of image quality. Since the present embodiments are capturing a monochromatic NIR image, the issues related to operation across a wide range of wavelengths are eliminated and system performance can improve. Additionally, unlike color sensors typically utilizing Bayer pattern where only one of three colors is sampled at each pixel location, the monochrome sensor samples the image at every pixel location, thus providing EDOF with full information for improved system performance.
- Although EDOF systems increase the depth of focus to improve the range where iris pictures may be taken, they also tend to decrease the MTF of the picture. In one embodiment, the image quality may be improved by using known super-resolution techniques to capture several images of the iris and process them to achieve a sharper picture. To achieve this, several pictures may be taken in succession with the sub-window rolling shutter to quickly capture the window of
interest 302 multiple times. If there is movement between the frames, image registration can be done before combining. The combining step may use a super-resolution technique, or could instead do a simple signal averaging to reduce noise. In a particular embodiment, this may be combined by synchronizing the NIR pulse with reading out of the window ofinterest 302 to achieve faster readouts of the window and ameliorating effects of movement within the frame. - Additional Functions
- In a further aspect of the exemplary embodiment, the architecture of the
biometric camera system 24 enables thebiometric camera system 24 to perform additional functions in addition to iris scanning, including proximity sensing, night-vision, 3D time-of-flight sensing, eye position and gaze tracking, and providing a structured light for 3D sensing, as further described below. - Proximity Sensor
- Proximity sensors utilize an imaging sensor and a NIR light source. The image sensor typically comprises a single pixel (not array of multiple pixels) and the proximity operation may comprise:
-
- 1) Capturing a first signal with the image sensor and measuring a first intensity of ambient light;
- 2) Enabling the NIR light source;
- 3) Capturing a second signal with the image sensor and measuring a second intensity of ambient and any reflected NIR light from the object/subject in proximity;
- 4) Disabling the NIR light source; and
- 5) Determining a difference between the first and second intensity of ambient light measurements, where a signal difference surpassing a threshold indicates an object present in proximity. In the exemplary embodiment, the sensor with shutter and NIR light captures pictures with and without an NIR light source flash, (optionally) sums pixel outputs in each image and subtracts the two values.
- Note that proximity sensors are typically designed to have a high dynamic range, such that they work outdoors in high ambient light. When a GaAs shutter is used, the shutter can open for a very short time (synchronized with the NIR light source pulsing) and thus greatly reduce the exposure from the ambient light. Therefore, the high dynamic functionality may not be necessary, thus potentially reducing imager cost and increasing image quality.
- Typically, every smartphone is equipped with a proximity sensor to turn off the touch screen while the user is holding the phone against his/her ear to prevent accidental touch input. Thus, importantly, since the
biometric camera system 24 may perform the function of the proximity sensor, the proximity sensor may be removed from the phone, and replaced with thebiometric camera system 24 without sacrificing the functionality and without unnecessarily growing smartphone size and cost. Exemplary embodiments also may reuse an already-existing NIR light source for both proximity sensing and biometric scanning. - Night Camera
- Since the
biometric camera 28 utilizes NIR and an NIRlight source 20, thebiometric camera 28 can capture images in darkness. Thesystem 24 can also detect a user in total darkness, scan an iris, capture user gestures and so on. Utilizing an auto-focus actuator instead of a fixed-focus (preferably EDOF) lens may be beneficial to be able to clearly capture objects distances further than 25 cm. - 3D Time-of-Flight Sensor
- An off-the-shelf CMOS camera equipped with a fast GaAs shutter and NIR light source is capable of capturing distance images. For example, a time-of-flight camera is a class of scannerless LIDAR, in which the entire scene is captured with each laser or light pulse. It should be noted that the distance measured to an object (e.g., a face) by the time-of-flight camera can be used for fast focusing using the auto-focus actuator.
- Eye Position Detection and Gaze Tracking Camera
- The availability of NIR
light source 20 and NIRbiometric camera 20 allows capturing iris and eyes continuously to track their location and gaze direction using existing techniques. - Motion Detection
- The NIR
biometric camera 28 can be used to detect motion, including motion in complete darkness, by utilizing imaging techniques known in the field (e.g., capturing frames and comparing their aggregated sharpness). - Structured Light for 3D Sensing
- A patterned slide can be inserted in front of the NIR
light source 20 to implement 3D sensing. The pattern will be reflected off the object being imaged and then captured by thebiometric camera 28. By comparing the relative distortion of the pattern between imager and projected version, the absolute distance of points on the object can be determined. - Typically structured light capture can only be used for imaging static scenes at high resolution because moving objects change position too rapidly to update the depth map without blur. Furthermore, in bright ambient light conditions, structured lighting depth capture systems suffer from wash out.
- In this system, the fast shutter and a strong NIR
light source 20 can mitigate against both of these issues. The problem of ambient wash out may be reduced if another NIR light source is added with a patterned slide that has a complementary pattern to the pattern on the first slide. For example, if the new slide has a pattern inverse to the first pattern, then an object can be rapidly illuminated with the first NIR light source and then the second, synchronized with the fast shutter, and then create a higher resolution pattern by using standard computational imaging operations on the two captured patterns. This would enable, e.g., 3D video capture, or the ability to track small changes in facial expression of a person holding the camera in front-facing mode. - 3D Face Recognition for Biometrics
- In the previous embodiment described above, the system can use the fast shutter and a structured light pattern over the IR illuminator to capture high precision, low blur, 3D depth maps of a person's face, paired with the RGB images of the face, captured from the standard front-facing camera. In this embodiment, the field of view of the NIR camera is designed to be wide enough to capture the entirety of the user's face at the target distance range for the application (typically 10 cm-25 cm). These 3D depth maps and RGB image pairs can be used to create a 3D model of a person's face, which can then be used to create a biometric signature, unique to that user. A benefit of this embodiment is that the same system setup can be used to simultaneously capture the person's iris image and 3D facial image, allowing for multi-factor authentication. Capture of the iris and 3D face may be done with the same IR image, or there may be possibly two IR illuminators, one with the structured light pattern on top of it and the other with no pattern on top. This can potentially simplify the processing of the image which would be needed to remove the structured light pattern from the iris image. The same camera could be used in rapid alternating sequence to capture iris and 3D face images.
- Capture with Addressable Shutter
- In some previous embodiments, the solid-state shutter of the
biometric camera 28 is addressable with a single set of electrodes and covers the entire field of view of the shutter. In one embodiment, electrodes may be patterned on the shutter so that there are two or more separately addressable areas of the shutter, that is, one portion of the shutter could be left open to the wavelengths of light measured while the other portion or portions could remain closed.FIGS. 4A, 4B and 4C are diagrams showing one embodiment in which ashutter 400 is patterned into left and right halves.FIG. 4B is a diagram showing theleft half 402, andFIG. 4C is a diagram showing theright half 404. Here, alternate pictures can be taken with each half successively blocked while the other side remains open. For example, a 3D stereo image may be captured through single lens technology with time-multiplex. A left stereo pair image may be captured with theleft half 402 of theshutter 400 transparent and theright half 404 opaque. A right stereo pair image may be captured with theright half 404 of theshutter 400 transparent and theleft half 402 opaque. The left and right captured images can then be processed using known techniques to provide stereo pair images using a single lens camera. In this application, strong ambient lighting might be used to capture the images. In a particular embodiment, the NIR light source may also be used to supplement or replace ambient light. The NIR pulses may be synchronized to the timing of the shutter halves while increasing exposure lighting without exceeding eye safety limits. - In another embodiment, the shutter of the
biometric camera 28 may be further subdivided into more independently addressable areas to create a patterned shutter with time-multiplex for the capture of light-field photos.FIGS. 5A and 5B are diagrams showing an example of theshutter 500 subdivided into more independentlyaddressable areas 502 so the biometric camera can be used as a plenoptic camera. In this example, successive frames may be taken, each with on of theareas 502 of theshutter 500 opened during its frame capture. The series of captured frames can then be processed as a light field image set. AlthoughFIG. 5B shows an example of only a singleaddressable area 502 open at one time for a frame capture, in another embodiment,multiple areas 502 of the shutter could be open at the same time in various patterns. - Combined Use with RGB Camera
- Since the
biometric camera 28 may have a relatively narrow field of view (˜30 degrees horizontal), it may not always capture the image of the entire user face, but only an image of one eye. To decide which eye this is—left or right—the output of the front-facingcamera 18 may be combined to determine which of the eyes is being imaged by thebiometric camera 28. The image from the front-facingcamera 18 may provide information regarding the position of the face with respect to the field of view imaged by thebiometric camera 28. - Other combined uses include:
- 1) The time-of-flight operation can be used to assist the front-facing
camera 18 for focusing; - 2) The front-facing
camera 18 can also automatically capture user's face during the iris scan, e.g. for reference purposes; - 3) The front-facing
camera 18 can capture the user's face to concurrently with the iris scan to additionally verify that the user is a living human (e.g. by tracking pulse using known methods based on spatio-temporal filtering and amplification of color); - 4) The front-facing
camera 18 may confirm the presence and determine the position of the user's head and eyes during iris scan to assist recognition in case of problems with head positioning, such that the system can guide the user to position his/her head properly (and also e.g. by showing the user a target to look at on the display or e.g. flash LED on the phone); and - 5) The front-facing camera can detect the presence of glasses, hair and identify other problems preventing successful iris scan. The system (e.g., smartphone) can then direct the user to address these problems, e.g. remove glasses, adjust hair and track and verify that the problems are addressed using the front-facing camera.
- A method and system for a biometric camera system has been disclosed. The present invention has been described in accordance with the embodiments shown, and there could be variations to the embodiments, and any variations would be within the spirit and scope of the present invention. For example, the exemplary embodiment can be implemented using hardware, software, a computer readable medium containing program instructions, or a combination thereof. Software written according to the present invention is to be either stored in some form of computer-readable medium such as a memory, a hard disk, or a CD/DVD-ROM and is to be executed by a processor. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.
Claims (20)
1. A biometric camera system for a mobile device, comprising:
a near infrared (NIR) light source on the mobile device that flashes a user of the mobile device with near infrared light during image capture;
a biometric camera located on the mobile device offset from the NIR light source, the biometric camera comprising:
an extended depth of field (EDOF) imaging lens;
a bandpass filter located adjacent to the EDOF imaging lens to reject ambient light during image capture; and
an imaging sensor located adjacent the bandpass filter that converts an optical image of an object into an electronic signal for image processing; and
a processor configured to receive video images of an iris of a user from the image sensor, and attempt to match the video images of the iris with previously registered images stored in an iris database, wherein if a match is found, the user is authenticated.
2. The system of claim 1 , wherein the bandpass filter comprises a near infrared bandpass filter.
3. The system of claim 1 , wherein the bandpass filter is combined with a solid-state shutter.
4. The method of claim 3 , wherein the solid-state shutter comprises gallium arsenide (GaAs).
5. The method of claim 3 , wherein the solid-state shutter has at least two separately addressable shutter areas, such that successive frames may be captured, each with at least one of the separately addressable shutter areas opened during the frame capture.
6. The system of claim 1 , further comprising a visible light filter located adjacent to the bandpass filter that rejects visible light.
7. The system of claim 1 , wherein the image sensor further includes at least one of a rolling shutter and a freeze-frame shutter.
8. The system of claim 1 , wherein the processor is further configured to reduce exposure from ambient light by:
starting exposure of the image sensor; and
reducing readout time by identifying an iris position on a user from a previously captured image and defining a window of interest around the iris position, and subsequently capturing and reading out from a portion of the image sensor defined by the window of interest.
9. The system of claim 1 , wherein the image sensor comprises a rolling shutter and wherein the NIR light source is synchronously pulsed with the readout of the rolling shutter.
10. The system of claim 1 , wherein the processor is further configured to implement a proximity sensor operation by:
capturing a first signal with the image sensor and measuring a first intensity of ambient light;
enabling the NIR light source;
capturing a second signal with the image sensor and measuring a second intensity of ambient light and any reflected NIR light;
disabling the NIR light source; and
determining a difference between the first and second intensity of ambient light measurement, wherein a signal difference surpassing a threshold indicates an object present in proximity.
11. The system of claim 1 , wherein the NIR light is covered with a structured light pattern to implement 3D sensing.
12. The system of claim 1 , wherein the NIR light source and the biometric camera are used for at least one of: a night camera, a 3D time-of-flight sensor, eye position detection and gaze tracking; and motion detection.
13. A biometric camera system for a mobile device, comprising:
a front-facing camera;
a near infrared (NIR) light source on the mobile device that flashes a user of the mobile device with near infrared light during image capture, the NIR light source being covered with a structured light pattern;
a biometric camera located on the mobile device offset from the NIR light source, the biometric camera comprising:
an extended depth of field (EDOF) imaging lens;
a bandpass filter located adjacent to the EDOF imaging lens to reject ambient light during image capture; and
an imaging sensor located adjacent the bandpass filter that converts an optical image of an object into an electronic signal for image processing; and
a processor configured to:
use the front-facing camera to capture RGB images of a user's face;
use the NIR light source with the structure light pattern and the biometric camera to capture one or more 3D depth maps of the user's face;
use the RGB images and the one or more 3D depth maps to create a 3D model of the user's face; and
use the 3D model to create a biometric signature for the user.
14. The system of claim 13 , wherein the biometric camera system is used to simultaneously capture an image of the user's iris to perform multi-factor authentication based on iris detection and 3D model detection.
15. The system of claim 13 , wherein the bandpass filter comprises a near infrared bandpass filter.
16. The system of claim 13 , wherein the bandpass filter is combined with a solid-state shutter.
17. The method of claim 16 , wherein the solid-state shutter comprises gallium arsenide (GaAs).
18. The method of claim 16 , wherein the solid-state shutter has at least two separately addressable shutter areas, such that successive frames may be captured, each with at least one of the separately addressable shutter areas opened during the frame capture.
19. The system of claim 13 , further comprising a visible light filter located adjacent to the bandpass filter that rejects visible light.
20. The system of claim 13 , wherein the image sensor further includes at least one of a rolling shutter and a freeze-frame shutter.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/500,674 US9971937B1 (en) | 2013-09-30 | 2014-09-29 | Biometric camera |
US15/968,578 US10311298B2 (en) | 2013-09-30 | 2018-05-01 | Biometric camera |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361884778P | 2013-09-30 | 2013-09-30 | |
US201361885291P | 2013-10-01 | 2013-10-01 | |
US201361913620P | 2013-12-09 | 2013-12-09 | |
US14/500,674 US9971937B1 (en) | 2013-09-30 | 2014-09-29 | Biometric camera |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/968,578 Continuation US10311298B2 (en) | 2013-09-30 | 2018-05-01 | Biometric camera |
Publications (2)
Publication Number | Publication Date |
---|---|
US20180121724A1 true US20180121724A1 (en) | 2018-05-03 |
US9971937B1 US9971937B1 (en) | 2018-05-15 |
Family
ID=53033511
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/500,674 Active 2037-01-10 US9971937B1 (en) | 2013-09-30 | 2014-09-29 | Biometric camera |
US15/968,578 Active US10311298B2 (en) | 2013-09-30 | 2018-05-01 | Biometric camera |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/968,578 Active US10311298B2 (en) | 2013-09-30 | 2018-05-01 | Biometric camera |
Country Status (2)
Country | Link |
---|---|
US (2) | US9971937B1 (en) |
KR (1) | KR102270674B1 (en) |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170142101A1 (en) * | 2015-11-16 | 2017-05-18 | International Business Machines Corporation | Secure Biometrics Matching with Split Phase Client-Server Matching Protocol |
US20170337440A1 (en) * | 2016-01-12 | 2017-11-23 | Princeton Identity, Inc. | Systems And Methods Of Biometric Analysis To Determine A Live Subject |
US20180150691A1 (en) * | 2016-11-29 | 2018-05-31 | Alibaba Group Holding Limited | Service control and user identity authentication based on virtual reality |
US20180165537A1 (en) * | 2016-12-12 | 2018-06-14 | Princeton Identity, Inc. | Systems And Methods Of Biometric Analysis |
US20180357470A1 (en) * | 2017-06-08 | 2018-12-13 | Stmicroelectronics, Inc. | Biometric ear identification |
US10324492B2 (en) * | 2015-09-14 | 2019-06-18 | Inodyn Newmedia Gmbh | Smartphone with front camera and maximized display screen |
US20190313007A1 (en) * | 2018-04-06 | 2019-10-10 | Samsung Electro-Mechanics Co., Ltd. | Infrared camera module, image sensor thereof, and electronic device |
US10445896B1 (en) | 2016-09-23 | 2019-10-15 | Apple Inc. | Systems and methods for determining object range |
US20190373171A1 (en) * | 2018-05-29 | 2019-12-05 | Sharp Kabushiki Kaisha | Electronic device, control device, method of controlling the electronic device, and storage medium |
US20200026389A1 (en) * | 2018-07-19 | 2020-01-23 | Suzhou Maxpad Technologies Co., Ltd | Electronic whiteboard capable of simultaneous writing and projection storage |
US10656275B1 (en) | 2015-09-25 | 2020-05-19 | Apple Inc. | Remote sensing for detection and ranging of objects |
WO2020101641A1 (en) * | 2018-11-12 | 2020-05-22 | Hewlett-Packard Development Company, L.P. | Camera shutters with light filters |
US10678897B2 (en) * | 2015-04-16 | 2020-06-09 | Tobii Ab | Identification, authentication, and/or guiding of a user using gaze information |
US20200191917A1 (en) * | 2017-08-23 | 2020-06-18 | Panasonic Intellectual Property Management Co., Ltd. | Image processing device, distance detection device, image processing method, and non-transitory storage medium |
US20200202153A1 (en) * | 2018-12-21 | 2020-06-25 | Oath Inc. | Biometric based self-sovereign information management |
EP3608814A4 (en) * | 2018-05-29 | 2020-07-15 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Verification template generation method and system, terminal, and computer apparatus |
US10733275B1 (en) * | 2016-04-01 | 2020-08-04 | Massachusetts Mutual Life Insurance Company | Access control through head imaging and biometric authentication |
US10846388B2 (en) | 2017-03-15 | 2020-11-24 | Advanced New Technologies Co., Ltd. | Virtual reality environment-based identity authentication method and apparatus |
US10956544B1 (en) | 2016-04-01 | 2021-03-23 | Massachusetts Mutual Life Insurance Company | Access control through head imaging and biometric authentication |
US20210097161A1 (en) * | 2017-09-28 | 2021-04-01 | Hewlett-Packard Development Company, L.P. | Magnetic switch |
US11062006B2 (en) | 2018-12-21 | 2021-07-13 | Verizon Media Inc. | Biometric based self-sovereign information management |
US11182608B2 (en) | 2018-12-21 | 2021-11-23 | Verizon Patent And Licensing Inc. | Biometric based self-sovereign information management |
US20210365671A1 (en) * | 2020-05-21 | 2021-11-25 | Dell Products L.P. | Information handling system partial spectrum camera shutter |
US11196740B2 (en) | 2018-12-21 | 2021-12-07 | Verizon Patent And Licensing Inc. | Method and system for secure information validation |
US11271747B2 (en) * | 2019-09-16 | 2022-03-08 | Lawrence Livermore National Security, Llc | Optical authentication of images |
US20220083796A1 (en) * | 2016-06-28 | 2022-03-17 | Intel Corporation | Iris or other body part identification on a computing device |
US11278452B1 (en) * | 2021-03-24 | 2022-03-22 | Stroma Medical Corporation | Systems and methods for eye imaging and position control |
US11281754B2 (en) | 2018-12-21 | 2022-03-22 | Verizon Patent And Licensing Inc. | Biometric based self-sovereign information management |
US11288387B2 (en) | 2018-12-21 | 2022-03-29 | Verizon Patent And Licensing Inc. | Method and system for self-sovereign information management |
US11288386B2 (en) | 2018-12-21 | 2022-03-29 | Verizon Patent And Licensing Inc. | Method and system for self-sovereign information management |
US20220156356A1 (en) * | 2019-03-26 | 2022-05-19 | Nec Corporation | Biometric-information authentication system, biometricinformation authentication method, authentication device, authentication method, measurement device, measurement method, and computer-readable recording medium having program recorded thereon |
US20220276482A1 (en) * | 2016-06-16 | 2022-09-01 | Intel Corporation | Combined biometrics capture system with ambient free infrared |
US11514177B2 (en) | 2018-12-21 | 2022-11-29 | Verizon Patent And Licensing Inc. | Method and system for self-sovereign information management |
US11861939B2 (en) * | 2015-09-09 | 2024-01-02 | Nec Corporation | Guidance acquisition device, guidance acquisition method, and program |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160350607A1 (en) * | 2015-05-26 | 2016-12-01 | Microsoft Technology Licensing, Llc | Biometric authentication device |
KR102388249B1 (en) * | 2015-11-27 | 2022-04-20 | 엘지이노텍 주식회사 | Camera module for taking picture using visible light or infrared ray |
CN108780498B (en) | 2016-01-06 | 2023-04-28 | Lg伊诺特有限公司 | Iris recognition camera system, lens module and iris recognition camera module |
US10547829B2 (en) | 2016-06-16 | 2020-01-28 | Samsung Electronics Co., Ltd. | Image detecting device and image detecting method using the same |
KR102525126B1 (en) | 2016-07-29 | 2023-04-25 | 삼성전자주식회사 | Electronic device comprising iris camera |
US10976549B2 (en) | 2016-09-28 | 2021-04-13 | Magic Leap, Inc. | Face model capture by a wearable device |
IL265734B (en) | 2016-10-05 | 2022-09-01 | Magic Leap Inc | Periocular test for mixed reality calibration |
CN110114777B (en) * | 2016-12-30 | 2023-10-20 | 托比股份公司 | Identification, authentication and/or guidance of a user using gaze information |
CN107424187B (en) * | 2017-04-17 | 2023-10-24 | 奥比中光科技集团股份有限公司 | Depth calculation processor, data processing method and 3D image device |
CN107506696A (en) * | 2017-07-29 | 2017-12-22 | 广东欧珀移动通信有限公司 | Anti-fake processing method and related product |
KR20190017551A (en) | 2017-08-11 | 2019-02-20 | 권민정 | Method and apparatus for user authentication |
KR102507746B1 (en) * | 2018-03-02 | 2023-03-09 | 삼성전자주식회사 | Method for generating plural information using camera to sense plural wave bandwidth and apparatus thereof |
US11508249B1 (en) | 2018-03-05 | 2022-11-22 | Intelligent Technologies International, Inc. | Secure testing using a smartphone |
WO2020023542A1 (en) | 2018-07-24 | 2020-01-30 | Magic Leap, Inc. | Display systems and methods for determining registration between a display and eyes of a user |
WO2021119212A1 (en) | 2019-12-09 | 2021-06-17 | Magic Leap, Inc. | Systems and methods for operating a head-mounted display system based on user identity |
KR102320723B1 (en) * | 2019-12-20 | 2021-11-02 | 라인플러스 주식회사 | Method and system for verifying users |
US20210358149A1 (en) * | 2020-05-18 | 2021-11-18 | Nec Laboratories America, Inc | Anti-spoofing 3d face reconstruction using infrared structure light |
US11523043B2 (en) * | 2020-10-12 | 2022-12-06 | Apple Inc. | Camera autofocus using time-of-flight assistance |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130236066A1 (en) * | 2012-03-06 | 2013-09-12 | Gary David Shubinsky | Biometric identification, authentication and verification using near-infrared structured illumination combined with 3d imaging of the human ear |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7369759B2 (en) | 2003-03-27 | 2008-05-06 | Matsushita Electric Industrial Co., Ltd. | Eye image pickup apparatus, iris authentication apparatus and portable terminal device having iris authentication function |
WO2007011375A1 (en) | 2004-09-13 | 2007-01-25 | Cdm Optics, Inc. | Iris image capture devices and associated systems |
KR100729280B1 (en) | 2005-01-08 | 2007-06-15 | 아이리텍 잉크 | Iris Identification System and Method using Mobile Device with Stereo Camera |
JP4265600B2 (en) * | 2005-12-26 | 2009-05-20 | 船井電機株式会社 | Compound eye imaging device |
EP2891918A1 (en) * | 2008-02-29 | 2015-07-08 | Global Bionic Optics Pty Ltd. | Single-lens extended depth-of-field imaging systems |
US20090245594A1 (en) * | 2008-03-31 | 2009-10-01 | General Electric Company | Iris imaging and iris-based identification |
US20120281280A1 (en) | 2009-10-15 | 2012-11-08 | Sony Corporation | Birefringent device with application specific pupil function and optical device |
KR20120020024A (en) * | 2010-08-28 | 2012-03-07 | 김인선 | Method and apparatus for iris recognition and iridodiagnosi using wireless mobile communications device |
KR20130104682A (en) * | 2012-03-15 | 2013-09-25 | 최상길 | Apparatus and method for automatically locking display and touch in mobile phone |
-
2014
- 2014-09-29 KR KR1020140130189A patent/KR102270674B1/en active IP Right Grant
- 2014-09-29 US US14/500,674 patent/US9971937B1/en active Active
-
2018
- 2018-05-01 US US15/968,578 patent/US10311298B2/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130236066A1 (en) * | 2012-03-06 | 2013-09-12 | Gary David Shubinsky | Biometric identification, authentication and verification using near-infrared structured illumination combined with 3d imaging of the human ear |
Cited By (53)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10678897B2 (en) * | 2015-04-16 | 2020-06-09 | Tobii Ab | Identification, authentication, and/or guiding of a user using gaze information |
US11861939B2 (en) * | 2015-09-09 | 2024-01-02 | Nec Corporation | Guidance acquisition device, guidance acquisition method, and program |
US10324492B2 (en) * | 2015-09-14 | 2019-06-18 | Inodyn Newmedia Gmbh | Smartphone with front camera and maximized display screen |
US10712446B1 (en) * | 2015-09-25 | 2020-07-14 | Apple Inc. | Remote sensing for detection and ranging of objects |
US10656275B1 (en) | 2015-09-25 | 2020-05-19 | Apple Inc. | Remote sensing for detection and ranging of objects |
US20170142101A1 (en) * | 2015-11-16 | 2017-05-18 | International Business Machines Corporation | Secure Biometrics Matching with Split Phase Client-Server Matching Protocol |
US10778672B2 (en) * | 2015-11-16 | 2020-09-15 | International Business Machines Corporation | Secure biometrics matching with split phase client-server matching protocol |
US10762367B2 (en) | 2016-01-12 | 2020-09-01 | Princeton Identity | Systems and methods of biometric analysis to determine natural reflectivity |
US10643087B2 (en) * | 2016-01-12 | 2020-05-05 | Princeton Identity, Inc. | Systems and methods of biometric analysis to determine a live subject |
US10643088B2 (en) * | 2016-01-12 | 2020-05-05 | Princeton Identity, Inc. | Systems and methods of biometric analysis with a specularity characteristic |
US20170337440A1 (en) * | 2016-01-12 | 2017-11-23 | Princeton Identity, Inc. | Systems And Methods Of Biometric Analysis To Determine A Live Subject |
US10943138B2 (en) | 2016-01-12 | 2021-03-09 | Princeton Identity, Inc. | Systems and methods of biometric analysis to determine lack of three-dimensionality |
US10733275B1 (en) * | 2016-04-01 | 2020-08-04 | Massachusetts Mutual Life Insurance Company | Access control through head imaging and biometric authentication |
US10956544B1 (en) | 2016-04-01 | 2021-03-23 | Massachusetts Mutual Life Insurance Company | Access control through head imaging and biometric authentication |
US20220276482A1 (en) * | 2016-06-16 | 2022-09-01 | Intel Corporation | Combined biometrics capture system with ambient free infrared |
US11698523B2 (en) * | 2016-06-16 | 2023-07-11 | Intel Corporation | Combined biometrics capture system with ambient free infrared |
US20220083796A1 (en) * | 2016-06-28 | 2022-03-17 | Intel Corporation | Iris or other body part identification on a computing device |
US11676424B2 (en) * | 2016-06-28 | 2023-06-13 | Intel Corporation | Iris or other body part identification on a computing device |
US10445896B1 (en) | 2016-09-23 | 2019-10-15 | Apple Inc. | Systems and methods for determining object range |
US11348369B2 (en) * | 2016-11-29 | 2022-05-31 | Advanced New Technologies Co., Ltd. | Service control and user identity authentication based on virtual reality |
US20180150691A1 (en) * | 2016-11-29 | 2018-05-31 | Alibaba Group Holding Limited | Service control and user identity authentication based on virtual reality |
US11783632B2 (en) * | 2016-11-29 | 2023-10-10 | Advanced New Technologies Co., Ltd. | Service control and user identity authentication based on virtual reality |
US20220284733A1 (en) * | 2016-11-29 | 2022-09-08 | Advanced New Technologies Co., Ltd. | Service control and user identity authentication based on virtual reality |
US20180165537A1 (en) * | 2016-12-12 | 2018-06-14 | Princeton Identity, Inc. | Systems And Methods Of Biometric Analysis |
US10846388B2 (en) | 2017-03-15 | 2020-11-24 | Advanced New Technologies Co., Ltd. | Virtual reality environment-based identity authentication method and apparatus |
US20180357470A1 (en) * | 2017-06-08 | 2018-12-13 | Stmicroelectronics, Inc. | Biometric ear identification |
US20200191917A1 (en) * | 2017-08-23 | 2020-06-18 | Panasonic Intellectual Property Management Co., Ltd. | Image processing device, distance detection device, image processing method, and non-transitory storage medium |
US20210097161A1 (en) * | 2017-09-28 | 2021-04-01 | Hewlett-Packard Development Company, L.P. | Magnetic switch |
US11481479B2 (en) * | 2017-09-28 | 2022-10-25 | Hewlett-Packard Development Company, L.P. | Magnetic switch |
US20190313007A1 (en) * | 2018-04-06 | 2019-10-10 | Samsung Electro-Mechanics Co., Ltd. | Infrared camera module, image sensor thereof, and electronic device |
EP3608814A4 (en) * | 2018-05-29 | 2020-07-15 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Verification template generation method and system, terminal, and computer apparatus |
US11210800B2 (en) | 2018-05-29 | 2021-12-28 | Shenzhen Heytap Technology Corp., Ltd. | Method, system and terminal for generating verification template |
US20190373171A1 (en) * | 2018-05-29 | 2019-12-05 | Sharp Kabushiki Kaisha | Electronic device, control device, method of controlling the electronic device, and storage medium |
US20200026389A1 (en) * | 2018-07-19 | 2020-01-23 | Suzhou Maxpad Technologies Co., Ltd | Electronic whiteboard capable of simultaneous writing and projection storage |
WO2020101641A1 (en) * | 2018-11-12 | 2020-05-22 | Hewlett-Packard Development Company, L.P. | Camera shutters with light filters |
US11288387B2 (en) | 2018-12-21 | 2022-03-29 | Verizon Patent And Licensing Inc. | Method and system for self-sovereign information management |
US10860874B2 (en) * | 2018-12-21 | 2020-12-08 | Oath Inc. | Biometric based self-sovereign information management |
US11288386B2 (en) | 2018-12-21 | 2022-03-29 | Verizon Patent And Licensing Inc. | Method and system for self-sovereign information management |
US11960583B2 (en) | 2018-12-21 | 2024-04-16 | Verizon Patent And Licensing Inc. | Biometric based self-sovereign information management based on reverse information search |
US20200202153A1 (en) * | 2018-12-21 | 2020-06-25 | Oath Inc. | Biometric based self-sovereign information management |
US11281754B2 (en) | 2018-12-21 | 2022-03-22 | Verizon Patent And Licensing Inc. | Biometric based self-sovereign information management |
US11196740B2 (en) | 2018-12-21 | 2021-12-07 | Verizon Patent And Licensing Inc. | Method and system for secure information validation |
US11062006B2 (en) | 2018-12-21 | 2021-07-13 | Verizon Media Inc. | Biometric based self-sovereign information management |
US11182608B2 (en) | 2018-12-21 | 2021-11-23 | Verizon Patent And Licensing Inc. | Biometric based self-sovereign information management |
US11514177B2 (en) | 2018-12-21 | 2022-11-29 | Verizon Patent And Licensing Inc. | Method and system for self-sovereign information management |
US20220156356A1 (en) * | 2019-03-26 | 2022-05-19 | Nec Corporation | Biometric-information authentication system, biometricinformation authentication method, authentication device, authentication method, measurement device, measurement method, and computer-readable recording medium having program recorded thereon |
US11641282B2 (en) | 2019-09-16 | 2023-05-02 | Lawrence Livermore National Security, Llc | Optical authentication of images |
US11271747B2 (en) * | 2019-09-16 | 2022-03-08 | Lawrence Livermore National Security, Llc | Optical authentication of images |
US20210365671A1 (en) * | 2020-05-21 | 2021-11-25 | Dell Products L.P. | Information handling system partial spectrum camera shutter |
US11892877B2 (en) * | 2020-05-21 | 2024-02-06 | Dell Products L.P. | Information handling system partial spectrum camera shutter |
US20220304850A1 (en) * | 2021-03-24 | 2022-09-29 | Stroma Medical Corporation | Systems and methods for eye imaging and position control |
US11278452B1 (en) * | 2021-03-24 | 2022-03-22 | Stroma Medical Corporation | Systems and methods for eye imaging and position control |
US11931295B2 (en) * | 2021-03-24 | 2024-03-19 | Stroma Medical Corporation | Systems and methods for eye imaging and position control |
Also Published As
Publication number | Publication date |
---|---|
US20180268215A1 (en) | 2018-09-20 |
KR20150037628A (en) | 2015-04-08 |
KR102270674B1 (en) | 2021-07-01 |
US9971937B1 (en) | 2018-05-15 |
US10311298B2 (en) | 2019-06-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10311298B2 (en) | Biometric camera | |
US11335079B2 (en) | Method and system of reflection suppression for image processing | |
JP4702441B2 (en) | Imaging apparatus and imaging method | |
KR102332320B1 (en) | Multi-band biometric camera system having iris color recognition | |
US20210014411A1 (en) | Method for image processing, electronic device, and computer readable storage medium | |
KR102305263B1 (en) | Multiple mode image acquisition for iris imaging | |
US10579871B2 (en) | Biometric composite imaging system and method reusable with visible light | |
WO2021073140A1 (en) | Monocular camera, and image processing system and image processing method | |
JP6564271B2 (en) | Imaging apparatus, image processing method, program, and storage medium | |
CN105094307A (en) | Mobile equipment with front-facing camera | |
JP2019512194A (en) | System and method for HDR video capture using a mobile device | |
Thavalengal et al. | Proof-of-concept and evaluation of a dual function visible/NIR camera for iris authentication in smartphones | |
US10395093B2 (en) | Image processing apparatus, image processing method, and non-transitory computer-readable storage medium | |
US20130265484A1 (en) | Image capturing apparatus and control method therefor | |
KR102506363B1 (en) | A device with exactly two cameras and how to create two images using the device | |
WO2023279289A1 (en) | Processing image data using multi-point depth sensing system information | |
JP2007251532A (en) | Imaging device and face area extraction method | |
JP2017157043A (en) | Image processing device, imaging device, and image processing method | |
JP2014049863A (en) | Image processing device, control method therefor, control program, and imaging device | |
JP2013005018A (en) | Imaging apparatus and imaging method | |
JP2009267733A (en) | Imaging apparatus and imaging control method | |
KR20090090851A (en) | Image capturing apparatus enabling non-contact shutter-operation and image capturing method | |
JP5946785B2 (en) | Visual effect evaluation device, method and program, and image pickup device | |
KR20230061799A (en) | Access guest book using iris recognition function | |
JP2012119912A (en) | Imaging device, image processing device, imaging method, and image processing method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:OVSIANNIKOV, ILIA;WANG, YIBING MICHELLE;PARK, YONG-HWA;AND OTHERS;SIGNING DATES FROM 20141204 TO 20150121;REEL/FRAME:038982/0270 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |