US20190356827A1 - Systems and methods of biometric analysis with adaptive trigger - Google Patents
Systems and methods of biometric analysis with adaptive trigger Download PDFInfo
- Publication number
- US20190356827A1 US20190356827A1 US16/531,488 US201916531488A US2019356827A1 US 20190356827 A1 US20190356827 A1 US 20190356827A1 US 201916531488 A US201916531488 A US 201916531488A US 2019356827 A1 US2019356827 A1 US 2019356827A1
- Authority
- US
- United States
- Prior art keywords
- illumination
- flash
- interest
- region
- biometric analysis
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H04N5/2256—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/689—Motion occurring during a rolling shutter mode
-
- H04N5/2329—
Definitions
- the present disclosure relates to systems and methods of biometric analysis and, in particular, to biometric analysis systems including an adaptive trigger configured to align an object (e.g., a subject, a barcode, or the like) within a camera field-of-view to improve application of flash illumination to the desired area of the object for capture and analysis.
- an adaptive trigger configured to align an object (e.g., a subject, a barcode, or the like) within a camera field-of-view to improve application of flash illumination to the desired area of the object for capture and analysis.
- iris recognition is a well-accepted and accurate means of biometric identification used in government and commercial systems around the world that enables secure transactions and an added layer of security beyond keys and/or passwords. Due to the increased security provided by iris recognition systems, an increase in use of such systems has occurred around the world.
- Video cameras that use complementary metal-oxide-semiconductor (CMOS) or semiconductor charge-coupled device (CCD) image sensors typically use electronic shutters to determine the time period over which the sensor measures light.
- a trigger signal opens the shutter for a predetermined time during which each pixel within the sensor array collects (integrates) incoming light.
- the signal collected (integrated) in each pixel during the exposure remains fixed and is then systematically read out, converted to a digital signal, and processed to become an image. Pixels are then cleared and readied for the next exposure to light.
- FIG. 1 is a diagram of a traditional global shutter camera schedule of events for signal collection, including the timeline for exposure of each row of an image sensor in the global shutter camera.
- Global shutters simultaneously expose an entire array of pixels, e.g., N rows by M columns, during signal collection.
- a single exposure event (during t exp )
- light is simultaneously collected in each pixel.
- the global shutter closes, the light signal within each pixel represents the image during the period of the single exposure. All pixels integrate signal over exactly the same period of time.
- Global shutter cameras avoid flash timing issues incurred when using rolling shutter cameras.
- global shutter cameras are expensive options for biometric analysis, thereby increasing the overall costs associated with biometric analysis systems.
- FIG. 2 is a diagram of a traditional rolling shutter schedule of events for signal collection, including the rolling shutter timeline.
- Rolling shutters expose an array of pixels differently from global shutters.
- a rolling shutter system exposes a first row of pixels for an exposure time (t exp ) and then commences to read-out the exposed row of pixels for digitization.
- the read-out process occupies a unique onboard resource for a period referred to as a read-out time during which no other row can be read-out.
- the rolling shutter exposes the second row of pixels during a time that is equal to but delayed from the first row by a read-out time.
- the second row is thereby exposed to light and ready to be read-out at the moment that the read-out process for the first row is complete.
- the third row is exposed to light for a time interval equal in length to that of the first two rows but delayed relative to the second row by a read-out time allowing for the required time to read-out the second row.
- the process “rolls” down the pixel array reading row-by-row in sequence taking a total time equal to the exposure time for a single row plus the read-out time interval, times the number of rows.
- the time interval during which a row is exposed and therefore the events captured by that row are different for each row for a rolling shutter sensor. This is a key difference from a global shutter sensor, especially when using a short flash.
- the light collection time period for each row of a sensor with a global shutter is simultaneous while the time periods of light collection for each row of a sensor equipped with a rolling shutter are not simultaneous. Rather, light collection time periods for each row of a rolling shutter are offset from one another with a delay between rows equal to the row read-out time.
- the different exposure techniques result in image shearing.
- FIG. 3 shows an image in which a moving fan blade was captured by an image sensor with a global shutter with no or little distortion as compared to the same moving fan captured by an image sensor with a rolling shutter shown in FIG. 4 .
- Image shearing is an inevitable consequence of the row-by-row time delays built into a rolling shutter sensor in which each row “sees” the scene over a slightly different and offset time interval.
- the row-by-row delay in exposure of a rolling shutter also has an effect on coordinating an exposure with flash illumination.
- a flash refers to a short, intense period of illumination of a subject during which the light applied by the flash dominates other sources of light on the scene (e.g., the area surrounding the subject).
- FIG. 5 shows the exposure of each row during t exp with the shaded region indicating the duration of time of the flash illumination occurring simultaneous during t exp .
- the exposure of all the pixels in a sensor can be coordinated with the application of the flash illumination to the scene. For example, if the period of the flash pulse is 1 ms, the global shutter can open simultaneously with the start of the pulse and close simultaneously with the end of the pulse 1 ms later.
- the flash illuminates the pixels during and only during their global exposure.
- Light forming the image is, by assumption, dominated by the light applied to the scene by the flash. If, for example, sunlight is present in the scene, the irradiance on the object from the flash is significantly brighter than that of the sunlight during the flash pulse when pixels are exposed by the global shutter.
- FIG. 6 shows the rolling shutter exposure during t exp for each row extending diagonally across the diagram, and the time period for flash illumination illustrated as the vertical shaded region t p .
- the flash pulse occurs between the start pulse and end pulse points of the diagram. Because the exposure period for each row is delayed from the previous row by a short read-out time interval, illumination of a full frame requires that a flash pulse provide illumination during a period when all rows are integrating light. Failure to meet this condition creates a situation in which some rows of pixels integrate light from the flash pulse while some do not, and perhaps some rows integrate light from only a portion of the flash pulse. In this case, the image is unevenly illuminated.
- some rows are finished integrating before the flash starts and other rows do not start integrating until after the flash ends.
- other rows integrate a partial flash and some integrate the full flash.
- a subset of lines on the rolling shutter sensor receive adequate illumination, but outside of this set of lines, the other parts of the sensor remain largely dark.
- FIG. 6 One example of a short flash pulse can be considered with respect to FIG. 6 , which, across the top horizontal line, four horizontal dashed lines, and bottom lines, respectively shows row numbers 0 , 200 , 450 , 600 , 850 , and 1000 .
- the flash pulse can start as row 200 of 1000 rows finishes integrating the signal and begins to read-out, indicated by the top dashed line.
- the flash pulse can end as row 850 of 1000 begins integrating the signal, indicated by the bottom dashed line.
- FIG. 6 shows that rows 450 through 599 receive the full illumination of the flash pulse, as bracketed by the middle two dashed lines.
- rows 200 to 449 and rows 600 to 849 only receive a portion of the flash illumination while rows outside of these ranges, e.g., rows 1 to 199 and 850 to 1000 , receive no flash illumination. Assuming insignificant ambient light, the resulting image would show an illumination stripe surrounded by dim regions. The transition from bright to dim at the top and bottom of the stripe is due to rows that receive flash illumination over a fraction of the total pulse time.
- FIG. 7 shown a portion of an image acquired using a rolling shutter camera with a delayed flash pulse in which the recorded irradiance is plotted to show dark regions before and after the flash, ramp-up and ramp-down regions of partial illumination, and a plateau region of complete flash illumination. The plateau region of FIG.
- an image captured using a rolling shutter camera would include a horizontal stripe with a vertical height proportional to the duration of the flash illumination. In cases with bright ambient illumination, the un-flashed portion of the image would appear, but might be significantly dimmer if the flash illumination is brighter than the ambient illumination.
- a trigger signal can be initiated by the sensor controller to fire the flash at a preset time relative to the start of image writing.
- the flash can fire when the first line is written and can remain on for 50 of 1000 lines.
- the resultant image would be flash illuminated for the top 5% of the image and would be dark elsewhere.
- the same flash can be delayed until the 500 th line of 1000 lines, resulting in an image with a stripe of illuminated content approximately halfway down the frame.
- the photographer would need to align the subject within the camera field-of-view such that the stripe of illumination detected by the sensor corresponds to the position of the desired object.
- an extended period of illumination e.g., a flash pulse that is started simultaneously with the beginning of the exposure of the first row of pixels and is not finished until the last row of pixels has been exposed.
- the extended period of illumination is needed to expose the entire image since image lines are written sequentially rather than all at once (as is the case with a camera including a more expensive and physically larger global shutter sensor).
- This technique necessitates a longer flash pulse compared to the global shutter case, and would show up in FIG. 6 as a shaded region covering all of the rows with a duration equal to the frame time.
- This technique would also illuminate the full frame shown in FIG. 7 . Additional requirements on the flash in terms of power output, heating and reliability are needed based on the longer pulse for this technique.
- a longer pulse might also challenge requirements for eye-safety. For these reasons, full frame pulses with rolling shutters are considered impractical.
- an exemplary biometric analysis system includes one or more illumination sources configured to provide dim illumination to a scene including an object, and further configured to provide flash illumination to the object in the scene.
- the dim illumination can be provided by an illumination source external and separate from the biometric analysis system, such as ambient light, sunlight, any other light source, or the like (e.g., one or more of the illumination sources can be ambient light).
- the biometric analysis system can include a single illumination source that provides the dim illumination, with the same illumination source providing the flash illumination at the determined time period.
- the biometric analysis system can include a first illumination source that provides the dim illumination, and a second (separate) illumination source that provides the flash illumination.
- the first illumination source can continue to provide the dim illumination during the flash illumination from the second illumination source. In some embodiments, the first illumination source can be automatically actuated into a non-illuminating configuration during the flash illumination provided by the second illumination source, and automatically actuated into an illuminating configuration after the flash illumination is complete.
- the biometric analysis system includes a rolling shutter camera configured to capture one or more images.
- the rolling shutter camera generally includes a frame with a field-of-view.
- image as used herein can include still frame images, video, combinations thereof, or the like.
- the biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device.
- the adaptive trigger module when executed, can be configured to analyze the scene to detect the object in the scene during dim illumination of the scene.
- the adaptive trigger module when executed, can be configured to determine a position in the frame of the rolling shutter camera that coincides with the detected object in the scene.
- the adaptive trigger module when executed, can be configured to arrange a delay (a time delay) between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
- the adaptive trigger module when executed, can be configured to track movement of the object within a field-of-view of the rolling shutter camera. In such embodiments, the adaptive trigger module can be configured to modify the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources based on detected movement of the object within the field-of-view.
- the adaptive trigger module can be configured to detect a region of interest of the object and arranges the delay such that the stripe of flash illumination coincides with the detected region of interest of the object.
- the region of interest of the object can include one or both eyes of a person.
- the object can be a person.
- the adaptive trigger module includes a face finder configured to detect a face (and/or features of the face) of the person.
- the object can be a physical item.
- the adaptive trigger module can include an identifier finder configured to detect a unique identifier (e.g., a barcode, a quick response (QR) code, combinations thereof, or the like) associated with the physical item.
- a unique identifier e.g., a barcode, a quick response (QR) code, combinations thereof, or the like
- the one or more illumination sources can be configured to provide the flash illumination as a synchronized pulse of flash illumination.
- the flash illumination provided by the one or more illumination sources is brighter than the dim illumination provided by the one or more illumination sources.
- the one or more illumination sources can be near infrared (NIR) illumination sources.
- the adaptive trigger module can be configured to sweep an illuminated stripe down the frame as the rolling shutter camera captures the one or more images, analyze an illuminated section of the one or more images to identify a region of interest in the illuminated section, and stop sweeping of the illuminated stripe when the region of interest is identified.
- an exemplary biometric analysis system includes one or more illumination sources configured to provide dim illumination to a scene including a subject and configured to provide flash illumination to the subject in the scene.
- the biometric analysis system includes a rolling shutter camera configured to capture one or more images.
- the biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device.
- the adaptive trigger module when executed, can be configured to analyze the scene to detect eyes of the subject in the scene during dim illumination of the scene.
- the adaptive trigger module when executed, can be configured to identify the eyes of the subject as a region of interest.
- the adaptive trigger module when executed, can be configured to determine a position in a frame of the rolling shutter camera that coincides with the identified region of interest.
- the adaptive trigger module when executed, can be configured to arrange a flash pulse delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the identified region of interest.
- the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected.
- the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest.
- the region of interest can be tracked by the adaptive trigger module after being detected.
- the flash pulse delay can ensure that the illuminated region of interest is substantially maintained in a center of the frame of the rolling shutter camera.
- the biometric analysis system can include a feedback module configured to be executed by the controller or processing device.
- the feedback module when executed, can be configured to analyze a captured image of the region of interest and determine if the region of interest is illuminated by the stripe of the flash illumination.
- the adaptive trigger module when executed, can be configured to adjust the flash pulse delay based on the determination of the feedback module to ensure that the region of interest is illuminated by the stripe of the flash illumination.
- an exemplary method of biometric analysis includes illuminating a scene and an object in the scene with dim illumination from one or more illumination sources.
- the method includes analyzing the scene with an adaptive trigger module to detect the object in scene during dim illumination.
- the method includes determining a position in a frame of a rolling shutter camera that coincides with the detected object in the scene.
- the method includes arranging a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
- the method can include tracking movement of the object within a field-of-view of the rolling shutter camera with the adaptive trigger module. In such embodiments, the method can include modifying the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources with the adaptive trigger module based on detected movement of the object within the field-of-view.
- an exemplary non-transitory computer-readable medium storing instructions for biometric analysis.
- Execution of the instructions by a processing device causes the processing device to illuminate a scene and an object in the scene with dim illumination from one or more illumination sources.
- Execution of the instructions by a processing device causes the processing device to analyze the scene with an adaptive trigger module to detect the object in scene during dim illumination.
- Execution of the instructions by a processing device causes the processing device to determine a position in a frame of a rolling shutter camera that coincides with the detected object in the scene.
- Execution of the instructions by a processing device causes the processing device to arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
- FIG. 1 (prior art) is a diagram of a traditional global shutter camera schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out.
- FIG. 2 (prior art) is a diagram of a traditional rolling shutter camera schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with delay between rows being equal to read-out time.
- FIG. 3 (prior art) is an image of a moving fan illustrating minimal distortion as captured by a traditional global shutter camera.
- FIG. 4 (prior art) is an image of a moving fan illustrating shearing distortion effects caused by cascading row delays as captured by a traditional rolling shutter camera.
- FIG. 5 is a diagram of a traditional global shutter schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out, with a flash firing during the same time interval as that in which the rows are exposed (t exp ).
- FIG. 6 is a diagram of a traditional rolling shutter schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with a flash firing after the beginning of the frame, illuminating some but not all rows, and some rows partially but others completely, resulting in a ramp-up, plateau and ramp-down periods of flash illumination.
- FIG. 7 is a diagram of a vertical strip of an image (re-oriented horizontally to align with the time axis) showing dark, ramp-up, plateau, ramp-down, and another dark region due to timing of flash in a traditional rolling shutter frame.
- FIG. 8 is a schematic block diagram of an exemplary biometric analysis system in accordance with the present disclosure.
- FIG. 9 is a representation of an image captured by a rolling shutter camera of the exemplary biometric analysis system including a stripe of flash illumination across a region of interest.
- FIG. 10 is a diagrammatic view of an exemplary biometric analysis system in accordance with the present disclosure.
- FIG. 11 is a diagrammatic view of a frame action of a rolling shutter camera of an exemplary biometric analysis system in accordance with the present disclosure.
- FIG. 12 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure.
- FIG. 13 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure.
- FIG. 14 is a block diagram of an exemplary computing device for implementing an exemplary biometric analysis system in accordance with the present disclosure.
- FIG. 15 is a block diagram of an exemplary biometric analysis system environment in accordance with the present disclosure.
- exemplary biometric analysis systems are described herein for a rolling shutter camera that include an adaptive trigger for ensuring that the stripe of light from the flash illumination illuminates a region of interest of an object, such as the eyes of a person, during image writing.
- the biometric analysis systems generally include one or more illumination sources that initially provide dim illumination to a scene to detect the object and identify the region of interest of the object. Based on the identified region of interest, a delay between the start of image writing by the rolling shutter camera and the trigger of the illumination source is arranged or coordinated to ensure that the stripe of the flash illumination coincides with the region of interest of the object.
- the exemplary biometric analysis systems use flash illumination in combination with inexpensive rolling shutter cameras of the type found in low-cost cameras and smartphones.
- the biometric analysis systems address a rolling shutter-specific problem that occurs when the brightness of the ambient light in the scene may need an exposure time less than the full frame time to avoid saturation, and the exposure time is sufficiently short that the first rows of the image are finished integrating light and reading out before the last lines of the image have begun integration.
- a single, short flash pulse will not illuminate an entire frame, leaving portions of the frame relatively dark. Long exposure times can result in motion blur.
- the exemplary biometric analysis systems balance the following parameters to ensure the captured image quality is sufficient for identification, and further ensuring that the region of interest is adequately illuminated by the flash: sensor frame rate, sensor exposure time, flash pulse intensity, flash pulse duration, flash pulse repetition, processed frame rate, and flash pulse offset.
- sensor frame rate over the available ranges increases the speed that the reset and read lines move over the image.
- increasing the frame rate effectively increases the number of exposed lines during that period.
- the execution time of the internal clock of the camera can be increased to operate the camera at a faster rate, thereby capturing a greater number of lines during the exposure time.
- the drawbacks of increased sensor frame rates are additional data transfers and additional heating due to higher clock speed. Lower frame rates may be desirable for dark indoor lighting conditions and higher frame rates may be desirable for outdoor bright lighting conditions.
- Increasing the sensor exposure time increases the distance between the reset and read times. This effectively increases the number of exposed lines during that period, allowing a short flash pulse to cover more lines.
- the drawback of increasing this period is that additional ambient light is accumulated over the entire period.
- the flash pulse needs to be bright enough to overpower direct sunlight for effective outdoor operation.
- the flash pulse is limited by available system power and eye safety limits.
- eye safety can be roughly proportional to the total amount of energy absorbed by the eye over a period of time (e.g., t ⁇ 0.75 ). Reducing the intensity, pulse duration, and pulse repetition all move the system into a safer standing. Longer flash pulses allow more lines to be exposed in a rolling shutter. However, longer pulses can lead to increased eye safety concerns forcing the system to use less intense pulses when longer durations are used.
- one method to reduce the total exposure can be to selectively process frames. While images can be captured at 60 frames per second (fps), only one in four can be illuminated with the flash, which would reduce the exposure to an equivalent of 15 fps with the advantages of capturing at 60 fps. Such technique allows for improved line read speeds, bringing the rolling shutter camera closer to the global shutter camera timing.
- the systems can include a continuous feedback loop to a user who is attempting to align their eye with a small box on the screen (e.g., in situations involving a smaller screen, such as a smartphone).
- the feedback loop allows the system to predetermine which lines should be exposed.
- a low level of continuous, dim illumination can be provided to ensure the user has a reasonable image to complete this task.
- the continuously illuminated frames can be processed to detect specular reflections, and further used to change the flash offset or delay. Such operation allows the user additional margin when attempting to place their eye in the box while keeping the exposure levels to a minimum.
- the biometric analysis systems adjust the portion of the frame that is illuminated such that the important portion of the image (e.g., a region of interest) is well-lit and the unimportant portions of the images are left dim.
- the biometric analysis systems can be used in imaging in the near infrared (NIR) spectrum using a short, intense, eye-safe NIR flash with an NIR filter to freeze motion and emphasize the flash relative to ambient light (including bright sun).
- the biometric analysis systems including the NIR flash with a rolling shutter camera can be used in, e.g., iris biometrics, or the like, in which cost reduction realized by using a rolling shutter instead of a global shutter is a primary driver in component selection.
- the biometric analysis systems include an adaptive step in which dim illumination allows a vision system (e.g., a sensor in a rolling shutter camera including a processor) to detect the position of the desired object in the field-of-view of the camera.
- a vision system e.g., a sensor in a rolling shutter camera including a processor
- the dim illumination remains on for the entire process of capturing an image, and reveals an under-illuminated frame.
- Dim illumination although dimmer than illumination by the flash, is bright enough to perform an automated scene analysis, e.g., face finding, object identification, or the like, that can output a recommendation of a position in the frame that would benefit from flash illumination.
- a controller or processor can arrange a delay between the start of image writing and the flash trigger such that a stripe of illumination coincides with the desired object. If the object moves in the field-of-view, biometric analysis systems can function dynamically to track movement of the object and change the delay accordingly.
- the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected.
- the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest.
- the region of interest can be tracked by the adaptive trigger module after being detected.
- Such arrangements to locate and track the object and indicate to the adaptive trigger module changes in the position of the object permits moving objects to be accurately imaged by the biometric analysis systems.
- the biometric analysis systems disclosed herein can be used for rolling shutter photography (e.g., NIR photography) of objects while outdoors or indoors.
- the biometric analysis systems can acquire NIR images with controlled illumination in full sun or in dim ambient lighting.
- An image of a desired object that fills less than a full frame, e.g., a person's face with the entire person in the field-of-view, can be captured by the biometric analysis systems. Due to the low cost, small size and availability of rolling shutter sensors, incorporation of the biometric analysis systems into cost-sensitive or size-sensitive platforms can benefit cost-sensitive or size-sensitive users.
- the biometric analysis systems can be used in cargo identification in which infrared dye for markings signify authenticity.
- a box marked with writing can be imaged with a small and inexpensive handheld infrared camera that would look for the marking surrounding an NIR watermark.
- the flash trigger can be adjusted to fire in order to illuminate the markings, leaving the rest of the field-of-view above and below the mark of interest in the dark.
- Handheld devices including smartphones, barcode readers, and other mass produced electronic devices, can incorporate the biometric analysis systems at a small and low cost.
- a rolling shutter camera would reduce the cost for such systems, but would require an extended flash to illuminate an entire scene to guarantee capturing the subject of interest.
- the exemplary biometric analysis systems include a rolling shutter camera with an adaptive image content-based trigger that issues a shorter flash of a near infrared light that allows for use of smaller batteries, provides for a longer battery life, provides for improved eye safety, and allows for longer illumination source lifetimes.
- the biometric analysis systems disclosed herein can provide flash illumination for illuminating an object sufficiently to outshine the sun, and in some embodiments can use strobing flash illumination as is disclosed in, e.g., U.S. Pat. Nos. 7,542,628; 7,627,147; 7,657,127; 9,131,141; U.S. Patent Application Publication No. 2015 / 0098630 ; U.S. Patent Application Publication No. 2016 / 0014121 ; U.S. Patent Application Publication No. 2016 / 0012218 ; and U.S. Patent Application Publication No. 2016 / 0012292 , which are incorporated by reference herein.
- the exemplary biometric analysis systems use a rolling shutter camera and a flash illumination that is triggered adaptively to illuminate the region of interest of the scene or object.
- the adaptive trigger can be configured to follow or track the subject, and adjusts the flash timing to maintain the stripe of flash illumination on the subject of interest.
- a face finder can be used as an image analytic to trigger the flash.
- a unique identifier such as a bar code or QR code, can be detected and illuminated by the stripe of flash illumination.
- the biometric analysis system can locate the object in a relatively dimly lit scene, and then initiates the repetitive flash that illuminates the object. Flash timing can be adjusted to maintain the object in the illuminated region of the field-of-view of the rolling shutter camera.
- a further assumption can be made that a flash pulse is started at some time later, e.g., delayed by a time ⁇ t.
- the delay time At the rows of full flash illumination can be adjusted from the top of the image frame to the bottom to focus on a region of interest located within the frame.
- the biometric analysis systems can be used to detect the region of interest within the frame of the rolling shutter camera, and intentionally delay the flash illumination in a coordinated manner to ensure that the region of interest is illuminated by the flash during exposure and read-out of the corresponding rows. Proper adjustment of the delay time can cause the stripe of flash illumination due to the short flash pulse to brighten the region of interest.
- the stripe can scan or search the image in successive frames of a photo or video stream. Such technique can be performed by systematically sweeping an illuminated stripe down the field-of-view and stopping when a region of interest is detected.
- the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest.
- the region of interest can be tracked by the adaptive trigger module after being detected.
- a pre-flash e.g., a single long, low intensity flash that illuminates the full frame and that is both eye-safe and within the operating range of the illumination hardware
- a pre-flash can provide a single, dim image of the scene.
- Automated image analysis of the full scene that can tolerate the low light can detect and output a region of interest that is the width of the image and the height of the flash illumination stripe.
- an automated delay signal can set the illumination stripe to the optimal position to illuminate the region of interest in succeeding frames of a photo or video stream using the full flash brightness over the narrowed region of interest. If the desired feature of the object begins to move out of the illumination stripe (as determined by video analysis performed at the frame rate), the delay can adapt using negative feedback to maintain the position of the desired feature centered within the well-lit region of the image.
- the biometric analysis systems can be used to capture NIR images of an object that is bathed in sunlight. While a fraction of the solar spectrum covers the NIR wavelengths, it can be assumed that filtered solar NIR irradiance (e.g., from 800 nm to 900 nm) is insufficient to illuminate a video frame that is adequately short in duration to suppress motion blur. It can be further assumed that using a long flash that covers the entire rolling shutter frame period is impractical, even with a short exposure time, due to hardware limitations or eye safety concerns. A short NIR flash pulse that can only illuminate a portion of the frame with an adaptive delay time to ensure that the region of interest is illuminated can therefore be used.
- filtered solar NIR irradiance e.g., from 800 nm to 900 nm
- the biometric analysis systems initially detect and identify a region of interest of an object, such as the eyes of a subject.
- Eye finding can be accomplished using a number of different processes, such as full frame NIR illumination with a relatively dim, eye-safe (and hardware-safe) long NIR flash pulse; partial frame illumination due to a short, intense pulse that is automatically swept frame-by-frame over the full field-of-view until the eyes appear in the search; and/or image acquisition by a color (or other spectrum) auxiliary camera that produces a full frame image allowing eye-finding or face-finding, and then translation from the auxiliary camera frame position to a NIR camera position.
- the biometric analysis systems automatically set the flash pulse delay to illuminate a stripe across the frame including the eyes (e.g., the region of interest).
- the delay that sets the vertical position of the illumination stripe can be automatically adjusted to maintain the eyes in substantially the center of the frame. While the portions of the frame outside of the illumination stripe remain unusably dark, the important information about the eyes remains well lit.
- the competing solar illumination may dimly illuminate the rest of the frame. This is not an issue as long as the flash illumination stripe remains resolved and does not saturate due to the sum of the in-band solar irradiance and the applied NIR light. If the illumination stripe nears the saturation level, the intensity of the applied flash can be throttled back. It is preferable to decrease the NIR illumination irradiance on the object rather than reducing the exposure or pulse duration, since the later strategies reduce the portions of the image that are illuminated.
- Any spectrum of light can be used as long as the camera uses a rolling shutter with an in-band flash illuminator.
- visible light NIR, short-wave infrared (SWIR), long-wave infrared (LWIR), ultraviolet (UV), or multi-spectral cameras with rolling shutters that use flash illumination.
- Any process of finding the region of interest can be used as long as the process returns a coordinate to the biometric analysis system that indicates the time delay between the start of frame and the start of the flash.
- use of an eye finder that seeks corneal reflections of the flash could be useful if the region of interest contains one or two eyes of the subject.
- a color face camera could be useful to find the eyes of a subject in iris recognition; but if nose-recognition were used, the region of interest would contain the subject's nose. If the same camera were used to find the region of interest and to acquire the information from the region of interest, a different illumination protocol could be used for each function. For example, a long, relatively dim pulse could illuminate the entire frame, followed by a short, bright pulse to illuminate the region of interest with better fidelity (e.g., improved signal-to-noise ratio). Any process that provides the information needed to set the delay time would also be useful.
- feedback can be used to maintain the region of interest in the illumination stripe.
- negative feedback can be used to correct the position of the region of interest within the illumination stripe. If the object of interest begins to move or slip to the top or bottom of the region of interest illuminated by the flash, feedback can be used to correct the delay time. For example, if the stripe illuminating the region of interest needs to move up in the frame to follow the object of interest, the delay would need to be corrected in one direction. If the object of interest moves down in the frame, the sign of the delay correction would need to be in the opposite direction.
- a feedback signal e.g., an audio alarm, a visual alarm, combinations thereof, or the like
- the width of the flash illumination stripe relative to the frame height can be set by practical considerations. For example, rolling shutter cameras with very short read-out times generally have wider flash illumination stripes, all other things being equal.
- the longest flash pulse can be dictated by practical issues, including available power to the illumination source, operating conditions in power, and temperature of the illumination source. For example, it is known that flashing an illumination source too often can damage it, and reduces the safety of human eyes in and around the illumination source.
- a flash illumination stripe that is narrow relative to the frame height generally should be more accurately timed than a flash illumination stripe that is wider and that can tolerate more motion of the object of interest within the flash illumination stripe.
- system 100 a schematic block diagram of an exemplary biometric analysis system 100 (hereinafter “system 100 ”) is provided.
- the system 100 is configured to detect and identify a region of interest of an object, such as the eyes of a person, and generates a time delay of the flash illumination such that the flash coincides with the region of interest.
- a rolling shutter camera can therefore be used for capturing one or more images for biometric analysis with the delayed flash illumination ensuring that the region of interest will be properly illuminated in the captured images.
- the system 100 is configured to track the region of interest within the frame or field-of-view of the rolling shutter camera, and adjusts the time delay of the flash pulse appropriately based on movement of the region of interest in and out of the frame.
- the system 100 generally includes one or more illumination sources 102 .
- the illumination sources 102 are configured to initially provide dim illumination 104 to a scene including an object to detect and identify the object (or the region of interest of the object).
- the dim illumination 104 can be separate from the system 100 and can be provided by, e.g., ambient light, any other illumination source, sunlight, combinations thereof, or the like.
- the illumination sources 102 are further configured to provide flash illumination 106 (e.g., a pulse of flash illumination) to illuminate the region of interest during exposure of the rows corresponding to the region of interest, thereby ensuring that the region of interest will be illuminated in the captured image.
- a single illumination source 102 can provide both the dim illumination 104 and the flash illumination 106 .
- separate illumination sources 102 can provide the dim illumination 104 and the flash illumination 106 .
- the illumination sources 102 can be NIR illumination sources. It should be understood that the dim illumination 104 is less bright than the flash illumination 106 , and provides sufficient ambient light for the system 100 to detect objects in a scene as viewed by the rolling shutter camera.
- the system 100 includes a rolling shutter camera 108 including a frame 110 with a field-of-view.
- the rolling shutter camera 108 is configured to capture one or more images (e.g., still frame images, video images, combinations thereof, or the like).
- the rolling shutter camera 108 is configured to initially capture an image of the scene and the object in the scene under dim illumination 104 conditions.
- the system 100 includes an adaptive trigger module 112 configured to be executed by a processing device 114 having a processor 116 (e.g., a controller). Although illustrated as separate components from the rolling shutter camera 108 , it should be understood that in some embodiments, the rolling shutter camera 108 can include the illumination sources 102 , the adaptive trigger module 112 , the processing device 114 , or the like.
- the processing device 114 can be part of a central computing system 118 .
- the adaptive trigger module 112 can analyze the captured image of the scene illuminated by the dim illumination 104 and detects the region of interest 120 within the scene. For example, the adaptive trigger module 112 can detect and identify the region of interest 120 as one of more eyes of the user, and stores the rows of the frame 110 corresponding to the region of interest 120 in a database 122 (e.g., a local or remote database).
- a database 122 e.g., a local or remote database
- the adaptive trigger module 112 Based on the rows of the frame 110 corresponding to the region of interest 120 , the adaptive trigger module 112 arranges a flash pulse delay 124 between the start of image writing by the rolling shutter camera 108 and a trigger of the flash illumination 106 such that the flash illumination 108 coincides with the identified region of interest 120 . For example, if the eyes of the user are the region of interest 120 corresponding with rows 400 - 600 of the frame 110 , the adaptive trigger module 112 arranges the flash pulse delay 124 such that the flash illumination 106 is triggered to illuminate the object during rows 400 - 600 .
- the image writing can start before triggering of the flash illumination 106 and the flash illumination 106 is only triggered to illuminate the region of interest 120 .
- Such flash delay 124 ensures that the region of interest 120 is illuminated in the images 126 captured by the rolling shutter camera 108 .
- the resulting image 126 generally includes a stripe of flash illumination 106 extending across the image 126 and over the eyes of the user (see, e.g., FIG. 9 ). Depending on the length of the exposure time, the height of the stripe of flash illumination 106 can be greater or smaller relative to the top and bottom of the image 126 .
- the adaptive trigger module 112 can include a face finder algorithm that locates the face and/or features of the face in the scene.
- the face finder algorithm can therefore locate the face identifies one or more eyes of the face as the region of interest 120 .
- the region of interest 120 can be a unique identifier associated with a physical item that can be detected by an identifier finder algorithm of the adaptive trigger module 112 .
- the identifier finder algorithm of the adaptive trigger module 112 can be used to detect, e.g., a barcode, a QR code, or the like, on a physical item, identifies such unique identifier as the region of interest 120 , and adjusts the flash delay 124 to ensure that the unique identifier is illuminated during capture of the image 126 .
- the adaptive trigger module 112 can be configured to track movement of the object within the field-of-view or frame 110 of the rolling shutter camera 108 .
- the adaptive trigger module 112 can be configured to continuously scan the scene in the frame 110 to detect the position of the region of interest 120 . If the user moves such that the region of interest 120 in the frame 110 is not centered or changes relative to the previously detected rows for the region of interest 120 , the adaptive trigger module 112 can determine the new rows corresponding to the region of interest 120 , stores the new rows in the database 122 , and adjusts the flash delay 124 to coincide with the new rows for the region of interest 120 .
- an object moving in the frame 110 of the rolling shutter camera 108 can be tracked by the adaptive trigger module 112 to allow for the region of interest 120 of a moving object to be properly illuminated by the flash illumination 106 .
- the system 100 includes a feedback module 128 configured to be executed by the processing device 114 and/or the central computing system 118 .
- the rolling shutter camera 108 can include the feedback module 128 .
- the feedback module 128 receives as input the images 126 captured by the rolling shutter camera 108 and analyzes the region of interest 120 captured in the image 126 to determine if the region of interest 120 is illuminated by the stripe of flash illumination 106 . If the region of interest 120 is illuminated by the stripe of flash illumination 106 , the feedback module 128 electronically outputs such findings to the adaptive trigger module 112 and no changes are made in the flash delay 124 .
- the feedback module 128 electronically outputs such findings to the adaptive trigger module 112 .
- the adaptive trigger module 112 analyzes the frame 110 to locate the region of interest 120 , determines the new rows associated with the region of interest 120 , and adjusts the flash pulse delay 124 to trigger the flash illumination 106 at the appropriate rows coinciding with the region of interest 120 .
- the feedback loop can be continuously performed to ensure the region of interest 120 is always illuminated by the stripe of flash illumination 106 .
- the system 100 can include a user interface 130 with a graphical user interface (GUI) 132 for receiving input from a user and for outputting information to the user.
- GUI graphical user interface
- the user interface 130 can be used to initiate the process performed by the system 100 .
- the user interface 130 can output an audio and/or visual alarm to the user when the region of interest 120 is no longer centered in the frame 110 .
- the system 100 can include a communication interface 134 configured to provide a means for electronic transmission between components of the system 100 , e.g., the illumination sources 102 , the processing device 114 , the rolling shutter camera 108 , the adaptive trigger module 112 , the database 122 , the user interface 130 , the central computing system 118 , and feedback module 128 , or the like.
- a communication interface 134 configured to provide a means for electronic transmission between components of the system 100 , e.g., the illumination sources 102 , the processing device 114 , the rolling shutter camera 108 , the adaptive trigger module 112 , the database 122 , the user interface 130 , the central computing system 118 , and feedback module 128 , or the like.
- FIG. 10 is a diagrammatic view of an exemplary biometric analysis system 200 (hereinafter “system 200 ”) of the present disclosure.
- the system 200 includes an illumination source 202 that provides dim illumination to the subject 204 in the scene.
- the illumination source 202 can provide 1 ms NIR pulses to the scene.
- the illumination source 202 can be an NIR light source not connected to the camera 208 .
- the illumination source 202 can be positioned a distance 206 from the subject 204 .
- the illumination source 202 can also be configured to provide the flash illumination to the subject 204 .
- the system 200 includes a camera 208 including a rolling shutter (e.g., a rolling shutter camera, such as a smartphone camera, or the like) configured to capture one or more images of the subject.
- the camera 208 can have a field-of-view represented by area 216 , including the region of interest of the subject 204 .
- the camera 208 includes an illumination source 210 (e.g., an integrated NIR light source) configured to provide flash illumination to a region of interest (e.g., the eyes of the subject 204 ).
- area 212 indicates the coverage of the NIR flash illumination provided by the illumination source 210 , with the region of interest on the subject 204 being within the area 212 .
- the camera 208 can be spaced by a distance 214 from the subject 204 .
- the illumination source 202 appears separate from camera 202 , however, one skilled in the art shall appreciate that, in some embodiments, the illumination source 202 can be incorporated into the camera 208 assembly such that the camera 208 assembly is capable of providing both the dim illumination and flash illumination to the subject 204 .
- FIG. 11 is a diagrammatic view of a frame action 300 of a rolling shutter camera of the exemplary biometric analysis systems disclosed herein.
- the frame action 300 includes a total frame time 302 .
- Ambient light 304 having a low illumination level 306 can be provided for all or part of the total frame time 302 by a dim illumination source and/or sunlight.
- the total camera exposure time 308 indicates when the rolling shutter camera begins and ends image writing and read-out.
- a pulse of flash illumination 310 (e.g., NIR from a light-emitting diode (LED) illumination source) is triggered to coincide with the rows of the region of interest.
- the pulse of flash illumination 310 has a time or pulse width 312 that is significantly smaller than the camera exposure time 308 , and an illumination level 314 that is brighter or higher than the level 306 of the ambient light 304 .
- the region of interest is thereby illuminated by the flash illumination 310 while the remaining rows of the captured image are only illuminated by the ambient light 304 .
- a 1280 ⁇ 960 sensor can be used with the rolling shutter camera.
- a 120 fps VGA (640 ⁇ 480) sensor can be used with the rolling shutter camera to allow for full frame exposure in 4 ms, and can be synchronized with an approximately 4 ms LED flash every 8 th frame.
- the VGA output can provide (unscaled) 15 fps feedback to the user facing the monitor of the system.
- the 4 ms can provide sufficient flash pulse while allowing for cost effective use of an LED and rolling shutter accommodation, and allowing for eye safety.
- a bright light (e.g., approximately 850 nm LED, two LEDs) can be used for the flash pulse.
- the systems can therefore have high pulsed current capabilities.
- the rolling shutter camera can have a high quality camera lens and NIR filtering.
- the NIR filter can be tighter than 100 nm around the LED center wavelength.
- the system can include optics with a modulation transfer function (MTF) optimized for the spatial frequencies that enter the system.
- the system can include preprocessing of iris images for a particular spectrum of spatial frequencies (e.g., low, medium, high, or the like).
- FIG. 12 is a flowchart illustrating an exemplary process 400 of implementing the biometric analysis systems disclosed herein.
- a scene and an object in the scene are illuminated with dim illumination from one or more illumination sources.
- the scene is analyzed with an adaptive trigger module to detect the object (or a region of interest associated with the object) in the scene during dim illumination.
- a position in a frame of a rolling shutter camera is determined with the adaptive trigger module, the position coinciding with the detected object (or region of interest associated with the object) in the scene.
- a delay is arranged by the adaptive trigger module between a start of image writing by the rolling shutter camera and a trigger of the illumination sources such that a stripe of flash illumination provided by the illumination sources coincides with the detected object (or region of interest associated with the object) in the scene.
- movement of the object is tracked within a field-of-view of the rolling shutter camera with the adaptive trigger module.
- the delay between the start of image writing by the rolling shutter camera and the trigger of the illumination sources is modified with the adaptive trigger module based on detected movement of the object (or region of interest associated with the object) out of the field-of-view.
- FIG. 13 is a flowchart illustrating an exemplary process 500 of implementing the biometric analysis systems disclosed herein.
- the adaptive trigger module can set the flash delay at 0 ms, the flash duration equal to the time of the full frame, and the flash brightness to low.
- a full frame image can be acquired with the rolling shutter camera under dim illumination.
- the adaptive trigger module analyzes the image captured at step 504 and locates the region of interest in the image.
- the adaptive trigger module sets the flash delay to a value greater than 0 ms, the flash duration to short (e.g., a pulse), and the flash brightness to high (e.g., higher than the dim illumination) to illuminate the region of interest.
- the adaptive trigger module measures the position of the object of interest within the region of interest (e.g., eyes of the user located on the user's face).
- the process 500 continues a loop to constantly (or at predetermined frequencies of time) ensure that the object is maintained within the region of interest.
- the process 500 performs step 516 to correct the flash delay such that the object is centered within the region of interest.
- the region of interest can thereby be maintained in the flash illumination stripe despite relative motion of the object and/or the camera.
- the process can include a loop that analyzes the captures images to control the intensity of the flash illumination, thereby avoiding saturation of the object.
- FIG. 14 is a block diagram of a computing device 600 in accordance with exemplary embodiments of the present disclosure.
- the computing device 600 includes one or more non-transitory computer-readable media for storing one or more computer-executable instructions or software for implementing exemplary embodiments.
- the non-transitory computer-readable media may include, but are not limited to, one or more types of hardware memory, non-transitory tangible media (for example, one or more magnetic storage disks, one or more optical disks, one or more flash drives), and the like.
- memory 606 included in the computing device 600 may store computer-readable and computer-executable instructions or software for implementing exemplary embodiments of the present disclosure (e.g., instructions for operating the illumination sources, instructions for operating the processing device, instructions for operating the rolling shutter camera, instructions for operating the adaptive trigger module, instructions for operating the communication interface, instructions for operating the user interface, instructions for operating the central computing system, instructions for operating the feedback module, combinations thereof, or the like).
- instructions for operating the illumination sources e.g., instructions for operating the illumination sources, instructions for operating the processing device, instructions for operating the rolling shutter camera, instructions for operating the adaptive trigger module, instructions for operating the communication interface, instructions for operating the user interface, instructions for operating the central computing system, instructions for operating the feedback module, combinations thereof, or the like.
- the computing device 600 also includes configurable and/or programmable processor 602 and associated core 604 , and optionally, one or more additional configurable and/or programmable processor(s) 602 ′ and associated core(s) 604 ′ (for example, in the case of computer systems having multiple processors/cores), for executing computer-readable and computer-executable instructions or software stored in the memory 606 and other programs for controlling system hardware.
- Processor 602 and processor(s) 602 ′ may each be a single core processor or multiple core ( 604 and 604 ′) processor.
- Virtualization may be employed in the computing device 600 so that infrastructure and resources in the computing device 600 may be shared dynamically.
- a virtual machine 614 may be provided to handle a process running on multiple processors so that the process appears to be using only one computing resource rather than multiple computing resources. Multiple virtual machines may also be used with one processor.
- Memory 606 may include a computer system memory or random access memory, such as DRAM, SRAM, EDO RAM, and the like. Memory 606 may include other types of memory as well, or combinations thereof.
- a user may interact with the computing device 600 through a visual display device 618 (e.g., a personal computer, a mobile smart device, or the like), such as a computer monitor, which may display one or more user interfaces 620 (e.g., a graphical user interface) that may be provided in accordance with exemplary embodiments.
- the computing device 600 may include other I/O devices for receiving input from a user, for example, a camera, a sensor, a keyboard or any suitable multi-point touch interface 608 , a pointing device 610 (e.g., a mouse).
- the keyboard 608 and the pointing device 610 may be coupled to the visual display device 618 .
- the computing device 600 may include other suitable conventional I/O peripherals.
- the computing device 600 may also include one or more storage devices 624 , such as a hard-drive, CD-ROM, eMMC (MultiMediaCard), SD (secure digital) card, flash drive, non-volatile storage media, or other computer readable media, for storing data and computer-readable instructions and/or software that implement exemplary embodiments of the biometric analysis systems described herein.
- Exemplary storage device 624 may also store one or more databases 626 for storing any suitable information required to implement exemplary embodiments.
- exemplary storage device 624 can store one or more databases 626 for storing information, such as data relating to captured images 126 under dim illumination 104 and flash illumination 106 , regions of interest 120 , flash delay 124 , combinations thereof, or the like, and computer-readable instructions and/or software that implement exemplary embodiments described herein.
- the databases 626 may be updated by manually or automatically at any suitable time to add, delete, and/or update one or more items in the databases.
- the computing device 600 can include a network interface 612 configured to interface via one or more network devices 622 with one or more networks, for example, Local Area Network (LAN), Wide Area Network (WAN) or the Internet through a variety of connections including, but not limited to, standard telephone lines, LAN or WAN links (for example, 802.11, T1, T3, 56 kb, X.25), broadband connections (for example, ISDN, Frame Relay, ATM), wireless connections, controller area network (CAN), or some combination of any or all of the above.
- LAN Local Area Network
- WAN Wide Area Network
- the Internet through a variety of connections including, but not limited to, standard telephone lines, LAN or WAN links (for example, 802.11, T1, T3, 56 kb, X.25), broadband connections (for example, ISDN, Frame Relay, ATM), wireless connections, controller area network (CAN), or some combination of any or all of the above.
- the network interface 612 may include a built-in network adapter, network interface card, PCMCIA network card, PCI/PCIe network adapter, SD adapter, Bluetooth adapter, card bus network adapter, wireless network adapter, USB network adapter, modem or any other device suitable for interfacing the computing device 600 to any type of network capable of communication and performing the operations described herein.
- the computing device 600 may be any computer system, such as a workstation, desktop computer, server, laptop, handheld computer, tablet computer (e.g., the tablet computer), mobile computing or communication device (e.g., the smart phone communication device), an embedded computing platform, or other form of computing or telecommunications device that is capable of communication and that has sufficient processor power and memory capacity to perform the operations described herein.
- the computing device 600 may run any operating system 616 , such as any of the versions of the Microsoft® Windows® operating systems, the different releases of the Unix and Linux operating systems, any version of the MacOS® for Macintosh computers, any embedded operating system, any real-time operating system, any open source operating system, any proprietary operating system, or any other operating system capable of running on the computing device and performing the operations described herein.
- the operating system 616 may be run in native mode or emulated mode.
- the operating system 616 may be run on one or more cloud machine instances.
- FIG. 15 is a block diagram of an exemplary biometric analysis system environment 700 in accordance with exemplary embodiments of the present disclosure.
- the environment 700 can include servers 702 , 704 configured to be in communication with one or more illumination sources 706 , one or more rolling shutter cameras 708 , one or more adaptive trigger modules 710 , a feedback module 712 , a user interface 714 , and a central computing system 716 via a communication platform 922 , which can be any network over which information can be transmitted between devices communicatively coupled to the network.
- the communication platform 722 can be the Internet, Intranet, virtual private network (VPN), wide area network (WAN), local area network (LAN), and the like.
- the communication platform 722 can be part of a cloud environment.
- the environment 700 can include repositories or databases 718 , 720 , which can be in communication with the servers 702 , 904 , as well as the one or more illumination sources 706 , one or more rolling shutter cameras 708 , one or more adaptive trigger modules 710 , the feedback module 712 , the user interface 714 , and the central computing system 716 , via the communications platform 722 .
- the servers 702 , 704 , one or more illumination sources 706 , one or more rolling shutter cameras 708 , one or more adaptive trigger modules 710 , the feedback module 712 , the user interface 714 , and the central computing system 716 can be implemented as computing devices (e.g., computing device 600 ).
- the databases 718 , 720 can be incorporated into one or more of the servers 702 , 704 .
- the database 718 can store data relating to captured images, regions of interest 120 , flash delay 124 , combinations thereof, or the like, can be distributed over multiple databases 718 , 720 .
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Studio Devices (AREA)
- Image Input (AREA)
- Exposure Control For Cameras (AREA)
Abstract
Exemplary embodiments are directed to biometric analysis systems including one or more illumination sources configured to provide dim illumination to a scene including an object and configured to provide flash illumination to the object in the scene. The biometric analysis systems include a rolling shutter camera configured to capture one or more images. The biometric analysis systems include an adaptive trigger module configured to analyze the scene to detect the object in the scene during dim illumination of the scene, determine a position in a frame of the rolling shutter camera that coincides with the detected object in the scene, and arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
Description
- The present application claims the benefit of priority to U.S. Provisional Patent Application No. 62/316,347, filed Mar. 31, 2016, which is hereby incorporated by reference in its entirety.
- The present disclosure relates to systems and methods of biometric analysis and, in particular, to biometric analysis systems including an adaptive trigger configured to align an object (e.g., a subject, a barcode, or the like) within a camera field-of-view to improve application of flash illumination to the desired area of the object for capture and analysis.
- Security is a concern in a variety of transactions involving private information. As an example, iris recognition is a well-accepted and accurate means of biometric identification used in government and commercial systems around the world that enables secure transactions and an added layer of security beyond keys and/or passwords. Due to the increased security provided by iris recognition systems, an increase in use of such systems has occurred around the world.
- Traditional cameras used in biometric identification are generally expensive. Video cameras that use complementary metal-oxide-semiconductor (CMOS) or semiconductor charge-coupled device (CCD) image sensors typically use electronic shutters to determine the time period over which the sensor measures light. A trigger signal opens the shutter for a predetermined time during which each pixel within the sensor array collects (integrates) incoming light. At the end of the exposure, the signal collected (integrated) in each pixel during the exposure remains fixed and is then systematically read out, converted to a digital signal, and processed to become an image. Pixels are then cleared and readied for the next exposure to light.
- There are different types of electronic shutters used in the industry.
FIG. 1 is a diagram of a traditional global shutter camera schedule of events for signal collection, including the timeline for exposure of each row of an image sensor in the global shutter camera. Global shutters simultaneously expose an entire array of pixels, e.g., N rows by M columns, during signal collection. During a single exposure event (during texp), light is simultaneously collected in each pixel. When the global shutter closes, the light signal within each pixel represents the image during the period of the single exposure. All pixels integrate signal over exactly the same period of time. Global shutter cameras avoid flash timing issues incurred when using rolling shutter cameras. However, global shutter cameras are expensive options for biometric analysis, thereby increasing the overall costs associated with biometric analysis systems. - Rolling shutter cameras save cost and size in their sensor design.
FIG. 2 is a diagram of a traditional rolling shutter schedule of events for signal collection, including the rolling shutter timeline. Rolling shutters expose an array of pixels differently from global shutters. A rolling shutter system exposes a first row of pixels for an exposure time (texp) and then commences to read-out the exposed row of pixels for digitization. The read-out process occupies a unique onboard resource for a period referred to as a read-out time during which no other row can be read-out. To minimize the duration of the total exposure including the read-out process, the rolling shutter exposes the second row of pixels during a time that is equal to but delayed from the first row by a read-out time. The second row is thereby exposed to light and ready to be read-out at the moment that the read-out process for the first row is complete. The third row is exposed to light for a time interval equal in length to that of the first two rows but delayed relative to the second row by a read-out time allowing for the required time to read-out the second row. The process “rolls” down the pixel array reading row-by-row in sequence taking a total time equal to the exposure time for a single row plus the read-out time interval, times the number of rows. The time interval during which a row is exposed and therefore the events captured by that row are different for each row for a rolling shutter sensor. This is a key difference from a global shutter sensor, especially when using a short flash. - As shown in
FIGS. 1 and 2 , the light collection time period for each row of a sensor with a global shutter is simultaneous while the time periods of light collection for each row of a sensor equipped with a rolling shutter are not simultaneous. Rather, light collection time periods for each row of a rolling shutter are offset from one another with a delay between rows equal to the row read-out time. The different exposure techniques result in image shearing. For example,FIG. 3 shows an image in which a moving fan blade was captured by an image sensor with a global shutter with no or little distortion as compared to the same moving fan captured by an image sensor with a rolling shutter shown inFIG. 4 . Image shearing is an inevitable consequence of the row-by-row time delays built into a rolling shutter sensor in which each row “sees” the scene over a slightly different and offset time interval. - The row-by-row delay in exposure of a rolling shutter also has an effect on coordinating an exposure with flash illumination. As discussed herein, a flash refers to a short, intense period of illumination of a subject during which the light applied by the flash dominates other sources of light on the scene (e.g., the area surrounding the subject).
FIG. 5 shows the exposure of each row during texp with the shaded region indicating the duration of time of the flash illumination occurring simultaneous during texp. In a global shutter, the exposure of all the pixels in a sensor can be coordinated with the application of the flash illumination to the scene. For example, if the period of the flash pulse is 1 ms, the global shutter can open simultaneously with the start of the pulse and close simultaneously with the end of thepulse 1 ms later. The flash illuminates the pixels during and only during their global exposure. Light forming the image is, by assumption, dominated by the light applied to the scene by the flash. If, for example, sunlight is present in the scene, the irradiance on the object from the flash is significantly brighter than that of the sunlight during the flash pulse when pixels are exposed by the global shutter. - Coordinating a flash pulse with a rolling shutter exposure is more complicated than with a global shutter.
FIG. 6 shows the rolling shutter exposure during texp for each row extending diagonally across the diagram, and the time period for flash illumination illustrated as the vertical shaded region tp. After a flash delay, the flash pulse occurs between the start pulse and end pulse points of the diagram. Because the exposure period for each row is delayed from the previous row by a short read-out time interval, illumination of a full frame requires that a flash pulse provide illumination during a period when all rows are integrating light. Failure to meet this condition creates a situation in which some rows of pixels integrate light from the flash pulse while some do not, and perhaps some rows integrate light from only a portion of the flash pulse. In this case, the image is unevenly illuminated. As shown inFIG. 6 , some rows are finished integrating before the flash starts and other rows do not start integrating until after the flash ends. In addition, other rows integrate a partial flash and some integrate the full flash. Thus, a subset of lines on the rolling shutter sensor receive adequate illumination, but outside of this set of lines, the other parts of the sensor remain largely dark. - One example of a short flash pulse can be considered with respect to
FIG. 6 , which, across the top horizontal line, four horizontal dashed lines, and bottom lines, respectively showsrow numbers row 200 of 1000 rows finishes integrating the signal and begins to read-out, indicated by the top dashed line. The flash pulse can end as row 850 of 1000 begins integrating the signal, indicated by the bottom dashed line.FIG. 6 shows that rows 450 through 599 receive the full illumination of the flash pulse, as bracketed by the middle two dashed lines. However,rows 200 to 449 androws 600 to 849 only receive a portion of the flash illumination while rows outside of these ranges, e.g.,rows 1 to 199 and 850 to 1000, receive no flash illumination. Assuming insignificant ambient light, the resulting image would show an illumination stripe surrounded by dim regions. The transition from bright to dim at the top and bottom of the stripe is due to rows that receive flash illumination over a fraction of the total pulse time.FIG. 7 shown a portion of an image acquired using a rolling shutter camera with a delayed flash pulse in which the recorded irradiance is plotted to show dark regions before and after the flash, ramp-up and ramp-down regions of partial illumination, and a plateau region of complete flash illumination. The plateau region ofFIG. 7 is not flat because the flash itself was not uniform over the field-of-view. As another example, an image captured using a rolling shutter camera would include a horizontal stripe with a vertical height proportional to the duration of the flash illumination. In cases with bright ambient illumination, the un-flashed portion of the image would appear, but might be significantly dimmer if the flash illumination is brighter than the ambient illumination. - When an image of a particular object is desired with a rolling shutter camera, a trigger signal can be initiated by the sensor controller to fire the flash at a preset time relative to the start of image writing. For example, the flash can fire when the first line is written and can remain on for 50 of 1000 lines. The resultant image would be flash illuminated for the top 5% of the image and would be dark elsewhere. The same flash can be delayed until the 500th line of 1000 lines, resulting in an image with a stripe of illuminated content approximately halfway down the frame. With such an arrangement, the photographer would need to align the subject within the camera field-of-view such that the stripe of illumination detected by the sensor corresponds to the position of the desired object.
- Traditionally, one solution to the problem of flash illuminating an image using a rolling shutter has been to use an extended period of illumination, e.g., a flash pulse that is started simultaneously with the beginning of the exposure of the first row of pixels and is not finished until the last row of pixels has been exposed. The extended period of illumination is needed to expose the entire image since image lines are written sequentially rather than all at once (as is the case with a camera including a more expensive and physically larger global shutter sensor). This technique necessitates a longer flash pulse compared to the global shutter case, and would show up in
FIG. 6 as a shaded region covering all of the rows with a duration equal to the frame time. This technique would also illuminate the full frame shown inFIG. 7 . Additional requirements on the flash in terms of power output, heating and reliability are needed based on the longer pulse for this technique. A longer pulse might also challenge requirements for eye-safety. For these reasons, full frame pulses with rolling shutters are considered impractical. - Thus, a need exists for improved biometric analysis systems including a rolling shutter that are capable of illuminating and capturing the desired area of an object for identification without an extended flash. These and other needs are addressed by the systems and methods of biometric analysis of the present disclosure.
- In accordance with embodiments of the present disclosure, an exemplary biometric analysis system is provided that includes one or more illumination sources configured to provide dim illumination to a scene including an object, and further configured to provide flash illumination to the object in the scene. In some embodiments, the dim illumination can be provided by an illumination source external and separate from the biometric analysis system, such as ambient light, sunlight, any other light source, or the like (e.g., one or more of the illumination sources can be ambient light). In some embodiments, the biometric analysis system can include a single illumination source that provides the dim illumination, with the same illumination source providing the flash illumination at the determined time period. In some embodiments, the biometric analysis system can include a first illumination source that provides the dim illumination, and a second (separate) illumination source that provides the flash illumination. In some embodiments, the first illumination source can continue to provide the dim illumination during the flash illumination from the second illumination source. In some embodiments, the first illumination source can be automatically actuated into a non-illuminating configuration during the flash illumination provided by the second illumination source, and automatically actuated into an illuminating configuration after the flash illumination is complete.
- The biometric analysis system includes a rolling shutter camera configured to capture one or more images. The rolling shutter camera generally includes a frame with a field-of-view. The term “image” as used herein can include still frame images, video, combinations thereof, or the like. The biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device. The adaptive trigger module, when executed, can be configured to analyze the scene to detect the object in the scene during dim illumination of the scene. The adaptive trigger module, when executed, can be configured to determine a position in the frame of the rolling shutter camera that coincides with the detected object in the scene. The adaptive trigger module, when executed, can be configured to arrange a delay (a time delay) between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
- In some embodiments, the adaptive trigger module, when executed, can be configured to track movement of the object within a field-of-view of the rolling shutter camera. In such embodiments, the adaptive trigger module can be configured to modify the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources based on detected movement of the object within the field-of-view.
- In some embodiments, the adaptive trigger module can be configured to detect a region of interest of the object and arranges the delay such that the stripe of flash illumination coincides with the detected region of interest of the object. In one embodiment, the region of interest of the object can include one or both eyes of a person.
- In some embodiments, the object can be a person. In such embodiments, the adaptive trigger module includes a face finder configured to detect a face (and/or features of the face) of the person. In some embodiments, the object can be a physical item. In such embodiments, the adaptive trigger module can include an identifier finder configured to detect a unique identifier (e.g., a barcode, a quick response (QR) code, combinations thereof, or the like) associated with the physical item.
- In some embodiments, the one or more illumination sources can be configured to provide the flash illumination as a synchronized pulse of flash illumination. The flash illumination provided by the one or more illumination sources is brighter than the dim illumination provided by the one or more illumination sources. In one embodiment, the one or more illumination sources can be near infrared (NIR) illumination sources. In some embodiments, the adaptive trigger module can be configured to sweep an illuminated stripe down the frame as the rolling shutter camera captures the one or more images, analyze an illuminated section of the one or more images to identify a region of interest in the illuminated section, and stop sweeping of the illuminated stripe when the region of interest is identified.
- In accordance with embodiments of the present disclosure, an exemplary biometric analysis system is provided that includes one or more illumination sources configured to provide dim illumination to a scene including a subject and configured to provide flash illumination to the subject in the scene. The biometric analysis system includes a rolling shutter camera configured to capture one or more images. The biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device. The adaptive trigger module, when executed, can be configured to analyze the scene to detect eyes of the subject in the scene during dim illumination of the scene. The adaptive trigger module, when executed, can be configured to identify the eyes of the subject as a region of interest. The adaptive trigger module, when executed, can be configured to determine a position in a frame of the rolling shutter camera that coincides with the identified region of interest. The adaptive trigger module, when executed, can be configured to arrange a flash pulse delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the identified region of interest. In some embodiments, rather than or in addition to using dim illumination, the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected. Thus, rather than using a single frame (as in dim illumination), the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest. The region of interest can be tracked by the adaptive trigger module after being detected.
- The flash pulse delay can ensure that the illuminated region of interest is substantially maintained in a center of the frame of the rolling shutter camera. The biometric analysis system can include a feedback module configured to be executed by the controller or processing device. The feedback module, when executed, can be configured to analyze a captured image of the region of interest and determine if the region of interest is illuminated by the stripe of the flash illumination. The adaptive trigger module, when executed, can be configured to adjust the flash pulse delay based on the determination of the feedback module to ensure that the region of interest is illuminated by the stripe of the flash illumination.
- In accordance with embodiments of the present disclosure, an exemplary method of biometric analysis is provided. The method includes illuminating a scene and an object in the scene with dim illumination from one or more illumination sources. The method includes analyzing the scene with an adaptive trigger module to detect the object in scene during dim illumination. The method includes determining a position in a frame of a rolling shutter camera that coincides with the detected object in the scene. The method includes arranging a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
- In some embodiments, the method can include tracking movement of the object within a field-of-view of the rolling shutter camera with the adaptive trigger module. In such embodiments, the method can include modifying the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources with the adaptive trigger module based on detected movement of the object within the field-of-view.
- In accordance with embodiments of the present disclosure, an exemplary non-transitory computer-readable medium storing instructions for biometric analysis is provided. Execution of the instructions by a processing device (or controller) causes the processing device to illuminate a scene and an object in the scene with dim illumination from one or more illumination sources. Execution of the instructions by a processing device causes the processing device to analyze the scene with an adaptive trigger module to detect the object in scene during dim illumination. Execution of the instructions by a processing device causes the processing device to determine a position in a frame of a rolling shutter camera that coincides with the detected object in the scene. Execution of the instructions by a processing device causes the processing device to arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
- Other objects and features will become apparent from the following detailed description considered in conjunction with the accompanying drawings. It is to be understood, however, that the drawings are designed as an illustration only and not as a definition of the limits of the invention.
- To assist those of skill in the art in making and using the disclosed systems and methods of biometric analysis, reference is made to the accompanying figures, wherein:
-
FIG. 1 (prior art) is a diagram of a traditional global shutter camera schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out. -
FIG. 2 (prior art) is a diagram of a traditional rolling shutter camera schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with delay between rows being equal to read-out time. -
FIG. 3 (prior art) is an image of a moving fan illustrating minimal distortion as captured by a traditional global shutter camera. -
FIG. 4 (prior art) is an image of a moving fan illustrating shearing distortion effects caused by cascading row delays as captured by a traditional rolling shutter camera. -
FIG. 5 (prior art) is a diagram of a traditional global shutter schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out, with a flash firing during the same time interval as that in which the rows are exposed (texp). -
FIG. 6 (prior art) is a diagram of a traditional rolling shutter schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with a flash firing after the beginning of the frame, illuminating some but not all rows, and some rows partially but others completely, resulting in a ramp-up, plateau and ramp-down periods of flash illumination. -
FIG. 7 (prior art) is a diagram of a vertical strip of an image (re-oriented horizontally to align with the time axis) showing dark, ramp-up, plateau, ramp-down, and another dark region due to timing of flash in a traditional rolling shutter frame. -
FIG. 8 is a schematic block diagram of an exemplary biometric analysis system in accordance with the present disclosure. -
FIG. 9 is a representation of an image captured by a rolling shutter camera of the exemplary biometric analysis system including a stripe of flash illumination across a region of interest. -
FIG. 10 is a diagrammatic view of an exemplary biometric analysis system in accordance with the present disclosure. -
FIG. 11 is a diagrammatic view of a frame action of a rolling shutter camera of an exemplary biometric analysis system in accordance with the present disclosure. -
FIG. 12 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure. -
FIG. 13 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure. -
FIG. 14 is a block diagram of an exemplary computing device for implementing an exemplary biometric analysis system in accordance with the present disclosure. -
FIG. 15 is a block diagram of an exemplary biometric analysis system environment in accordance with the present disclosure. - In accordance with embodiments of the present disclosure, exemplary biometric analysis systems are described herein for a rolling shutter camera that include an adaptive trigger for ensuring that the stripe of light from the flash illumination illuminates a region of interest of an object, such as the eyes of a person, during image writing. The biometric analysis systems generally include one or more illumination sources that initially provide dim illumination to a scene to detect the object and identify the region of interest of the object. Based on the identified region of interest, a delay between the start of image writing by the rolling shutter camera and the trigger of the illumination source is arranged or coordinated to ensure that the stripe of the flash illumination coincides with the region of interest of the object.
- Cameras that include rolling shutter image sensors are typically cheaper than those with global shutter image sensors. The exemplary biometric analysis systems use flash illumination in combination with inexpensive rolling shutter cameras of the type found in low-cost cameras and smartphones. The biometric analysis systems address a rolling shutter-specific problem that occurs when the brightness of the ambient light in the scene may need an exposure time less than the full frame time to avoid saturation, and the exposure time is sufficiently short that the first rows of the image are finished integrating light and reading out before the last lines of the image have begun integration. In cases of short exposure times, a single, short flash pulse will not illuminate an entire frame, leaving portions of the frame relatively dark. Long exposure times can result in motion blur.
- The exemplary biometric analysis systems balance the following parameters to ensure the captured image quality is sufficient for identification, and further ensuring that the region of interest is adequately illuminated by the flash: sensor frame rate, sensor exposure time, flash pulse intensity, flash pulse duration, flash pulse repetition, processed frame rate, and flash pulse offset. Increasing the sensor frame rate over the available ranges increases the speed that the reset and read lines move over the image. For a given sensor exposure time, increasing the frame rate effectively increases the number of exposed lines during that period. For example, with the exposure time remaining the same, the execution time of the internal clock of the camera can be increased to operate the camera at a faster rate, thereby capturing a greater number of lines during the exposure time. The drawbacks of increased sensor frame rates are additional data transfers and additional heating due to higher clock speed. Lower frame rates may be desirable for dark indoor lighting conditions and higher frame rates may be desirable for outdoor bright lighting conditions.
- Increasing the sensor exposure time increases the distance between the reset and read times. This effectively increases the number of exposed lines during that period, allowing a short flash pulse to cover more lines. The drawback of increasing this period is that additional ambient light is accumulated over the entire period. During outdoor operation, it may be beneficial to keep the exposure time as short as possible, but allowing enough lines to be covered by the flash. Keeping the exposure time shorter is also important to avoid introducing motion blur which can degrade recognition performance.
- The flash pulse needs to be bright enough to overpower direct sunlight for effective outdoor operation. The flash pulse is limited by available system power and eye safety limits. In general, eye safety can be roughly proportional to the total amount of energy absorbed by the eye over a period of time (e.g., t−0.75). Reducing the intensity, pulse duration, and pulse repetition all move the system into a safer standing. Longer flash pulses allow more lines to be exposed in a rolling shutter. However, longer pulses can lead to increased eye safety concerns forcing the system to use less intense pulses when longer durations are used.
- Several short flash pulses may be used back-to-back to illuminate additional lines with minimal overlap between exposed lines. This technique can reduce the total eye exposure and power requirements while expanding the total number of lines that can be properly illuminated. Pulse rate, duration, and intensity can all impact the system power requirements. Increasing any of these parameters generally increases the demand for power.
- In some embodiments, since eye safety depends on the total exposure over a period of time, one method to reduce the total exposure can be to selectively process frames. While images can be captured at 60 frames per second (fps), only one in four can be illuminated with the flash, which would reduce the exposure to an equivalent of 15 fps with the advantages of capturing at 60 fps. Such technique allows for improved line read speeds, bringing the rolling shutter camera closer to the global shutter camera timing.
- In some embodiments, the systems can include a continuous feedback loop to a user who is attempting to align their eye with a small box on the screen (e.g., in situations involving a smaller screen, such as a smartphone). The feedback loop allows the system to predetermine which lines should be exposed. A low level of continuous, dim illumination can be provided to ensure the user has a reasonable image to complete this task. Depending on the available processing power and flash offset controls, the continuously illuminated frames can be processed to detect specular reflections, and further used to change the flash offset or delay. Such operation allows the user additional margin when attempting to place their eye in the box while keeping the exposure levels to a minimum.
- The biometric analysis systems adjust the portion of the frame that is illuminated such that the important portion of the image (e.g., a region of interest) is well-lit and the unimportant portions of the images are left dim. The biometric analysis systems can be used in imaging in the near infrared (NIR) spectrum using a short, intense, eye-safe NIR flash with an NIR filter to freeze motion and emphasize the flash relative to ambient light (including bright sun). The biometric analysis systems including the NIR flash with a rolling shutter camera can be used in, e.g., iris biometrics, or the like, in which cost reduction realized by using a rolling shutter instead of a global shutter is a primary driver in component selection.
- The biometric analysis systems include an adaptive step in which dim illumination allows a vision system (e.g., a sensor in a rolling shutter camera including a processor) to detect the position of the desired object in the field-of-view of the camera. In some embodiments, the dim illumination remains on for the entire process of capturing an image, and reveals an under-illuminated frame. Dim illumination, although dimmer than illumination by the flash, is bright enough to perform an automated scene analysis, e.g., face finding, object identification, or the like, that can output a recommendation of a position in the frame that would benefit from flash illumination. For example, if a face appeared between 20% and 40% of the frame height measured from the bottom, a controller or processor can arrange a delay between the start of image writing and the flash trigger such that a stripe of illumination coincides with the desired object. If the object moves in the field-of-view, biometric analysis systems can function dynamically to track movement of the object and change the delay accordingly. In some embodiments, rather than or in addition to using dim illumination, the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected. Thus, rather than using a single frame (as in dim illumination), the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest. The region of interest can be tracked by the adaptive trigger module after being detected. Such arrangements to locate and track the object and indicate to the adaptive trigger module changes in the position of the object permits moving objects to be accurately imaged by the biometric analysis systems.
- The biometric analysis systems disclosed herein can be used for rolling shutter photography (e.g., NIR photography) of objects while outdoors or indoors. The biometric analysis systems can acquire NIR images with controlled illumination in full sun or in dim ambient lighting. An image of a desired object that fills less than a full frame, e.g., a person's face with the entire person in the field-of-view, can be captured by the biometric analysis systems. Due to the low cost, small size and availability of rolling shutter sensors, incorporation of the biometric analysis systems into cost-sensitive or size-sensitive platforms can benefit cost-sensitive or size-sensitive users. As one example, the biometric analysis systems can be used in cargo identification in which infrared dye for markings signify authenticity. As another example, a box marked with writing can be imaged with a small and inexpensive handheld infrared camera that would look for the marking surrounding an NIR watermark.
- The flash trigger can be adjusted to fire in order to illuminate the markings, leaving the rest of the field-of-view above and below the mark of interest in the dark. Handheld devices, including smartphones, barcode readers, and other mass produced electronic devices, can incorporate the biometric analysis systems at a small and low cost. Traditionally, a rolling shutter camera would reduce the cost for such systems, but would require an extended flash to illuminate an entire scene to guarantee capturing the subject of interest. The exemplary biometric analysis systems include a rolling shutter camera with an adaptive image content-based trigger that issues a shorter flash of a near infrared light that allows for use of smaller batteries, provides for a longer battery life, provides for improved eye safety, and allows for longer illumination source lifetimes.
- The biometric analysis systems disclosed herein can provide flash illumination for illuminating an object sufficiently to outshine the sun, and in some embodiments can use strobing flash illumination as is disclosed in, e.g., U.S. Pat. Nos. 7,542,628; 7,627,147; 7,657,127; 9,131,141; U.S. Patent Application Publication No. 2015/0098630; U.S. Patent Application Publication No. 2016/0014121; U.S. Patent Application Publication No. 2016/0012218; and U.S. Patent Application Publication No. 2016/0012292, which are incorporated by reference herein.
- The exemplary biometric analysis systems use a rolling shutter camera and a flash illumination that is triggered adaptively to illuminate the region of interest of the scene or object. The adaptive trigger can be configured to follow or track the subject, and adjusts the flash timing to maintain the stripe of flash illumination on the subject of interest. In one embodiment, a face finder can be used as an image analytic to trigger the flash. In one embodiment, a unique identifier, such as a bar code or QR code, can be detected and illuminated by the stripe of flash illumination. In either case, the biometric analysis system can locate the object in a relatively dimly lit scene, and then initiates the repetitive flash that illuminates the object. Flash timing can be adjusted to maintain the object in the illuminated region of the field-of-view of the rolling shutter camera.
- With respect to the delayed flash, an assumption can be made that the start of integration of the first row of a rolling shutter sensor coincides with time t=0. A further assumption can be made that a flash pulse is started at some time later, e.g., delayed by a time Δt. Rows that are read-out (e.g., integrate light) before At miss any light from the flash. Rows that are in the process of integrating light when the flash starts—some nearly finished, some just starting—receive fractions of the flash and form the ramp-up region of partial illumination. Pixels in rows that start integrating at or after the start of the flash and that finish integrating light by the end of the flash are illuminated by the full flash. By adjusting the delay time At, the rows of full flash illumination can be adjusted from the top of the image frame to the bottom to focus on a region of interest located within the frame.
- In particular, the biometric analysis systems can be used to detect the region of interest within the frame of the rolling shutter camera, and intentionally delay the flash illumination in a coordinated manner to ensure that the region of interest is illuminated by the flash during exposure and read-out of the corresponding rows. Proper adjustment of the delay time can cause the stripe of flash illumination due to the short flash pulse to brighten the region of interest. In some embodiments, the stripe can scan or search the image in successive frames of a photo or video stream. Such technique can be performed by systematically sweeping an illuminated stripe down the field-of-view and stopping when a region of interest is detected. Thus, rather than using a single frame (as in dim illumination), the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest. The region of interest can be tracked by the adaptive trigger module after being detected.
- In some embodiments, a pre-flash (e.g., a single long, low intensity flash that illuminates the full frame and that is both eye-safe and within the operating range of the illumination hardware) can provide a single, dim image of the scene. Automated image analysis of the full scene that can tolerate the low light can detect and output a region of interest that is the width of the image and the height of the flash illumination stripe. With the region of interest defined, an automated delay signal can set the illumination stripe to the optimal position to illuminate the region of interest in succeeding frames of a photo or video stream using the full flash brightness over the narrowed region of interest. If the desired feature of the object begins to move out of the illumination stripe (as determined by video analysis performed at the frame rate), the delay can adapt using negative feedback to maintain the position of the desired feature centered within the well-lit region of the image.
- In some embodiments, the biometric analysis systems can be used to capture NIR images of an object that is bathed in sunlight. While a fraction of the solar spectrum covers the NIR wavelengths, it can be assumed that filtered solar NIR irradiance (e.g., from 800 nm to 900 nm) is insufficient to illuminate a video frame that is adequately short in duration to suppress motion blur. It can be further assumed that using a long flash that covers the entire rolling shutter frame period is impractical, even with a short exposure time, due to hardware limitations or eye safety concerns. A short NIR flash pulse that can only illuminate a portion of the frame with an adaptive delay time to ensure that the region of interest is illuminated can therefore be used.
- The biometric analysis systems initially detect and identify a region of interest of an object, such as the eyes of a subject. Eye finding can be accomplished using a number of different processes, such as full frame NIR illumination with a relatively dim, eye-safe (and hardware-safe) long NIR flash pulse; partial frame illumination due to a short, intense pulse that is automatically swept frame-by-frame over the full field-of-view until the eyes appear in the search; and/or image acquisition by a color (or other spectrum) auxiliary camera that produces a full frame image allowing eye-finding or face-finding, and then translation from the auxiliary camera frame position to a NIR camera position.
- After the eyes are detected and their position located in the NIR camera frame, the biometric analysis systems automatically set the flash pulse delay to illuminate a stripe across the frame including the eyes (e.g., the region of interest). The delay that sets the vertical position of the illumination stripe can be automatically adjusted to maintain the eyes in substantially the center of the frame. While the portions of the frame outside of the illumination stripe remain unusably dark, the important information about the eyes remains well lit.
- The competing solar illumination may dimly illuminate the rest of the frame. This is not an issue as long as the flash illumination stripe remains resolved and does not saturate due to the sum of the in-band solar irradiance and the applied NIR light. If the illumination stripe nears the saturation level, the intensity of the applied flash can be throttled back. It is preferable to decrease the NIR illumination irradiance on the object rather than reducing the exposure or pulse duration, since the later strategies reduce the portions of the image that are illuminated.
- Any spectrum of light can be used as long as the camera uses a rolling shutter with an in-band flash illuminator. For example, visible light, NIR, short-wave infrared (SWIR), long-wave infrared (LWIR), ultraviolet (UV), or multi-spectral cameras with rolling shutters that use flash illumination. Any process of finding the region of interest can be used as long as the process returns a coordinate to the biometric analysis system that indicates the time delay between the start of frame and the start of the flash. For example, use of an eye finder that seeks corneal reflections of the flash could be useful if the region of interest contains one or two eyes of the subject. As a further example, a color face camera could be useful to find the eyes of a subject in iris recognition; but if nose-recognition were used, the region of interest would contain the subject's nose. If the same camera were used to find the region of interest and to acquire the information from the region of interest, a different illumination protocol could be used for each function. For example, a long, relatively dim pulse could illuminate the entire frame, followed by a short, bright pulse to illuminate the region of interest with better fidelity (e.g., improved signal-to-noise ratio). Any process that provides the information needed to set the delay time would also be useful.
- In some embodiments, feedback can be used to maintain the region of interest in the illumination stripe. For example, negative feedback can be used to correct the position of the region of interest within the illumination stripe. If the object of interest begins to move or slip to the top or bottom of the region of interest illuminated by the flash, feedback can be used to correct the delay time. For example, if the stripe illuminating the region of interest needs to move up in the frame to follow the object of interest, the delay would need to be corrected in one direction. If the object of interest moves down in the frame, the sign of the delay correction would need to be in the opposite direction. If the delay reaches the minimum or maximum value or, in other words, the object of interest begins to leave the camera field-of-view, the controller or processing device automatically reorients the camera to maintain the object in the field-of-view. In some embodiments, a feedback signal (e.g., an audio alarm, a visual alarm, combinations thereof, or the like) can provide feedback to an operator regarding the object leaving the field-of-view.
- The width of the flash illumination stripe relative to the frame height can be set by practical considerations. For example, rolling shutter cameras with very short read-out times generally have wider flash illumination stripes, all other things being equal. The longest flash pulse can be dictated by practical issues, including available power to the illumination source, operating conditions in power, and temperature of the illumination source. For example, it is known that flashing an illumination source too often can damage it, and reduces the safety of human eyes in and around the illumination source. A flash illumination stripe that is narrow relative to the frame height generally should be more accurately timed than a flash illumination stripe that is wider and that can tolerate more motion of the object of interest within the flash illumination stripe.
- With reference to
FIG. 8 , a schematic block diagram of an exemplary biometric analysis system 100 (hereinafter “system 100”) is provided. Thesystem 100 is configured to detect and identify a region of interest of an object, such as the eyes of a person, and generates a time delay of the flash illumination such that the flash coincides with the region of interest. A rolling shutter camera can therefore be used for capturing one or more images for biometric analysis with the delayed flash illumination ensuring that the region of interest will be properly illuminated in the captured images. Thesystem 100 is configured to track the region of interest within the frame or field-of-view of the rolling shutter camera, and adjusts the time delay of the flash pulse appropriately based on movement of the region of interest in and out of the frame. - The
system 100 generally includes one ormore illumination sources 102. Theillumination sources 102 are configured to initially providedim illumination 104 to a scene including an object to detect and identify the object (or the region of interest of the object). In some embodiments, thedim illumination 104 can be separate from thesystem 100 and can be provided by, e.g., ambient light, any other illumination source, sunlight, combinations thereof, or the like. Theillumination sources 102 are further configured to provide flash illumination 106 (e.g., a pulse of flash illumination) to illuminate the region of interest during exposure of the rows corresponding to the region of interest, thereby ensuring that the region of interest will be illuminated in the captured image. In some embodiments, asingle illumination source 102 can provide both thedim illumination 104 and theflash illumination 106. In some embodiments,separate illumination sources 102 can provide thedim illumination 104 and theflash illumination 106. In one embodiment, theillumination sources 102 can be NIR illumination sources. It should be understood that thedim illumination 104 is less bright than theflash illumination 106, and provides sufficient ambient light for thesystem 100 to detect objects in a scene as viewed by the rolling shutter camera. - The
system 100 includes a rollingshutter camera 108 including aframe 110 with a field-of-view. The rollingshutter camera 108 is configured to capture one or more images (e.g., still frame images, video images, combinations thereof, or the like). The rollingshutter camera 108 is configured to initially capture an image of the scene and the object in the scene underdim illumination 104 conditions. Thesystem 100 includes anadaptive trigger module 112 configured to be executed by aprocessing device 114 having a processor 116 (e.g., a controller). Although illustrated as separate components from the rollingshutter camera 108, it should be understood that in some embodiments, the rollingshutter camera 108 can include theillumination sources 102, theadaptive trigger module 112, theprocessing device 114, or the like. In some embodiments, theprocessing device 114 can be part of acentral computing system 118. When executed, theadaptive trigger module 112 can analyze the captured image of the scene illuminated by thedim illumination 104 and detects the region ofinterest 120 within the scene. For example, theadaptive trigger module 112 can detect and identify the region ofinterest 120 as one of more eyes of the user, and stores the rows of theframe 110 corresponding to the region ofinterest 120 in a database 122 (e.g., a local or remote database). - Based on the rows of the
frame 110 corresponding to the region ofinterest 120, theadaptive trigger module 112 arranges aflash pulse delay 124 between the start of image writing by the rollingshutter camera 108 and a trigger of theflash illumination 106 such that theflash illumination 108 coincides with the identified region ofinterest 120. For example, if the eyes of the user are the region ofinterest 120 corresponding with rows 400-600 of theframe 110, theadaptive trigger module 112 arranges theflash pulse delay 124 such that theflash illumination 106 is triggered to illuminate the object during rows 400-600. Thus, rather than triggering theflash illumination 106 at the same time as the start of image writing by the rollingshutter camera 108, the image writing can start before triggering of theflash illumination 106 and theflash illumination 106 is only triggered to illuminate the region ofinterest 120.Such flash delay 124 ensures that the region ofinterest 120 is illuminated in theimages 126 captured by the rollingshutter camera 108. The resultingimage 126 generally includes a stripe offlash illumination 106 extending across theimage 126 and over the eyes of the user (see, e.g.,FIG. 9 ). Depending on the length of the exposure time, the height of the stripe offlash illumination 106 can be greater or smaller relative to the top and bottom of theimage 126. - In some embodiments, the
adaptive trigger module 112 can include a face finder algorithm that locates the face and/or features of the face in the scene. The face finder algorithm can therefore locate the face identifies one or more eyes of the face as the region ofinterest 120. Although discussed herein as eyes of a user, it should be understood that the region ofinterest 120 can be a unique identifier associated with a physical item that can be detected by an identifier finder algorithm of theadaptive trigger module 112. For example, the identifier finder algorithm of theadaptive trigger module 112 can be used to detect, e.g., a barcode, a QR code, or the like, on a physical item, identifies such unique identifier as the region ofinterest 120, and adjusts theflash delay 124 to ensure that the unique identifier is illuminated during capture of theimage 126. - The
adaptive trigger module 112 can be configured to track movement of the object within the field-of-view orframe 110 of the rollingshutter camera 108. In some embodiments, theadaptive trigger module 112 can be configured to continuously scan the scene in theframe 110 to detect the position of the region ofinterest 120. If the user moves such that the region ofinterest 120 in theframe 110 is not centered or changes relative to the previously detected rows for the region ofinterest 120, theadaptive trigger module 112 can determine the new rows corresponding to the region ofinterest 120, stores the new rows in thedatabase 122, and adjusts theflash delay 124 to coincide with the new rows for the region ofinterest 120. Thus, an object moving in theframe 110 of the rollingshutter camera 108 can be tracked by theadaptive trigger module 112 to allow for the region ofinterest 120 of a moving object to be properly illuminated by theflash illumination 106. - The
system 100 includes afeedback module 128 configured to be executed by theprocessing device 114 and/or thecentral computing system 118. In some embodiments, the rollingshutter camera 108 can include thefeedback module 128. When executed, thefeedback module 128 receives as input theimages 126 captured by the rollingshutter camera 108 and analyzes the region ofinterest 120 captured in theimage 126 to determine if the region ofinterest 120 is illuminated by the stripe offlash illumination 106. If the region ofinterest 120 is illuminated by the stripe offlash illumination 106, thefeedback module 128 electronically outputs such findings to theadaptive trigger module 112 and no changes are made in theflash delay 124. If the region ofinterest 120 is not illuminated (or only partially illuminated) by the stripe offlash illumination 106, thefeedback module 128 electronically outputs such findings to theadaptive trigger module 112. Based on the feedback from thefeedback module 128, theadaptive trigger module 112 analyzes theframe 110 to locate the region ofinterest 120, determines the new rows associated with the region ofinterest 120, and adjusts theflash pulse delay 124 to trigger theflash illumination 106 at the appropriate rows coinciding with the region ofinterest 120. The feedback loop can be continuously performed to ensure the region ofinterest 120 is always illuminated by the stripe offlash illumination 106. - The
system 100 can include auser interface 130 with a graphical user interface (GUI) 132 for receiving input from a user and for outputting information to the user. For example, theuser interface 130 can be used to initiate the process performed by thesystem 100. As a further example, theuser interface 130 can output an audio and/or visual alarm to the user when the region ofinterest 120 is no longer centered in theframe 110. Thesystem 100 can include acommunication interface 134 configured to provide a means for electronic transmission between components of thesystem 100, e.g., theillumination sources 102, theprocessing device 114, the rollingshutter camera 108, theadaptive trigger module 112, thedatabase 122, theuser interface 130, thecentral computing system 118, andfeedback module 128, or the like. -
FIG. 10 is a diagrammatic view of an exemplary biometric analysis system 200 (hereinafter “system 200”) of the present disclosure. Thesystem 200 includes anillumination source 202 that provides dim illumination to the subject 204 in the scene. In some embodiments, theillumination source 202 can provide 1 ms NIR pulses to the scene. In some embodiments, theillumination source 202 can be an NIR light source not connected to thecamera 208. Theillumination source 202 can be positioned adistance 206 from the subject 204. In some embodiments, theillumination source 202 can also be configured to provide the flash illumination to the subject 204. - The
system 200 includes acamera 208 including a rolling shutter (e.g., a rolling shutter camera, such as a smartphone camera, or the like) configured to capture one or more images of the subject. Thecamera 208 can have a field-of-view represented byarea 216, including the region of interest of the subject 204. In some embodiments, thecamera 208 includes an illumination source 210 (e.g., an integrated NIR light source) configured to provide flash illumination to a region of interest (e.g., the eyes of the subject 204). As an example,area 212 indicates the coverage of the NIR flash illumination provided by theillumination source 210, with the region of interest on the subject 204 being within thearea 212. Thecamera 208 can be spaced by adistance 214 from the subject 204. InFIG. 10 , theillumination source 202 appears separate fromcamera 202, however, one skilled in the art shall appreciate that, in some embodiments, theillumination source 202 can be incorporated into thecamera 208 assembly such that thecamera 208 assembly is capable of providing both the dim illumination and flash illumination to the subject 204. -
FIG. 11 is a diagrammatic view of aframe action 300 of a rolling shutter camera of the exemplary biometric analysis systems disclosed herein. Theframe action 300 includes atotal frame time 302.Ambient light 304 having alow illumination level 306 can be provided for all or part of thetotal frame time 302 by a dim illumination source and/or sunlight. The totalcamera exposure time 308 indicates when the rolling shutter camera begins and ends image writing and read-out. - Based on the determination of the region of interest and the corresponding rows of the frame, a pulse of flash illumination 310 (e.g., NIR from a light-emitting diode (LED) illumination source) is triggered to coincide with the rows of the region of interest. The pulse of
flash illumination 310 has a time orpulse width 312 that is significantly smaller than thecamera exposure time 308, and anillumination level 314 that is brighter or higher than thelevel 306 of theambient light 304. The region of interest is thereby illuminated by theflash illumination 310 while the remaining rows of the captured image are only illuminated by theambient light 304. - In some embodiments, a 1280×960 sensor can be used with the rolling shutter camera. In some embodiments, a 120 fps VGA (640×480) sensor can be used with the rolling shutter camera to allow for full frame exposure in 4 ms, and can be synchronized with an approximately 4 ms LED flash every 8th frame. The VGA output can provide (unscaled) 15 fps feedback to the user facing the monitor of the system. The 4 ms can provide sufficient flash pulse while allowing for cost effective use of an LED and rolling shutter accommodation, and allowing for eye safety.
- In some embodiments, a bright light (e.g., approximately 850 nm LED, two LEDs) can be used for the flash pulse. The systems can therefore have high pulsed current capabilities. The rolling shutter camera can have a high quality camera lens and NIR filtering. In some embodiments, the NIR filter can be tighter than 100 nm around the LED center wavelength. In some embodiments, the system can include optics with a modulation transfer function (MTF) optimized for the spatial frequencies that enter the system. In some embodiments, the system can include preprocessing of iris images for a particular spectrum of spatial frequencies (e.g., low, medium, high, or the like).
-
FIG. 12 is a flowchart illustrating anexemplary process 400 of implementing the biometric analysis systems disclosed herein. To begin, atstep 402, a scene and an object in the scene are illuminated with dim illumination from one or more illumination sources. Atstep 404, the scene is analyzed with an adaptive trigger module to detect the object (or a region of interest associated with the object) in the scene during dim illumination. Atstep 406, a position in a frame of a rolling shutter camera is determined with the adaptive trigger module, the position coinciding with the detected object (or region of interest associated with the object) in the scene. - At
step 408, a delay is arranged by the adaptive trigger module between a start of image writing by the rolling shutter camera and a trigger of the illumination sources such that a stripe of flash illumination provided by the illumination sources coincides with the detected object (or region of interest associated with the object) in the scene. Atstep 410, movement of the object is tracked within a field-of-view of the rolling shutter camera with the adaptive trigger module. Atstep 412, the delay between the start of image writing by the rolling shutter camera and the trigger of the illumination sources is modified with the adaptive trigger module based on detected movement of the object (or region of interest associated with the object) out of the field-of-view. -
FIG. 13 is a flowchart illustrating anexemplary process 500 of implementing the biometric analysis systems disclosed herein. To begin, atstep 502, the adaptive trigger module can set the flash delay at 0 ms, the flash duration equal to the time of the full frame, and the flash brightness to low. Atstep 504, a full frame image can be acquired with the rolling shutter camera under dim illumination. Atstep 506, the adaptive trigger module analyzes the image captured atstep 504 and locates the region of interest in the image. - At
step 508, the adaptive trigger module sets the flash delay to a value greater than 0 ms, the flash duration to short (e.g., a pulse), and the flash brightness to high (e.g., higher than the dim illumination) to illuminate the region of interest. Atstep 510, the adaptive trigger module measures the position of the object of interest within the region of interest (e.g., eyes of the user located on the user's face). Atstep 512, if the object of interest is within the region of interest, theprocess 500 continues a loop to constantly (or at predetermined frequencies of time) ensure that the object is maintained within the region of interest. Atstep 514, if the object of interest is near an edge of the frame or region of interest, theprocess 500 performsstep 516 to correct the flash delay such that the object is centered within the region of interest. The region of interest can thereby be maintained in the flash illumination stripe despite relative motion of the object and/or the camera. In some embodiments, the process can include a loop that analyzes the captures images to control the intensity of the flash illumination, thereby avoiding saturation of the object. -
FIG. 14 is a block diagram of acomputing device 600 in accordance with exemplary embodiments of the present disclosure. Thecomputing device 600 includes one or more non-transitory computer-readable media for storing one or more computer-executable instructions or software for implementing exemplary embodiments. The non-transitory computer-readable media may include, but are not limited to, one or more types of hardware memory, non-transitory tangible media (for example, one or more magnetic storage disks, one or more optical disks, one or more flash drives), and the like. For example,memory 606 included in thecomputing device 600 may store computer-readable and computer-executable instructions or software for implementing exemplary embodiments of the present disclosure (e.g., instructions for operating the illumination sources, instructions for operating the processing device, instructions for operating the rolling shutter camera, instructions for operating the adaptive trigger module, instructions for operating the communication interface, instructions for operating the user interface, instructions for operating the central computing system, instructions for operating the feedback module, combinations thereof, or the like). Thecomputing device 600 also includes configurable and/orprogrammable processor 602 and associatedcore 604, and optionally, one or more additional configurable and/or programmable processor(s) 602′ and associated core(s) 604′ (for example, in the case of computer systems having multiple processors/cores), for executing computer-readable and computer-executable instructions or software stored in thememory 606 and other programs for controlling system hardware.Processor 602 and processor(s) 602′ may each be a single core processor or multiple core (604 and 604′) processor. - Virtualization may be employed in the
computing device 600 so that infrastructure and resources in thecomputing device 600 may be shared dynamically. Avirtual machine 614 may be provided to handle a process running on multiple processors so that the process appears to be using only one computing resource rather than multiple computing resources. Multiple virtual machines may also be used with one processor.Memory 606 may include a computer system memory or random access memory, such as DRAM, SRAM, EDO RAM, and the like.Memory 606 may include other types of memory as well, or combinations thereof. - A user may interact with the
computing device 600 through a visual display device 618 (e.g., a personal computer, a mobile smart device, or the like), such as a computer monitor, which may display one or more user interfaces 620 (e.g., a graphical user interface) that may be provided in accordance with exemplary embodiments. Thecomputing device 600 may include other I/O devices for receiving input from a user, for example, a camera, a sensor, a keyboard or any suitablemulti-point touch interface 608, a pointing device 610 (e.g., a mouse). Thekeyboard 608 and thepointing device 610 may be coupled to thevisual display device 618. Thecomputing device 600 may include other suitable conventional I/O peripherals. - The
computing device 600 may also include one ormore storage devices 624, such as a hard-drive, CD-ROM, eMMC (MultiMediaCard), SD (secure digital) card, flash drive, non-volatile storage media, or other computer readable media, for storing data and computer-readable instructions and/or software that implement exemplary embodiments of the biometric analysis systems described herein.Exemplary storage device 624 may also store one ormore databases 626 for storing any suitable information required to implement exemplary embodiments. For example,exemplary storage device 624 can store one ormore databases 626 for storing information, such as data relating to capturedimages 126 underdim illumination 104 andflash illumination 106, regions ofinterest 120,flash delay 124, combinations thereof, or the like, and computer-readable instructions and/or software that implement exemplary embodiments described herein. Thedatabases 626 may be updated by manually or automatically at any suitable time to add, delete, and/or update one or more items in the databases. - The
computing device 600 can include anetwork interface 612 configured to interface via one ormore network devices 622 with one or more networks, for example, Local Area Network (LAN), Wide Area Network (WAN) or the Internet through a variety of connections including, but not limited to, standard telephone lines, LAN or WAN links (for example, 802.11, T1, T3, 56 kb, X.25), broadband connections (for example, ISDN, Frame Relay, ATM), wireless connections, controller area network (CAN), or some combination of any or all of the above. Thenetwork interface 612 may include a built-in network adapter, network interface card, PCMCIA network card, PCI/PCIe network adapter, SD adapter, Bluetooth adapter, card bus network adapter, wireless network adapter, USB network adapter, modem or any other device suitable for interfacing thecomputing device 600 to any type of network capable of communication and performing the operations described herein. Moreover, thecomputing device 600 may be any computer system, such as a workstation, desktop computer, server, laptop, handheld computer, tablet computer (e.g., the tablet computer), mobile computing or communication device (e.g., the smart phone communication device), an embedded computing platform, or other form of computing or telecommunications device that is capable of communication and that has sufficient processor power and memory capacity to perform the operations described herein. - The
computing device 600 may run anyoperating system 616, such as any of the versions of the Microsoft® Windows® operating systems, the different releases of the Unix and Linux operating systems, any version of the MacOS® for Macintosh computers, any embedded operating system, any real-time operating system, any open source operating system, any proprietary operating system, or any other operating system capable of running on the computing device and performing the operations described herein. In exemplary embodiments, theoperating system 616 may be run in native mode or emulated mode. In an exemplary embodiment, theoperating system 616 may be run on one or more cloud machine instances. -
FIG. 15 is a block diagram of an exemplary biometricanalysis system environment 700 in accordance with exemplary embodiments of the present disclosure. Theenvironment 700 can includeservers more illumination sources 706, one or more rollingshutter cameras 708, one or moreadaptive trigger modules 710, afeedback module 712, auser interface 714, and acentral computing system 716 via a communication platform 922, which can be any network over which information can be transmitted between devices communicatively coupled to the network. For example, thecommunication platform 722 can be the Internet, Intranet, virtual private network (VPN), wide area network (WAN), local area network (LAN), and the like. In some embodiments, thecommunication platform 722 can be part of a cloud environment. - The
environment 700 can include repositories ordatabases servers 702, 904, as well as the one ormore illumination sources 706, one or more rollingshutter cameras 708, one or moreadaptive trigger modules 710, thefeedback module 712, theuser interface 714, and thecentral computing system 716, via thecommunications platform 722. - In exemplary embodiments, the
servers more illumination sources 706, one or more rollingshutter cameras 708, one or moreadaptive trigger modules 710, thefeedback module 712, theuser interface 714, and thecentral computing system 716 can be implemented as computing devices (e.g., computing device 600). Those skilled in the art will recognize that thedatabases servers database 718 can store data relating to captured images, regions ofinterest 120,flash delay 124, combinations thereof, or the like, can be distributed overmultiple databases - While exemplary embodiments have been described herein, it is expressly noted that these embodiments should not be construed as limiting, but rather that additions and modifications to what is expressly described herein also are included within the scope of the invention. Moreover, it is to be understood that the features of the various embodiments described herein are not mutually exclusive and can exist in various combinations and permutations, even if such combinations or permutations are not made express herein, without departing from the spirit and scope of the invention.
Claims (21)
1. A biometric analysis system, comprising:
one or more illumination sources configured to provide dim illumination to a scene including an object and configured to provide flash illumination to the object in the scene;
a rolling shutter camera configured to capture one or more images; and
an adaptive trigger module configured to (i) analyze the scene to detect the object in the scene during dim illumination of the scene, (ii) determine a position in a frame of the rolling shutter camera that coincides with the detected object in the scene, and (iii) arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
2. The biometric analysis system of claim 1 , wherein the adaptive trigger module is configured to track movement of the object within a field-of-view of the rolling shutter camera.
3. The biometric analysis system of claim 2 , wherein the adaptive trigger module is configured to modify the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources based on detected movement of the object within the field-of-view.
4. The biometric analysis system of claim 1 , wherein the adaptive trigger module is configured to detect a region of interest of the object and arranges the delay such that the stripe of flash illumination coincides with the detected region of interest of the object.
5. The biometric analysis system of claim 4 , wherein the region of interest of the object includes eyes of a person.
6. The biometric analysis system of claim 1 , wherein the object is a person, and wherein the adaptive trigger module comprises a face finder configured to detect a face of the person.
7. The biometric analysis system of claim 1 , wherein the object is a physical item, and the adaptive trigger module comprises an identifier finder configured to detect a unique identifier associated with the physical item.
8. The biometric analysis system of claim 7 , wherein the unique identifier is a barcode or a quick response (QR) code.
9. The biometric analysis system of claim 1 , wherein the one or more illumination sources are configured to provide the flash illumination as a synchronized pulse of flash illumination.
10. The biometric analysis system of claim 1 , wherein the flash illumination provided by the one or more illumination sources is brighter than the dim illumination provided by the one or more illumination sources.
11. The biometric analysis system of claim 1 , wherein the one or more illumination sources comprise a first illumination source configured to provide the dim illumination and a second illumination source configured to provide the flash illumination.
12. The biometric analysis system of claim 1 , wherein the one or more illumination sources are near infrared illumination sources.
13. The biometric analysis system of claim 1 , wherein the one or more illumination sources are ambient light.
14. The biometric analysis system of claim 1 , wherein the adaptive trigger module is configured to sweep an illuminated stripe down the frame as the rolling shutter camera captures the one or more images, analyzes an illuminated section of the one or more images to identify a region of interest in the illuminated section, and stops sweeping of the illuminated stripe when the region of interest is identified.
15. The biometric analysis system of claim 1 provided as a smartphone having said one or more illumination sources, said rolling shutter camera, and said adaptive trigger module.
16. A biometric analysis system, comprising:
one or more illumination sources configured to provide dim illumination to a scene including a subject and configured to provide flash illumination to the subject in the scene;
a rolling shutter camera configured to capture one or more images; and
an adaptive trigger module configured to (i) analyze the scene to detect eyes of the subject in the scene during dim illumination of the scene, (ii) identify the eyes of the subject as a region of interest, (iii) determine a position in a frame of the rolling shutter camera that coincides with the identified region of interest, and (iv) arrange a flash pulse delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the identified region of interest.
17. The biometric analysis system of claim 16 , wherein the flash pulse delay ensures that the illuminated region of interest is maintained in a center of the frame of the rolling shutter camera.
18. The biometric analysis system of claim 16 , comprising a feedback module configured to analyze a captured image of the region of interest and determine if the region of interest is illuminated by the stripe of the flash illumination.
19. The biometric analysis system of claim 18 , wherein the adaptive trigger module is configured to adjust the flash pulse delay based on the determination of the feedback module to ensure that the region of interest is illuminated by the stripe of the flash illumination.
20. The biometric analysis system of claim 16 provided as a smartphone having said one or more illumination sources, said rolling shutter camera, and said adaptive trigger module.
21-24. (canceled)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/531,488 US20190356827A1 (en) | 2016-03-31 | 2019-08-05 | Systems and methods of biometric analysis with adaptive trigger |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201662316347P | 2016-03-31 | 2016-03-31 | |
US15/471,131 US10373008B2 (en) | 2016-03-31 | 2017-03-28 | Systems and methods of biometric analysis with adaptive trigger |
US16/531,488 US20190356827A1 (en) | 2016-03-31 | 2019-08-05 | Systems and methods of biometric analysis with adaptive trigger |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/471,131 Continuation US10373008B2 (en) | 2016-03-31 | 2017-03-28 | Systems and methods of biometric analysis with adaptive trigger |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190356827A1 true US20190356827A1 (en) | 2019-11-21 |
Family
ID=59961086
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/471,131 Active 2037-07-20 US10373008B2 (en) | 2016-03-31 | 2017-03-28 | Systems and methods of biometric analysis with adaptive trigger |
US16/531,488 Abandoned US20190356827A1 (en) | 2016-03-31 | 2019-08-05 | Systems and methods of biometric analysis with adaptive trigger |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/471,131 Active 2037-07-20 US10373008B2 (en) | 2016-03-31 | 2017-03-28 | Systems and methods of biometric analysis with adaptive trigger |
Country Status (2)
Country | Link |
---|---|
US (2) | US10373008B2 (en) |
WO (1) | WO2017172695A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112040139A (en) * | 2020-09-09 | 2020-12-04 | 绍兴埃瓦科技有限公司 | Light supplementing method for camera imaging |
US11816198B2 (en) | 2021-04-06 | 2023-11-14 | Bank Of America Corporation | Systems and methods for geolocation security using biometric analysis |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10425814B2 (en) | 2014-09-24 | 2019-09-24 | Princeton Identity, Inc. | Control of wireless communication device capability in a mobile device with a biometric key |
EP3227816A4 (en) | 2014-12-03 | 2018-07-04 | Princeton Identity, Inc. | System and method for mobile device biometric add-on |
JP2019506694A (en) | 2016-01-12 | 2019-03-07 | プリンストン・アイデンティティー・インコーポレーテッド | Biometric analysis system and method |
WO2017173228A1 (en) | 2016-03-31 | 2017-10-05 | Princeton Identity, Inc. | Biometric enrollment systems and methods |
US10373008B2 (en) | 2016-03-31 | 2019-08-06 | Princeton Identity, Inc. | Systems and methods of biometric analysis with adaptive trigger |
JP2017211891A (en) * | 2016-05-27 | 2017-11-30 | ソニー株式会社 | Information processing device, information processing method, and recording medium |
US10607096B2 (en) | 2017-04-04 | 2020-03-31 | Princeton Identity, Inc. | Z-dimension user feedback biometric system |
JP2020529073A (en) | 2017-07-26 | 2020-10-01 | プリンストン・アイデンティティー・インコーポレーテッド | Biosecurity systems and methods |
CN109474788B (en) * | 2017-09-08 | 2021-11-30 | 超威半导体公司 | Illumination control techniques for color and IR camera sensors |
US10691969B2 (en) * | 2017-11-06 | 2020-06-23 | EagleSens Systems Corporation | Asynchronous object ROI detection in video mode |
WO2019111785A1 (en) * | 2017-12-08 | 2019-06-13 | キヤノン株式会社 | Identification device and sorting system |
CN108197546B (en) * | 2017-12-26 | 2020-12-11 | 深圳市友信长丰科技有限公司 | Illumination processing method and device in face recognition, computer equipment and storage medium |
NL2021044B1 (en) * | 2018-06-01 | 2019-12-10 | Eldolab Holding Bv | LED driver and method of operating a camera |
US10728447B2 (en) * | 2018-10-26 | 2020-07-28 | Alibaba Group Holding Limited | Capturing images using sub-frame illumination |
CN114009002B (en) * | 2018-11-28 | 2022-11-22 | 亮锐有限责任公司 | Method for obtaining digital image |
EP3820134A1 (en) * | 2019-11-06 | 2021-05-12 | Koninklijke Philips N.V. | A system for performing ambient light image correction |
US11941897B2 (en) * | 2021-12-09 | 2024-03-26 | Fotonation Limited | Vehicle occupant monitoring system including an image acquisition device with a rolling shutter image sensor |
Family Cites Families (215)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3852592A (en) | 1973-06-07 | 1974-12-03 | Stanley Works | Automatic door operator |
US3993888A (en) | 1974-10-29 | 1976-11-23 | Calspan Corporation | Scanning line filter |
US4109237A (en) | 1977-01-17 | 1978-08-22 | Hill Robert B | Apparatus and method for identifying individuals through their retinal vasculature patterns |
US4641349A (en) | 1985-02-20 | 1987-02-03 | Leonard Flom | Iris recognition system |
US5291560A (en) | 1991-07-15 | 1994-03-01 | Iri Scan Incorporated | Biometric personal identification system based on iris analysis |
US6850252B1 (en) | 1999-10-05 | 2005-02-01 | Steven M. Hoffberg | Intelligent electronic appliance system and method |
US5337104A (en) | 1992-11-09 | 1994-08-09 | Smith Neil D | Apparatus and method for redirecting camera flash light so as to avoid red-eye |
JP2882266B2 (en) | 1993-12-28 | 1999-04-12 | 株式会社日立製作所 | Signal transmission device and circuit block |
US5835616A (en) | 1994-02-18 | 1998-11-10 | University Of Central Florida | Face detection using templates |
US5481622A (en) | 1994-03-01 | 1996-01-02 | Rensselaer Polytechnic Institute | Eye tracking apparatus and method employing grayscale threshold values |
US6714665B1 (en) | 1994-09-02 | 2004-03-30 | Sarnoff Corporation | Fully automated iris recognition system utilizing wide and narrow fields of view |
US5572596A (en) | 1994-09-02 | 1996-11-05 | David Sarnoff Research Center, Inc. | Automated, non-invasive iris recognition system and method |
US5833188A (en) | 1994-12-20 | 1998-11-10 | Twofish Unlimited | Accessory mounting apparatus |
WO1997014873A1 (en) | 1995-10-16 | 1997-04-24 | Valentin Semenovich Gorelykh | Method of converting energy and a device for applying the said method |
EP0865637A4 (en) | 1995-12-04 | 1999-08-18 | Sarnoff David Res Center | Wide field of view/narrow field of view recognition system and method |
US6226016B1 (en) | 1996-02-05 | 2001-05-01 | Seiko Epson Corporation | Display apparatus and method capable of rotating an image by 180 degrees |
US5987459A (en) | 1996-03-15 | 1999-11-16 | Regents Of The University Of Minnesota | Image and document management system for content-based retrieval |
US6320610B1 (en) | 1998-12-31 | 2001-11-20 | Sensar, Inc. | Compact imaging device incorporating rotatably mounted cameras |
JP3436293B2 (en) | 1996-07-25 | 2003-08-11 | 沖電気工業株式会社 | Animal individual identification device and individual identification system |
US5933515A (en) | 1996-07-25 | 1999-08-03 | California Institute Of Technology | User identification through sequential input of fingerprints |
GB2315858A (en) | 1996-08-01 | 1998-02-11 | Sharp Kk | System for eye detection and gaze direction determination |
AU727389B2 (en) | 1996-08-25 | 2000-12-14 | Sarnoff Corporation, The | Apparatus for the iris acquiring images |
US6123733A (en) | 1996-11-27 | 2000-09-26 | Voxel, Inc. | Method and apparatus for rapidly evaluating digital data processing parameters |
US6215891B1 (en) | 1997-03-26 | 2001-04-10 | Oki Electric Industry Co., Ltd. | Eye image recognition method eye image selection method and system therefor |
US6144754A (en) | 1997-03-28 | 2000-11-07 | Oki Electric Industry Co., Ltd. | Method and apparatus for identifying individuals |
US6204858B1 (en) | 1997-05-30 | 2001-03-20 | Adobe Systems Incorporated | System and method for adjusting color data of pixels in a digital image |
US6119096A (en) | 1997-07-31 | 2000-09-12 | Eyeticket Corporation | System and method for aircraft passenger check-in and boarding using iris recognition |
US6252976B1 (en) | 1997-08-29 | 2001-06-26 | Eastman Kodak Company | Computer program product for redeye detection |
EP0910986A1 (en) | 1997-10-24 | 1999-04-28 | BRITISH TELECOMMUNICATIONS public limited company | Imaging apparatus |
US6055322A (en) | 1997-12-01 | 2000-04-25 | Sensor, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
US5953440A (en) | 1997-12-02 | 1999-09-14 | Sensar, Inc. | Method of measuring the focus of close-up images of eyes |
EP0934986A1 (en) | 1997-12-17 | 1999-08-11 | Roche Diagnostics GmbH | Dye-polysaccharide or -cyclosaccharide conjugates and their use as diagnostic agent |
US6421462B1 (en) | 1998-02-06 | 2002-07-16 | Compaq Computer Corporation | Technique for differencing an image |
US6301370B1 (en) | 1998-04-13 | 2001-10-09 | Eyematic Interfaces, Inc. | Face recognition from video images |
US5966197A (en) | 1998-04-21 | 1999-10-12 | Visx, Incorporated | Linear array eye tracker |
JP3315648B2 (en) | 1998-07-17 | 2002-08-19 | 沖電気工業株式会社 | Iris code generation device and iris recognition system |
JP3610234B2 (en) | 1998-07-17 | 2005-01-12 | 株式会社メディア・テクノロジー | Iris information acquisition device and iris identification device |
US6424727B1 (en) | 1998-11-25 | 2002-07-23 | Iridian Technologies, Inc. | System and method of animal identification and animal transaction authorization using iris patterns |
JP2000189403A (en) | 1998-12-25 | 2000-07-11 | Oki Electric Ind Co Ltd | Iris region extraction and individual identifying device |
US6247813B1 (en) | 1999-04-09 | 2001-06-19 | Iritech, Inc. | Iris identification system and method of identifying a person through iris recognition |
JP3636927B2 (en) | 1999-05-18 | 2005-04-06 | 三菱電機株式会社 | Face image processing device |
US6433326B1 (en) | 1999-07-14 | 2002-08-13 | Sarnoff Corporation | CMOS/CCD line transfer imager with low dark current |
CN1361894A (en) | 1999-07-22 | 2002-07-31 | 瑞士电信流动电话公司 | Transaction method and suitable devices therefor |
US6525303B1 (en) | 1999-09-23 | 2003-02-25 | Mitutoyo Corporation | Control systems and methods for synchronizing diffuse illumination with a camera |
US6836554B1 (en) | 2000-06-16 | 2004-12-28 | International Business Machines Corporation | System and method for distorting a biometric for transactions with enhanced security and privacy |
JP3466173B2 (en) | 2000-07-24 | 2003-11-10 | 株式会社ソニー・コンピュータエンタテインメント | Image processing system, device, method and computer program |
JP4469476B2 (en) | 2000-08-09 | 2010-05-26 | パナソニック株式会社 | Eye position detection method and eye position detection apparatus |
US7277561B2 (en) | 2000-10-07 | 2007-10-02 | Qritek Co., Ltd. | Iris identification |
US7155036B2 (en) | 2000-12-04 | 2006-12-26 | Sony Corporation | Face detection under varying rotation |
US7015955B2 (en) | 2000-12-22 | 2006-03-21 | Eastman Kodak Company | Camera having verification display with viewer adaptation compensation for reference illuminants and method |
JP3586431B2 (en) | 2001-02-28 | 2004-11-10 | 松下電器産業株式会社 | Personal authentication method and device |
KR100374708B1 (en) | 2001-03-06 | 2003-03-04 | 에버미디어 주식회사 | Non-contact type human iris recognition method by correction of rotated iris image |
US7095901B2 (en) | 2001-03-15 | 2006-08-22 | Lg Electronics, Inc. | Apparatus and method for adjusting focus position in iris recognition system |
JP2002330318A (en) | 2001-04-27 | 2002-11-15 | Matsushita Electric Ind Co Ltd | Mobile terminal |
US6937135B2 (en) | 2001-05-30 | 2005-08-30 | Hewlett-Packard Development Company, L.P. | Face and environment sensing watch |
US6895103B2 (en) | 2001-06-19 | 2005-05-17 | Eastman Kodak Company | Method for automatically locating eyes in an image |
US20030169334A1 (en) | 2001-08-06 | 2003-09-11 | Michael Braithwaite | Iris capture device having expanded capture volume |
US7167987B2 (en) | 2001-08-29 | 2007-01-23 | Hewlett-Packard Development Company, L.P. | Use of biometrics to provide physical and logic access to computer devices |
US7269737B2 (en) | 2001-09-21 | 2007-09-11 | Pay By Touch Checking Resources, Inc. | System and method for biometric authorization for financial transactions |
US8200980B1 (en) | 2001-09-21 | 2012-06-12 | Open Invention Network, Llc | System and method for enrolling in a biometric system |
KR100456619B1 (en) | 2001-12-05 | 2004-11-10 | 한국전자통신연구원 | A system for registering and authenticating human face using support vector machines and method thereof |
DE10163583A1 (en) | 2001-12-21 | 2003-07-03 | Philips Intellectual Property | Method and device for exposing x-rays |
KR100854890B1 (en) | 2001-12-28 | 2008-08-28 | 엘지전자 주식회사 | Iris recording and recognition method using of several led for iris recognition system |
US7715595B2 (en) | 2002-01-16 | 2010-05-11 | Iritech, Inc. | System and method for iris identification using stereoscopic face recognition |
US7118042B2 (en) | 2002-01-18 | 2006-10-10 | Microscan Systems Incorporated | Method and apparatus for rapid image capture in an image system |
US8819788B2 (en) | 2002-10-21 | 2014-08-26 | Clearone Communications Hong Kong, Limited | Method and system for providing security data to security stations |
AU2003298731A1 (en) | 2002-11-26 | 2004-06-18 | Digimarc Id Systems | Systems and methods for managing and detecting fraud in image databases used with identification documents |
TWI224287B (en) | 2003-01-23 | 2004-11-21 | Ind Tech Res Inst | Iris extraction method |
US7280678B2 (en) | 2003-02-28 | 2007-10-09 | Avago Technologies General Ip Pte Ltd | Apparatus and method for detecting pupils |
US7380938B2 (en) | 2003-03-25 | 2008-06-03 | Sarnoff Corporation | Apparatus to detect and measure saccade and pupilary changes |
US7599524B2 (en) | 2003-04-04 | 2009-10-06 | Sarnoff Corporation | Method and apparatus for providing a robust object finder |
KR20030066512A (en) | 2003-07-04 | 2003-08-09 | 김재민 | Iris Recognition System Robust to noises |
US7379567B2 (en) | 2003-07-17 | 2008-05-27 | Matsushita Electric Industrial Co., Ltd. | Iris code generation method, individual authentication method, iris code entry device, individual authentication device, and individual certification program |
DE10335261A1 (en) | 2003-08-01 | 2005-02-17 | Daimlerchrysler Ag | Compressor and / or turbine wheel for a secondary air conveyor |
US6912298B1 (en) | 2003-08-11 | 2005-06-28 | Adobe Systems Incorporation | Object detection using dynamic probability scans |
US7333653B2 (en) | 2003-08-29 | 2008-02-19 | Hewlett-Packard Development Company, L.P. | Detecting and correcting redeye in an image |
KR100682889B1 (en) | 2003-08-29 | 2007-02-15 | 삼성전자주식회사 | Method and Apparatus for image-based photorealistic 3D face modeling |
WO2005024698A2 (en) | 2003-09-04 | 2005-03-17 | Sarnoff Corporation | Method and apparatus for performing iris recognition from an image |
US8098901B2 (en) | 2005-01-26 | 2012-01-17 | Honeywell International Inc. | Standoff iris recognition system |
US8090157B2 (en) | 2005-01-26 | 2012-01-03 | Honeywell International Inc. | Approaches and apparatus for eye detection in a digital image |
US7593550B2 (en) | 2005-01-26 | 2009-09-22 | Honeywell International Inc. | Distance iris recognition |
US7310443B1 (en) | 2003-09-17 | 2007-12-18 | Sonic Solutions, Inc. | Automatic red eye detection and correction in digital images |
US7295686B2 (en) | 2003-09-29 | 2007-11-13 | Primax Electronics Ltd. | Method of processing red eye in digital images |
US20050210267A1 (en) | 2004-03-18 | 2005-09-22 | Jun Sugano | User authentication method and system, information terminal device and service providing server, subject identification method and system, correspondence confirmation method and system, object confirmation method and system, and program products for them |
JP4059224B2 (en) | 2004-04-13 | 2008-03-12 | 株式会社デンソー | Driver appearance recognition system |
JP2005334402A (en) | 2004-05-28 | 2005-12-08 | Sanyo Electric Co Ltd | Method and device for authentication |
US7466308B2 (en) | 2004-06-28 | 2008-12-16 | Microsoft Corporation | Disposing identifying codes on a user's hand to provide input to an interactive display application |
JP4574277B2 (en) | 2004-08-09 | 2010-11-04 | 興和株式会社 | Ophthalmic imaging equipment |
WO2007011375A1 (en) | 2004-09-13 | 2007-01-25 | Cdm Optics, Inc. | Iris image capture devices and associated systems |
US7856558B2 (en) | 2004-10-21 | 2010-12-21 | Honeywell International Inc. | Biometric verification and duress detection system and method |
US20060184243A1 (en) | 2004-10-22 | 2006-08-17 | Omer Yilmaz | System and method for aligning an optic with an axis of an eye |
US7565183B2 (en) | 2004-11-15 | 2009-07-21 | Sony Ericsson Mobile Communications Ab | Mobile device with selectable camera position |
US7428320B2 (en) | 2004-12-07 | 2008-09-23 | Aoptix Technologies, Inc. | Iris imaging using reflection from the eye |
KR101224408B1 (en) | 2005-01-26 | 2013-01-22 | 허니웰 인터내셔널 인코포레이티드 | A distance iris recognition system |
US7568628B2 (en) * | 2005-03-11 | 2009-08-04 | Hand Held Products, Inc. | Bar code reading device with global electronic shutter control |
JP4702598B2 (en) | 2005-03-15 | 2011-06-15 | オムロン株式会社 | Monitoring system, monitoring apparatus and method, recording medium, and program |
US20060222212A1 (en) | 2005-04-05 | 2006-10-05 | Yingzi Du | One-dimensional iris signature generation system and method |
US7542628B2 (en) | 2005-04-11 | 2009-06-02 | Sarnoff Corporation | Method and apparatus for providing strobed image capture |
US7634114B2 (en) | 2006-09-01 | 2009-12-15 | Sarnoff Corporation | Method and apparatus for iris biometric systems for use in an entryway |
US20060274918A1 (en) | 2005-06-03 | 2006-12-07 | Sarnoff Corporation | Method and apparatus for designing iris biometric systems for use in minimally constrained settings |
US20070047770A1 (en) | 2005-06-13 | 2007-03-01 | Swope Guy G | Multiple biometrics enrollment and verification system |
JP2007011667A (en) | 2005-06-30 | 2007-01-18 | Matsushita Electric Ind Co Ltd | Iris authentication device and iris authentication method |
JP4664147B2 (en) | 2005-07-29 | 2011-04-06 | 株式会社山武 | Iris authentication device |
US20100176802A1 (en) | 2005-08-16 | 2010-07-15 | Koninklijke Philips Electronics, N.V. | Portable electronic device having a rotary unit |
WO2007025258A2 (en) | 2005-08-25 | 2007-03-01 | Sarnoff Corporation | Methods and systems for biometric identification |
US7873490B2 (en) | 2005-12-28 | 2011-01-18 | Solmetric Corporation | Solar access measurement device |
US20070160266A1 (en) | 2006-01-11 | 2007-07-12 | Jones Michael J | Method for extracting features of irises in images using difference of sum filters |
US7583823B2 (en) | 2006-01-11 | 2009-09-01 | Mitsubishi Electric Research Laboratories, Inc. | Method for localizing irises in images using gradients and textures |
GB0603411D0 (en) | 2006-02-21 | 2006-03-29 | Xvista Ltd | Method of processing an image of an eye |
JP2007225763A (en) | 2006-02-22 | 2007-09-06 | Sony Corp | Imaging apparatus, imaging method, and program |
US7844084B2 (en) | 2006-02-27 | 2010-11-30 | Donald Martin Monro | Rotation compensated iris comparison |
DE602007007062D1 (en) | 2006-03-03 | 2010-07-22 | Honeywell Int Inc | IRISER IDENTIFICATION SYSTEM WITH IMAGE QUALITY METERING |
US7903168B2 (en) | 2006-04-06 | 2011-03-08 | Eastman Kodak Company | Camera and method with additional evaluation image capture based on scene brightness changes |
WO2007124020A2 (en) | 2006-04-21 | 2007-11-01 | Sick, Inc. | Image quality analysis with test pattern |
US7682026B2 (en) | 2006-08-22 | 2010-03-23 | Southwest Research Institute | Eye location and gaze detection system and method |
WO2008032329A2 (en) | 2006-09-13 | 2008-03-20 | Alon Atsmon | Providing content responsive to multimedia signals |
WO2008091401A2 (en) | 2006-09-15 | 2008-07-31 | Retica Systems, Inc | Multimodal ocular biometric system and methods |
US8121356B2 (en) | 2006-09-15 | 2012-02-21 | Identix Incorporated | Long distance multimodal biometric system and method |
EP2062197A4 (en) * | 2006-09-15 | 2010-10-06 | Retica Systems Inc | Long distance multimodal biometric system and method |
US7574021B2 (en) | 2006-09-18 | 2009-08-11 | Sarnoff Corporation | Iris recognition for a secure facility |
US20080121721A1 (en) | 2006-11-25 | 2008-05-29 | Chun-Hsien Chen | Reflective Reading Structure For Portable Computer |
WO2008073962A2 (en) | 2006-12-12 | 2008-06-19 | Rutgers, The State University Of New Jersey | System and method for detecting and tracking features in images |
US7825958B2 (en) * | 2007-01-25 | 2010-11-02 | Research In Motion Limited | Method and apparatus for controlling a camera module to compensate for the light level of a white LED |
KR100807940B1 (en) | 2007-03-08 | 2008-02-28 | 박경진 | Intraocular lens |
US8023699B2 (en) | 2007-03-09 | 2011-09-20 | Jiris Co., Ltd. | Iris recognition system, a method thereof, and an encryption system using the same |
US9002073B2 (en) | 2007-09-01 | 2015-04-07 | Eyelock, Inc. | Mobile identity platform |
WO2009029757A1 (en) | 2007-09-01 | 2009-03-05 | Global Rainmakers, Inc. | System and method for iris data acquisition for biometric identification |
EP2215579A4 (en) | 2007-11-29 | 2013-01-30 | Wavefront Biometric Technologies Pty Ltd | Biometric authentication using the eye |
CN104866553A (en) | 2007-12-31 | 2015-08-26 | 应用识别公司 | Method, system, and computer program for identification and sharing of digital images with face signatures |
CA2736609C (en) | 2008-02-14 | 2016-09-20 | Iristrac, Llc | System and method for animal identification using iris images |
US8930238B2 (en) | 2008-02-21 | 2015-01-06 | International Business Machines Corporation | Pervasive symbiotic advertising system and methods therefor |
US9131141B2 (en) | 2008-05-12 | 2015-09-08 | Sri International | Image sensor with integrated region of interest calculation for iris capture, autofocus, and gain control |
US8994877B2 (en) * | 2008-07-30 | 2015-03-31 | Semiconductor Components Industries, Llc | Method and system for synchronizing a flash to an imager |
US8213782B2 (en) | 2008-08-07 | 2012-07-03 | Honeywell International Inc. | Predictive autofocusing system |
US8081254B2 (en) | 2008-08-14 | 2011-12-20 | DigitalOptics Corporation Europe Limited | In-camera based method of detecting defect eye with high accuracy |
US9633261B2 (en) | 2008-08-22 | 2017-04-25 | International Business Machines Corporation | Salting system and method for cancelable iris biometric |
US20100082398A1 (en) | 2008-09-29 | 2010-04-01 | Yahoo! Inc. | System for providing contextually relevant data |
US20100278394A1 (en) | 2008-10-29 | 2010-11-04 | Raguin Daniel H | Apparatus for Iris Capture |
US8317325B2 (en) | 2008-10-31 | 2012-11-27 | Cross Match Technologies, Inc. | Apparatus and method for two eye imaging for iris identification |
CN101750697A (en) | 2008-12-10 | 2010-06-23 | 深圳富泰宏精密工业有限公司 | External mirror frame of electronic device |
US8615596B1 (en) | 2009-01-14 | 2013-12-24 | Sprint Communications Company L.P. | Communication method and system for providing content to a communication device according to a user preference |
US8374404B2 (en) | 2009-02-13 | 2013-02-12 | Raytheon Company | Iris recognition using hyper-spectral signatures |
WO2010121227A1 (en) | 2009-04-17 | 2010-10-21 | Ioculi | Image diversion systems & methods for image capture devices |
US8387858B2 (en) | 2009-06-01 | 2013-03-05 | Synderesis Technologies, Inc. | Consumer rewards systems and methods |
US20100328420A1 (en) | 2009-06-29 | 2010-12-30 | Roman Kendyl A | Optical adapters for mobile devices with a camera |
US20110007205A1 (en) | 2009-07-08 | 2011-01-13 | Dechnia, LLC | Rear to forward facing camera adapter |
US8452131B2 (en) | 2009-09-25 | 2013-05-28 | International Business Machines Corporation | Iris recognition system and method |
US20110081946A1 (en) | 2009-10-07 | 2011-04-07 | Byron Singh N John Singh | Apparatus and method for changing an image-capturing position of a mobile phone camera using a mirror device |
US20110142297A1 (en) | 2009-12-16 | 2011-06-16 | Eye Controls, Llc | Camera Angle Compensation in Iris Identification |
US20110187878A1 (en) | 2010-02-02 | 2011-08-04 | Primesense Ltd. | Synchronization of projected illumination with rolling shutter of image sensor |
JP5003774B2 (en) | 2010-02-15 | 2012-08-15 | 東京エレクトロン株式会社 | Developing device, developing method, and storage medium |
US10178290B2 (en) | 2010-02-17 | 2019-01-08 | Sri International | Method and apparatus for automatically acquiring facial, ocular, and iris images from moving subjects at long-range |
KR101046459B1 (en) | 2010-05-13 | 2011-07-04 | 아이리텍 잉크 | An iris recognition apparatus and a method using multiple iris templates |
US8957956B2 (en) | 2010-06-09 | 2015-02-17 | Honeywell International Inc. | Method and system for iris image capture |
US8150255B2 (en) * | 2010-06-25 | 2012-04-03 | Apple Inc. | Flash control for electronic rolling shutter |
US20120086645A1 (en) | 2010-10-11 | 2012-04-12 | Siemens Corporation | Eye typing system using a three-layer user interface |
US8955001B2 (en) | 2011-07-06 | 2015-02-10 | Symphony Advanced Media | Mobile remote media control platform apparatuses and methods |
KR20120067761A (en) | 2010-12-16 | 2012-06-26 | 한국전자통신연구원 | Apparatus for measuring biometric information using user terminal and method thereof |
US8254768B2 (en) | 2010-12-22 | 2012-08-28 | Michael Braithwaite | System and method for illuminating and imaging the iris of a person |
US8195576B1 (en) | 2011-01-31 | 2012-06-05 | Bank Of America Corporation | Mobile transaction device security system |
US8625847B2 (en) | 2011-03-21 | 2014-01-07 | Blackberry Limited | Login method based on direction of gaze |
US8639058B2 (en) | 2011-04-28 | 2014-01-28 | Sri International | Method of generating a normalized digital image of an iris of an eye |
US8682073B2 (en) | 2011-04-28 | 2014-03-25 | Sri International | Method of pupil segmentation |
US8854446B2 (en) | 2011-04-28 | 2014-10-07 | Iristrac, Llc | Method of capturing image data for iris code based identification of vertebrates |
US8755607B2 (en) | 2011-04-28 | 2014-06-17 | Sri International | Method of normalizing a digital image of an iris of an eye |
EP2710514A4 (en) | 2011-05-18 | 2015-04-01 | Nextgenid Inc | Multi-biometric enrollment kiosk including biometric enrollment and verification, face recognition and fingerprint matching systems |
CN103763972A (en) | 2011-06-20 | 2014-04-30 | Rhp多媒体公司 | Case for portable electronic device |
EP2731490B1 (en) | 2011-07-14 | 2015-01-21 | Koninklijke Philips N.V. | System and method for remote measurement of optical focus |
KR102024954B1 (en) | 2011-08-22 | 2019-09-24 | 아이락 엘엘씨 | Systems and methods for capturing artifact free images |
US8473748B2 (en) | 2011-09-27 | 2013-06-25 | George P. Sampas | Mobile device-based authentication |
WO2013049778A1 (en) | 2011-09-30 | 2013-04-04 | Intuitive Medical Technologies, Llc | Optical adapter for ophthalmological imaging apparatus |
US20130089240A1 (en) | 2011-10-07 | 2013-04-11 | Aoptix Technologies, Inc. | Handheld iris imager |
US9241200B2 (en) | 2011-10-11 | 2016-01-19 | Verizon Patent And Licensing Inc. | Targeted advertising |
WO2013056001A1 (en) | 2011-10-12 | 2013-04-18 | Carnegie Mellon University | System and method for the long range acquisition of iris images for stationary and mobile subjects |
TW201324180A (en) | 2011-12-07 | 2013-06-16 | Wistron Corp | Method for integrating a mobile communication device with a portable electronic device and portable computer module thereof |
GB2497553B (en) | 2011-12-13 | 2018-05-16 | Irisguard Inc | Improvements relating to iris cameras |
US20130188943A1 (en) | 2012-01-19 | 2013-07-25 | Shing-Wai David Wu | Camera Adapter |
US9330294B2 (en) | 2012-01-26 | 2016-05-03 | Aware, Inc. | System and method of capturing and producing biometric-matching quality fingerprints and other types of dactylographic images with a mobile device |
US9373023B2 (en) | 2012-02-22 | 2016-06-21 | Sri International | Method and apparatus for robustly collecting facial, ocular, and iris images using a single sensor |
US9100825B2 (en) | 2012-02-28 | 2015-08-04 | Verizon Patent And Licensing Inc. | Method and system for multi-factor biometric authentication based on different device capture modalities |
US9138140B2 (en) | 2012-03-20 | 2015-09-22 | Lrs Identity, Inc. | Compact iris imaging system |
CN102708357A (en) | 2012-04-12 | 2012-10-03 | 北京释码大华科技有限公司 | Single-image sensor based double-eye iris recognition equipment |
US8977560B2 (en) | 2012-08-08 | 2015-03-10 | Ebay Inc. | Cross-browser, cross-machine recoverable user identifiers |
US8437513B1 (en) | 2012-08-10 | 2013-05-07 | EyeVerify LLC | Spoof detection for biometric authentication |
US20140055337A1 (en) | 2012-08-22 | 2014-02-27 | Mobitv, Inc. | Device eye tracking calibration |
CN104781714A (en) | 2012-09-12 | 2015-07-15 | 欧乐柯利普有限责任公司 | Removable optical devices for mobile electronic devices |
US9154677B2 (en) | 2012-09-20 | 2015-10-06 | Apple Inc. | Camera accessory for angled camera viewing |
KR102039725B1 (en) | 2012-11-23 | 2019-11-04 | 엘지디스플레이 주식회사 | Array substrate for liquid crystal display and manufacturing method of the same |
US9195890B2 (en) | 2012-12-10 | 2015-11-24 | Sri International | Iris biometric matching system |
ES2718493T3 (en) | 2012-12-18 | 2019-07-02 | Eyesmatch Ltd | Aspect capture and visualization procedure |
US9019420B2 (en) | 2012-12-19 | 2015-04-28 | Covr Llc | Image redirection system for mobile devices |
CN103048848A (en) | 2012-12-21 | 2013-04-17 | 吴刚 | Corner lens applied to camera of mobile device and assembly of corner lens |
CN103099624A (en) | 2013-01-11 | 2013-05-15 | 北京释码大华科技有限公司 | Iris distance measurement plate, iris recognition all-in-one machine and iris recognition method thereof |
US9304376B2 (en) | 2013-02-20 | 2016-04-05 | Hand Held Products, Inc. | Optical redirection adapter |
US20140327815A1 (en) | 2013-05-06 | 2014-11-06 | Cvs Pharmacy, Inc. | Smartphone scanner adapter |
US9390327B2 (en) | 2013-09-16 | 2016-07-12 | Eyeverify, Llc | Feature extraction and matching for biometric authentication |
EP3074924A4 (en) | 2013-10-08 | 2017-11-22 | Princeton Identity, Inc. | Iris biometric recognition module and access control assembly |
US10042994B2 (en) | 2013-10-08 | 2018-08-07 | Princeton Identity, Inc. | Validation of the right to access an object |
US10025982B2 (en) | 2013-10-08 | 2018-07-17 | Princeton Identity, Inc. | Collecting and targeting marketing data and information based upon iris identification |
US10038691B2 (en) | 2013-10-08 | 2018-07-31 | Princeton Identity, Inc. | Authorization of a financial transaction |
US9300769B2 (en) | 2013-11-01 | 2016-03-29 | Symbol Technologies, Llc | System for and method of adapting a mobile device having a camera to a reader for electro-optically reading targets |
US10121073B2 (en) * | 2014-02-12 | 2018-11-06 | Samsung Electronics Co., Ltd | Agile biometric camera with bandpass filter and variable light source |
WO2016018488A2 (en) | 2014-05-09 | 2016-02-04 | Eyefluence, Inc. | Systems and methods for discerning eye signals and continuous biometric identification |
US9767358B2 (en) | 2014-10-22 | 2017-09-19 | Veridium Ip Limited | Systems and methods for performing iris identification and verification using mobile devices |
US9412169B2 (en) | 2014-11-21 | 2016-08-09 | iProov | Real-time visual feedback for user positioning with respect to a camera and a display |
EP3227816A4 (en) | 2014-12-03 | 2018-07-04 | Princeton Identity, Inc. | System and method for mobile device biometric add-on |
JP6535223B2 (en) | 2015-05-28 | 2019-06-26 | 浜松ホトニクス株式会社 | Blink measurement method, blink measurement apparatus, and blink measurement program |
US20160364609A1 (en) | 2015-06-12 | 2016-12-15 | Delta ID Inc. | Apparatuses and methods for iris based biometric recognition |
TWI558202B (en) | 2015-10-14 | 2016-11-11 | 緯創資通股份有限公司 | Imaging device, imaging system and control method thereof |
US20170124314A1 (en) | 2015-11-03 | 2017-05-04 | Tavita Laumea | Device and Method for Keypad with an Integral Fingerprint Scanner |
KR102402829B1 (en) | 2015-11-10 | 2022-05-30 | 삼성전자 주식회사 | Method for user authentication and electronic device implementing the same |
JP2019506694A (en) | 2016-01-12 | 2019-03-07 | プリンストン・アイデンティティー・インコーポレーテッド | Biometric analysis system and method |
WO2017173228A1 (en) | 2016-03-31 | 2017-10-05 | Princeton Identity, Inc. | Biometric enrollment systems and methods |
US10373008B2 (en) | 2016-03-31 | 2019-08-06 | Princeton Identity, Inc. | Systems and methods of biometric analysis with adaptive trigger |
WO2018111860A1 (en) | 2016-12-12 | 2018-06-21 | Princeton Identity, Inc. | Systems and methods of biometric analysis |
-
2017
- 2017-03-28 US US15/471,131 patent/US10373008B2/en active Active
- 2017-03-28 WO PCT/US2017/024444 patent/WO2017172695A1/en active Application Filing
-
2019
- 2019-08-05 US US16/531,488 patent/US20190356827A1/en not_active Abandoned
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112040139A (en) * | 2020-09-09 | 2020-12-04 | 绍兴埃瓦科技有限公司 | Light supplementing method for camera imaging |
US11816198B2 (en) | 2021-04-06 | 2023-11-14 | Bank Of America Corporation | Systems and methods for geolocation security using biometric analysis |
Also Published As
Publication number | Publication date |
---|---|
WO2017172695A1 (en) | 2017-10-05 |
US10373008B2 (en) | 2019-08-06 |
US20170286792A1 (en) | 2017-10-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10373008B2 (en) | Systems and methods of biometric analysis with adaptive trigger | |
US11132771B2 (en) | Bright spot removal using a neural network | |
US20210014411A1 (en) | Method for image processing, electronic device, and computer readable storage medium | |
US9953428B2 (en) | Digital camera unit with simultaneous structured and unstructured illumination | |
KR102270674B1 (en) | Biometric camera | |
CN108805198B (en) | Image processing method, image processing device, computer-readable storage medium and electronic equipment | |
US20180165537A1 (en) | Systems And Methods Of Biometric Analysis | |
US20050212913A1 (en) | Method and arrangement for recording regions of interest of moving objects | |
WO2021073140A1 (en) | Monocular camera, and image processing system and image processing method | |
CN104137529A (en) | Method and apparatus for enhanced automatic adjustment of focus, exposure and white balance in digital photography | |
CN110536068A (en) | Focusing method and device, electronic equipment, computer readable storage medium | |
CN109862269A (en) | Image-pickup method, device, electronic equipment and computer readable storage medium | |
CN108322651B (en) | Photographing method and device, electronic equipment and computer readable storage medium | |
CN109327626B (en) | Image acquisition method and device, electronic equipment and computer readable storage medium | |
US9489750B2 (en) | Exposure metering based on background pixels | |
US20110081142A1 (en) | Pulsed control of camera flash | |
US9195884B2 (en) | Method, apparatus, and manufacture for smiling face detection | |
CN108600740A (en) | Optical element detection method, device, electronic equipment and storage medium | |
CN110248101A (en) | Focusing method and device, electronic equipment, computer readable storage medium | |
US20130286218A1 (en) | Image recognition device that recognizes specific object area, method of controlling the device, and storage medium, as well as image pickup apparatus, and display device | |
US8442349B2 (en) | Removal of artifacts in flash images | |
CN107040697A (en) | Use the method for imaging and relevant camera system of gaze detection | |
US20200020086A1 (en) | Information processing method and information processing apparatus | |
US20160350607A1 (en) | Biometric authentication device | |
JP2023059952A (en) | Image processing device, imaging device, image processing method, image processing program, and recording medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |