US20190356827A1 - Systems and methods of biometric analysis with adaptive trigger - Google Patents

Systems and methods of biometric analysis with adaptive trigger Download PDF

Info

Publication number
US20190356827A1
US20190356827A1 US16/531,488 US201916531488A US2019356827A1 US 20190356827 A1 US20190356827 A1 US 20190356827A1 US 201916531488 A US201916531488 A US 201916531488A US 2019356827 A1 US2019356827 A1 US 2019356827A1
Authority
US
United States
Prior art keywords
illumination
flash
interest
region
biometric analysis
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/531,488
Inventor
David Alan Ackerman
James R. Bergen
Barry E. Mapen
Steven N. Perna
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Princeton Identity Inc
Original Assignee
Princeton Identity Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Princeton Identity Inc filed Critical Princeton Identity Inc
Priority to US16/531,488 priority Critical patent/US20190356827A1/en
Publication of US20190356827A1 publication Critical patent/US20190356827A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • H04N5/2256
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/56Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/68Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
    • H04N23/689Motion occurring during a rolling shutter mode
    • H04N5/2329

Definitions

  • the present disclosure relates to systems and methods of biometric analysis and, in particular, to biometric analysis systems including an adaptive trigger configured to align an object (e.g., a subject, a barcode, or the like) within a camera field-of-view to improve application of flash illumination to the desired area of the object for capture and analysis.
  • an adaptive trigger configured to align an object (e.g., a subject, a barcode, or the like) within a camera field-of-view to improve application of flash illumination to the desired area of the object for capture and analysis.
  • iris recognition is a well-accepted and accurate means of biometric identification used in government and commercial systems around the world that enables secure transactions and an added layer of security beyond keys and/or passwords. Due to the increased security provided by iris recognition systems, an increase in use of such systems has occurred around the world.
  • Video cameras that use complementary metal-oxide-semiconductor (CMOS) or semiconductor charge-coupled device (CCD) image sensors typically use electronic shutters to determine the time period over which the sensor measures light.
  • a trigger signal opens the shutter for a predetermined time during which each pixel within the sensor array collects (integrates) incoming light.
  • the signal collected (integrated) in each pixel during the exposure remains fixed and is then systematically read out, converted to a digital signal, and processed to become an image. Pixels are then cleared and readied for the next exposure to light.
  • FIG. 1 is a diagram of a traditional global shutter camera schedule of events for signal collection, including the timeline for exposure of each row of an image sensor in the global shutter camera.
  • Global shutters simultaneously expose an entire array of pixels, e.g., N rows by M columns, during signal collection.
  • a single exposure event (during t exp )
  • light is simultaneously collected in each pixel.
  • the global shutter closes, the light signal within each pixel represents the image during the period of the single exposure. All pixels integrate signal over exactly the same period of time.
  • Global shutter cameras avoid flash timing issues incurred when using rolling shutter cameras.
  • global shutter cameras are expensive options for biometric analysis, thereby increasing the overall costs associated with biometric analysis systems.
  • FIG. 2 is a diagram of a traditional rolling shutter schedule of events for signal collection, including the rolling shutter timeline.
  • Rolling shutters expose an array of pixels differently from global shutters.
  • a rolling shutter system exposes a first row of pixels for an exposure time (t exp ) and then commences to read-out the exposed row of pixels for digitization.
  • the read-out process occupies a unique onboard resource for a period referred to as a read-out time during which no other row can be read-out.
  • the rolling shutter exposes the second row of pixels during a time that is equal to but delayed from the first row by a read-out time.
  • the second row is thereby exposed to light and ready to be read-out at the moment that the read-out process for the first row is complete.
  • the third row is exposed to light for a time interval equal in length to that of the first two rows but delayed relative to the second row by a read-out time allowing for the required time to read-out the second row.
  • the process “rolls” down the pixel array reading row-by-row in sequence taking a total time equal to the exposure time for a single row plus the read-out time interval, times the number of rows.
  • the time interval during which a row is exposed and therefore the events captured by that row are different for each row for a rolling shutter sensor. This is a key difference from a global shutter sensor, especially when using a short flash.
  • the light collection time period for each row of a sensor with a global shutter is simultaneous while the time periods of light collection for each row of a sensor equipped with a rolling shutter are not simultaneous. Rather, light collection time periods for each row of a rolling shutter are offset from one another with a delay between rows equal to the row read-out time.
  • the different exposure techniques result in image shearing.
  • FIG. 3 shows an image in which a moving fan blade was captured by an image sensor with a global shutter with no or little distortion as compared to the same moving fan captured by an image sensor with a rolling shutter shown in FIG. 4 .
  • Image shearing is an inevitable consequence of the row-by-row time delays built into a rolling shutter sensor in which each row “sees” the scene over a slightly different and offset time interval.
  • the row-by-row delay in exposure of a rolling shutter also has an effect on coordinating an exposure with flash illumination.
  • a flash refers to a short, intense period of illumination of a subject during which the light applied by the flash dominates other sources of light on the scene (e.g., the area surrounding the subject).
  • FIG. 5 shows the exposure of each row during t exp with the shaded region indicating the duration of time of the flash illumination occurring simultaneous during t exp .
  • the exposure of all the pixels in a sensor can be coordinated with the application of the flash illumination to the scene. For example, if the period of the flash pulse is 1 ms, the global shutter can open simultaneously with the start of the pulse and close simultaneously with the end of the pulse 1 ms later.
  • the flash illuminates the pixels during and only during their global exposure.
  • Light forming the image is, by assumption, dominated by the light applied to the scene by the flash. If, for example, sunlight is present in the scene, the irradiance on the object from the flash is significantly brighter than that of the sunlight during the flash pulse when pixels are exposed by the global shutter.
  • FIG. 6 shows the rolling shutter exposure during t exp for each row extending diagonally across the diagram, and the time period for flash illumination illustrated as the vertical shaded region t p .
  • the flash pulse occurs between the start pulse and end pulse points of the diagram. Because the exposure period for each row is delayed from the previous row by a short read-out time interval, illumination of a full frame requires that a flash pulse provide illumination during a period when all rows are integrating light. Failure to meet this condition creates a situation in which some rows of pixels integrate light from the flash pulse while some do not, and perhaps some rows integrate light from only a portion of the flash pulse. In this case, the image is unevenly illuminated.
  • some rows are finished integrating before the flash starts and other rows do not start integrating until after the flash ends.
  • other rows integrate a partial flash and some integrate the full flash.
  • a subset of lines on the rolling shutter sensor receive adequate illumination, but outside of this set of lines, the other parts of the sensor remain largely dark.
  • FIG. 6 One example of a short flash pulse can be considered with respect to FIG. 6 , which, across the top horizontal line, four horizontal dashed lines, and bottom lines, respectively shows row numbers 0 , 200 , 450 , 600 , 850 , and 1000 .
  • the flash pulse can start as row 200 of 1000 rows finishes integrating the signal and begins to read-out, indicated by the top dashed line.
  • the flash pulse can end as row 850 of 1000 begins integrating the signal, indicated by the bottom dashed line.
  • FIG. 6 shows that rows 450 through 599 receive the full illumination of the flash pulse, as bracketed by the middle two dashed lines.
  • rows 200 to 449 and rows 600 to 849 only receive a portion of the flash illumination while rows outside of these ranges, e.g., rows 1 to 199 and 850 to 1000 , receive no flash illumination. Assuming insignificant ambient light, the resulting image would show an illumination stripe surrounded by dim regions. The transition from bright to dim at the top and bottom of the stripe is due to rows that receive flash illumination over a fraction of the total pulse time.
  • FIG. 7 shown a portion of an image acquired using a rolling shutter camera with a delayed flash pulse in which the recorded irradiance is plotted to show dark regions before and after the flash, ramp-up and ramp-down regions of partial illumination, and a plateau region of complete flash illumination. The plateau region of FIG.
  • an image captured using a rolling shutter camera would include a horizontal stripe with a vertical height proportional to the duration of the flash illumination. In cases with bright ambient illumination, the un-flashed portion of the image would appear, but might be significantly dimmer if the flash illumination is brighter than the ambient illumination.
  • a trigger signal can be initiated by the sensor controller to fire the flash at a preset time relative to the start of image writing.
  • the flash can fire when the first line is written and can remain on for 50 of 1000 lines.
  • the resultant image would be flash illuminated for the top 5% of the image and would be dark elsewhere.
  • the same flash can be delayed until the 500 th line of 1000 lines, resulting in an image with a stripe of illuminated content approximately halfway down the frame.
  • the photographer would need to align the subject within the camera field-of-view such that the stripe of illumination detected by the sensor corresponds to the position of the desired object.
  • an extended period of illumination e.g., a flash pulse that is started simultaneously with the beginning of the exposure of the first row of pixels and is not finished until the last row of pixels has been exposed.
  • the extended period of illumination is needed to expose the entire image since image lines are written sequentially rather than all at once (as is the case with a camera including a more expensive and physically larger global shutter sensor).
  • This technique necessitates a longer flash pulse compared to the global shutter case, and would show up in FIG. 6 as a shaded region covering all of the rows with a duration equal to the frame time.
  • This technique would also illuminate the full frame shown in FIG. 7 . Additional requirements on the flash in terms of power output, heating and reliability are needed based on the longer pulse for this technique.
  • a longer pulse might also challenge requirements for eye-safety. For these reasons, full frame pulses with rolling shutters are considered impractical.
  • an exemplary biometric analysis system includes one or more illumination sources configured to provide dim illumination to a scene including an object, and further configured to provide flash illumination to the object in the scene.
  • the dim illumination can be provided by an illumination source external and separate from the biometric analysis system, such as ambient light, sunlight, any other light source, or the like (e.g., one or more of the illumination sources can be ambient light).
  • the biometric analysis system can include a single illumination source that provides the dim illumination, with the same illumination source providing the flash illumination at the determined time period.
  • the biometric analysis system can include a first illumination source that provides the dim illumination, and a second (separate) illumination source that provides the flash illumination.
  • the first illumination source can continue to provide the dim illumination during the flash illumination from the second illumination source. In some embodiments, the first illumination source can be automatically actuated into a non-illuminating configuration during the flash illumination provided by the second illumination source, and automatically actuated into an illuminating configuration after the flash illumination is complete.
  • the biometric analysis system includes a rolling shutter camera configured to capture one or more images.
  • the rolling shutter camera generally includes a frame with a field-of-view.
  • image as used herein can include still frame images, video, combinations thereof, or the like.
  • the biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device.
  • the adaptive trigger module when executed, can be configured to analyze the scene to detect the object in the scene during dim illumination of the scene.
  • the adaptive trigger module when executed, can be configured to determine a position in the frame of the rolling shutter camera that coincides with the detected object in the scene.
  • the adaptive trigger module when executed, can be configured to arrange a delay (a time delay) between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
  • the adaptive trigger module when executed, can be configured to track movement of the object within a field-of-view of the rolling shutter camera. In such embodiments, the adaptive trigger module can be configured to modify the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources based on detected movement of the object within the field-of-view.
  • the adaptive trigger module can be configured to detect a region of interest of the object and arranges the delay such that the stripe of flash illumination coincides with the detected region of interest of the object.
  • the region of interest of the object can include one or both eyes of a person.
  • the object can be a person.
  • the adaptive trigger module includes a face finder configured to detect a face (and/or features of the face) of the person.
  • the object can be a physical item.
  • the adaptive trigger module can include an identifier finder configured to detect a unique identifier (e.g., a barcode, a quick response (QR) code, combinations thereof, or the like) associated with the physical item.
  • a unique identifier e.g., a barcode, a quick response (QR) code, combinations thereof, or the like
  • the one or more illumination sources can be configured to provide the flash illumination as a synchronized pulse of flash illumination.
  • the flash illumination provided by the one or more illumination sources is brighter than the dim illumination provided by the one or more illumination sources.
  • the one or more illumination sources can be near infrared (NIR) illumination sources.
  • the adaptive trigger module can be configured to sweep an illuminated stripe down the frame as the rolling shutter camera captures the one or more images, analyze an illuminated section of the one or more images to identify a region of interest in the illuminated section, and stop sweeping of the illuminated stripe when the region of interest is identified.
  • an exemplary biometric analysis system includes one or more illumination sources configured to provide dim illumination to a scene including a subject and configured to provide flash illumination to the subject in the scene.
  • the biometric analysis system includes a rolling shutter camera configured to capture one or more images.
  • the biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device.
  • the adaptive trigger module when executed, can be configured to analyze the scene to detect eyes of the subject in the scene during dim illumination of the scene.
  • the adaptive trigger module when executed, can be configured to identify the eyes of the subject as a region of interest.
  • the adaptive trigger module when executed, can be configured to determine a position in a frame of the rolling shutter camera that coincides with the identified region of interest.
  • the adaptive trigger module when executed, can be configured to arrange a flash pulse delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the identified region of interest.
  • the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected.
  • the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest.
  • the region of interest can be tracked by the adaptive trigger module after being detected.
  • the flash pulse delay can ensure that the illuminated region of interest is substantially maintained in a center of the frame of the rolling shutter camera.
  • the biometric analysis system can include a feedback module configured to be executed by the controller or processing device.
  • the feedback module when executed, can be configured to analyze a captured image of the region of interest and determine if the region of interest is illuminated by the stripe of the flash illumination.
  • the adaptive trigger module when executed, can be configured to adjust the flash pulse delay based on the determination of the feedback module to ensure that the region of interest is illuminated by the stripe of the flash illumination.
  • an exemplary method of biometric analysis includes illuminating a scene and an object in the scene with dim illumination from one or more illumination sources.
  • the method includes analyzing the scene with an adaptive trigger module to detect the object in scene during dim illumination.
  • the method includes determining a position in a frame of a rolling shutter camera that coincides with the detected object in the scene.
  • the method includes arranging a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
  • the method can include tracking movement of the object within a field-of-view of the rolling shutter camera with the adaptive trigger module. In such embodiments, the method can include modifying the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources with the adaptive trigger module based on detected movement of the object within the field-of-view.
  • an exemplary non-transitory computer-readable medium storing instructions for biometric analysis.
  • Execution of the instructions by a processing device causes the processing device to illuminate a scene and an object in the scene with dim illumination from one or more illumination sources.
  • Execution of the instructions by a processing device causes the processing device to analyze the scene with an adaptive trigger module to detect the object in scene during dim illumination.
  • Execution of the instructions by a processing device causes the processing device to determine a position in a frame of a rolling shutter camera that coincides with the detected object in the scene.
  • Execution of the instructions by a processing device causes the processing device to arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
  • FIG. 1 (prior art) is a diagram of a traditional global shutter camera schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out.
  • FIG. 2 (prior art) is a diagram of a traditional rolling shutter camera schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with delay between rows being equal to read-out time.
  • FIG. 3 (prior art) is an image of a moving fan illustrating minimal distortion as captured by a traditional global shutter camera.
  • FIG. 4 (prior art) is an image of a moving fan illustrating shearing distortion effects caused by cascading row delays as captured by a traditional rolling shutter camera.
  • FIG. 5 is a diagram of a traditional global shutter schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out, with a flash firing during the same time interval as that in which the rows are exposed (t exp ).
  • FIG. 6 is a diagram of a traditional rolling shutter schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with a flash firing after the beginning of the frame, illuminating some but not all rows, and some rows partially but others completely, resulting in a ramp-up, plateau and ramp-down periods of flash illumination.
  • FIG. 7 is a diagram of a vertical strip of an image (re-oriented horizontally to align with the time axis) showing dark, ramp-up, plateau, ramp-down, and another dark region due to timing of flash in a traditional rolling shutter frame.
  • FIG. 8 is a schematic block diagram of an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 9 is a representation of an image captured by a rolling shutter camera of the exemplary biometric analysis system including a stripe of flash illumination across a region of interest.
  • FIG. 10 is a diagrammatic view of an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 11 is a diagrammatic view of a frame action of a rolling shutter camera of an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 12 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 13 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 14 is a block diagram of an exemplary computing device for implementing an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 15 is a block diagram of an exemplary biometric analysis system environment in accordance with the present disclosure.
  • exemplary biometric analysis systems are described herein for a rolling shutter camera that include an adaptive trigger for ensuring that the stripe of light from the flash illumination illuminates a region of interest of an object, such as the eyes of a person, during image writing.
  • the biometric analysis systems generally include one or more illumination sources that initially provide dim illumination to a scene to detect the object and identify the region of interest of the object. Based on the identified region of interest, a delay between the start of image writing by the rolling shutter camera and the trigger of the illumination source is arranged or coordinated to ensure that the stripe of the flash illumination coincides with the region of interest of the object.
  • the exemplary biometric analysis systems use flash illumination in combination with inexpensive rolling shutter cameras of the type found in low-cost cameras and smartphones.
  • the biometric analysis systems address a rolling shutter-specific problem that occurs when the brightness of the ambient light in the scene may need an exposure time less than the full frame time to avoid saturation, and the exposure time is sufficiently short that the first rows of the image are finished integrating light and reading out before the last lines of the image have begun integration.
  • a single, short flash pulse will not illuminate an entire frame, leaving portions of the frame relatively dark. Long exposure times can result in motion blur.
  • the exemplary biometric analysis systems balance the following parameters to ensure the captured image quality is sufficient for identification, and further ensuring that the region of interest is adequately illuminated by the flash: sensor frame rate, sensor exposure time, flash pulse intensity, flash pulse duration, flash pulse repetition, processed frame rate, and flash pulse offset.
  • sensor frame rate over the available ranges increases the speed that the reset and read lines move over the image.
  • increasing the frame rate effectively increases the number of exposed lines during that period.
  • the execution time of the internal clock of the camera can be increased to operate the camera at a faster rate, thereby capturing a greater number of lines during the exposure time.
  • the drawbacks of increased sensor frame rates are additional data transfers and additional heating due to higher clock speed. Lower frame rates may be desirable for dark indoor lighting conditions and higher frame rates may be desirable for outdoor bright lighting conditions.
  • Increasing the sensor exposure time increases the distance between the reset and read times. This effectively increases the number of exposed lines during that period, allowing a short flash pulse to cover more lines.
  • the drawback of increasing this period is that additional ambient light is accumulated over the entire period.
  • the flash pulse needs to be bright enough to overpower direct sunlight for effective outdoor operation.
  • the flash pulse is limited by available system power and eye safety limits.
  • eye safety can be roughly proportional to the total amount of energy absorbed by the eye over a period of time (e.g., t ⁇ 0.75 ). Reducing the intensity, pulse duration, and pulse repetition all move the system into a safer standing. Longer flash pulses allow more lines to be exposed in a rolling shutter. However, longer pulses can lead to increased eye safety concerns forcing the system to use less intense pulses when longer durations are used.
  • one method to reduce the total exposure can be to selectively process frames. While images can be captured at 60 frames per second (fps), only one in four can be illuminated with the flash, which would reduce the exposure to an equivalent of 15 fps with the advantages of capturing at 60 fps. Such technique allows for improved line read speeds, bringing the rolling shutter camera closer to the global shutter camera timing.
  • the systems can include a continuous feedback loop to a user who is attempting to align their eye with a small box on the screen (e.g., in situations involving a smaller screen, such as a smartphone).
  • the feedback loop allows the system to predetermine which lines should be exposed.
  • a low level of continuous, dim illumination can be provided to ensure the user has a reasonable image to complete this task.
  • the continuously illuminated frames can be processed to detect specular reflections, and further used to change the flash offset or delay. Such operation allows the user additional margin when attempting to place their eye in the box while keeping the exposure levels to a minimum.
  • the biometric analysis systems adjust the portion of the frame that is illuminated such that the important portion of the image (e.g., a region of interest) is well-lit and the unimportant portions of the images are left dim.
  • the biometric analysis systems can be used in imaging in the near infrared (NIR) spectrum using a short, intense, eye-safe NIR flash with an NIR filter to freeze motion and emphasize the flash relative to ambient light (including bright sun).
  • the biometric analysis systems including the NIR flash with a rolling shutter camera can be used in, e.g., iris biometrics, or the like, in which cost reduction realized by using a rolling shutter instead of a global shutter is a primary driver in component selection.
  • the biometric analysis systems include an adaptive step in which dim illumination allows a vision system (e.g., a sensor in a rolling shutter camera including a processor) to detect the position of the desired object in the field-of-view of the camera.
  • a vision system e.g., a sensor in a rolling shutter camera including a processor
  • the dim illumination remains on for the entire process of capturing an image, and reveals an under-illuminated frame.
  • Dim illumination although dimmer than illumination by the flash, is bright enough to perform an automated scene analysis, e.g., face finding, object identification, or the like, that can output a recommendation of a position in the frame that would benefit from flash illumination.
  • a controller or processor can arrange a delay between the start of image writing and the flash trigger such that a stripe of illumination coincides with the desired object. If the object moves in the field-of-view, biometric analysis systems can function dynamically to track movement of the object and change the delay accordingly.
  • the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected.
  • the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest.
  • the region of interest can be tracked by the adaptive trigger module after being detected.
  • Such arrangements to locate and track the object and indicate to the adaptive trigger module changes in the position of the object permits moving objects to be accurately imaged by the biometric analysis systems.
  • the biometric analysis systems disclosed herein can be used for rolling shutter photography (e.g., NIR photography) of objects while outdoors or indoors.
  • the biometric analysis systems can acquire NIR images with controlled illumination in full sun or in dim ambient lighting.
  • An image of a desired object that fills less than a full frame, e.g., a person's face with the entire person in the field-of-view, can be captured by the biometric analysis systems. Due to the low cost, small size and availability of rolling shutter sensors, incorporation of the biometric analysis systems into cost-sensitive or size-sensitive platforms can benefit cost-sensitive or size-sensitive users.
  • the biometric analysis systems can be used in cargo identification in which infrared dye for markings signify authenticity.
  • a box marked with writing can be imaged with a small and inexpensive handheld infrared camera that would look for the marking surrounding an NIR watermark.
  • the flash trigger can be adjusted to fire in order to illuminate the markings, leaving the rest of the field-of-view above and below the mark of interest in the dark.
  • Handheld devices including smartphones, barcode readers, and other mass produced electronic devices, can incorporate the biometric analysis systems at a small and low cost.
  • a rolling shutter camera would reduce the cost for such systems, but would require an extended flash to illuminate an entire scene to guarantee capturing the subject of interest.
  • the exemplary biometric analysis systems include a rolling shutter camera with an adaptive image content-based trigger that issues a shorter flash of a near infrared light that allows for use of smaller batteries, provides for a longer battery life, provides for improved eye safety, and allows for longer illumination source lifetimes.
  • the biometric analysis systems disclosed herein can provide flash illumination for illuminating an object sufficiently to outshine the sun, and in some embodiments can use strobing flash illumination as is disclosed in, e.g., U.S. Pat. Nos. 7,542,628; 7,627,147; 7,657,127; 9,131,141; U.S. Patent Application Publication No. 2015 / 0098630 ; U.S. Patent Application Publication No. 2016 / 0014121 ; U.S. Patent Application Publication No. 2016 / 0012218 ; and U.S. Patent Application Publication No. 2016 / 0012292 , which are incorporated by reference herein.
  • the exemplary biometric analysis systems use a rolling shutter camera and a flash illumination that is triggered adaptively to illuminate the region of interest of the scene or object.
  • the adaptive trigger can be configured to follow or track the subject, and adjusts the flash timing to maintain the stripe of flash illumination on the subject of interest.
  • a face finder can be used as an image analytic to trigger the flash.
  • a unique identifier such as a bar code or QR code, can be detected and illuminated by the stripe of flash illumination.
  • the biometric analysis system can locate the object in a relatively dimly lit scene, and then initiates the repetitive flash that illuminates the object. Flash timing can be adjusted to maintain the object in the illuminated region of the field-of-view of the rolling shutter camera.
  • a further assumption can be made that a flash pulse is started at some time later, e.g., delayed by a time ⁇ t.
  • the delay time At the rows of full flash illumination can be adjusted from the top of the image frame to the bottom to focus on a region of interest located within the frame.
  • the biometric analysis systems can be used to detect the region of interest within the frame of the rolling shutter camera, and intentionally delay the flash illumination in a coordinated manner to ensure that the region of interest is illuminated by the flash during exposure and read-out of the corresponding rows. Proper adjustment of the delay time can cause the stripe of flash illumination due to the short flash pulse to brighten the region of interest.
  • the stripe can scan or search the image in successive frames of a photo or video stream. Such technique can be performed by systematically sweeping an illuminated stripe down the field-of-view and stopping when a region of interest is detected.
  • the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest.
  • the region of interest can be tracked by the adaptive trigger module after being detected.
  • a pre-flash e.g., a single long, low intensity flash that illuminates the full frame and that is both eye-safe and within the operating range of the illumination hardware
  • a pre-flash can provide a single, dim image of the scene.
  • Automated image analysis of the full scene that can tolerate the low light can detect and output a region of interest that is the width of the image and the height of the flash illumination stripe.
  • an automated delay signal can set the illumination stripe to the optimal position to illuminate the region of interest in succeeding frames of a photo or video stream using the full flash brightness over the narrowed region of interest. If the desired feature of the object begins to move out of the illumination stripe (as determined by video analysis performed at the frame rate), the delay can adapt using negative feedback to maintain the position of the desired feature centered within the well-lit region of the image.
  • the biometric analysis systems can be used to capture NIR images of an object that is bathed in sunlight. While a fraction of the solar spectrum covers the NIR wavelengths, it can be assumed that filtered solar NIR irradiance (e.g., from 800 nm to 900 nm) is insufficient to illuminate a video frame that is adequately short in duration to suppress motion blur. It can be further assumed that using a long flash that covers the entire rolling shutter frame period is impractical, even with a short exposure time, due to hardware limitations or eye safety concerns. A short NIR flash pulse that can only illuminate a portion of the frame with an adaptive delay time to ensure that the region of interest is illuminated can therefore be used.
  • filtered solar NIR irradiance e.g., from 800 nm to 900 nm
  • the biometric analysis systems initially detect and identify a region of interest of an object, such as the eyes of a subject.
  • Eye finding can be accomplished using a number of different processes, such as full frame NIR illumination with a relatively dim, eye-safe (and hardware-safe) long NIR flash pulse; partial frame illumination due to a short, intense pulse that is automatically swept frame-by-frame over the full field-of-view until the eyes appear in the search; and/or image acquisition by a color (or other spectrum) auxiliary camera that produces a full frame image allowing eye-finding or face-finding, and then translation from the auxiliary camera frame position to a NIR camera position.
  • the biometric analysis systems automatically set the flash pulse delay to illuminate a stripe across the frame including the eyes (e.g., the region of interest).
  • the delay that sets the vertical position of the illumination stripe can be automatically adjusted to maintain the eyes in substantially the center of the frame. While the portions of the frame outside of the illumination stripe remain unusably dark, the important information about the eyes remains well lit.
  • the competing solar illumination may dimly illuminate the rest of the frame. This is not an issue as long as the flash illumination stripe remains resolved and does not saturate due to the sum of the in-band solar irradiance and the applied NIR light. If the illumination stripe nears the saturation level, the intensity of the applied flash can be throttled back. It is preferable to decrease the NIR illumination irradiance on the object rather than reducing the exposure or pulse duration, since the later strategies reduce the portions of the image that are illuminated.
  • Any spectrum of light can be used as long as the camera uses a rolling shutter with an in-band flash illuminator.
  • visible light NIR, short-wave infrared (SWIR), long-wave infrared (LWIR), ultraviolet (UV), or multi-spectral cameras with rolling shutters that use flash illumination.
  • Any process of finding the region of interest can be used as long as the process returns a coordinate to the biometric analysis system that indicates the time delay between the start of frame and the start of the flash.
  • use of an eye finder that seeks corneal reflections of the flash could be useful if the region of interest contains one or two eyes of the subject.
  • a color face camera could be useful to find the eyes of a subject in iris recognition; but if nose-recognition were used, the region of interest would contain the subject's nose. If the same camera were used to find the region of interest and to acquire the information from the region of interest, a different illumination protocol could be used for each function. For example, a long, relatively dim pulse could illuminate the entire frame, followed by a short, bright pulse to illuminate the region of interest with better fidelity (e.g., improved signal-to-noise ratio). Any process that provides the information needed to set the delay time would also be useful.
  • feedback can be used to maintain the region of interest in the illumination stripe.
  • negative feedback can be used to correct the position of the region of interest within the illumination stripe. If the object of interest begins to move or slip to the top or bottom of the region of interest illuminated by the flash, feedback can be used to correct the delay time. For example, if the stripe illuminating the region of interest needs to move up in the frame to follow the object of interest, the delay would need to be corrected in one direction. If the object of interest moves down in the frame, the sign of the delay correction would need to be in the opposite direction.
  • a feedback signal e.g., an audio alarm, a visual alarm, combinations thereof, or the like
  • the width of the flash illumination stripe relative to the frame height can be set by practical considerations. For example, rolling shutter cameras with very short read-out times generally have wider flash illumination stripes, all other things being equal.
  • the longest flash pulse can be dictated by practical issues, including available power to the illumination source, operating conditions in power, and temperature of the illumination source. For example, it is known that flashing an illumination source too often can damage it, and reduces the safety of human eyes in and around the illumination source.
  • a flash illumination stripe that is narrow relative to the frame height generally should be more accurately timed than a flash illumination stripe that is wider and that can tolerate more motion of the object of interest within the flash illumination stripe.
  • system 100 a schematic block diagram of an exemplary biometric analysis system 100 (hereinafter “system 100 ”) is provided.
  • the system 100 is configured to detect and identify a region of interest of an object, such as the eyes of a person, and generates a time delay of the flash illumination such that the flash coincides with the region of interest.
  • a rolling shutter camera can therefore be used for capturing one or more images for biometric analysis with the delayed flash illumination ensuring that the region of interest will be properly illuminated in the captured images.
  • the system 100 is configured to track the region of interest within the frame or field-of-view of the rolling shutter camera, and adjusts the time delay of the flash pulse appropriately based on movement of the region of interest in and out of the frame.
  • the system 100 generally includes one or more illumination sources 102 .
  • the illumination sources 102 are configured to initially provide dim illumination 104 to a scene including an object to detect and identify the object (or the region of interest of the object).
  • the dim illumination 104 can be separate from the system 100 and can be provided by, e.g., ambient light, any other illumination source, sunlight, combinations thereof, or the like.
  • the illumination sources 102 are further configured to provide flash illumination 106 (e.g., a pulse of flash illumination) to illuminate the region of interest during exposure of the rows corresponding to the region of interest, thereby ensuring that the region of interest will be illuminated in the captured image.
  • a single illumination source 102 can provide both the dim illumination 104 and the flash illumination 106 .
  • separate illumination sources 102 can provide the dim illumination 104 and the flash illumination 106 .
  • the illumination sources 102 can be NIR illumination sources. It should be understood that the dim illumination 104 is less bright than the flash illumination 106 , and provides sufficient ambient light for the system 100 to detect objects in a scene as viewed by the rolling shutter camera.
  • the system 100 includes a rolling shutter camera 108 including a frame 110 with a field-of-view.
  • the rolling shutter camera 108 is configured to capture one or more images (e.g., still frame images, video images, combinations thereof, or the like).
  • the rolling shutter camera 108 is configured to initially capture an image of the scene and the object in the scene under dim illumination 104 conditions.
  • the system 100 includes an adaptive trigger module 112 configured to be executed by a processing device 114 having a processor 116 (e.g., a controller). Although illustrated as separate components from the rolling shutter camera 108 , it should be understood that in some embodiments, the rolling shutter camera 108 can include the illumination sources 102 , the adaptive trigger module 112 , the processing device 114 , or the like.
  • the processing device 114 can be part of a central computing system 118 .
  • the adaptive trigger module 112 can analyze the captured image of the scene illuminated by the dim illumination 104 and detects the region of interest 120 within the scene. For example, the adaptive trigger module 112 can detect and identify the region of interest 120 as one of more eyes of the user, and stores the rows of the frame 110 corresponding to the region of interest 120 in a database 122 (e.g., a local or remote database).
  • a database 122 e.g., a local or remote database
  • the adaptive trigger module 112 Based on the rows of the frame 110 corresponding to the region of interest 120 , the adaptive trigger module 112 arranges a flash pulse delay 124 between the start of image writing by the rolling shutter camera 108 and a trigger of the flash illumination 106 such that the flash illumination 108 coincides with the identified region of interest 120 . For example, if the eyes of the user are the region of interest 120 corresponding with rows 400 - 600 of the frame 110 , the adaptive trigger module 112 arranges the flash pulse delay 124 such that the flash illumination 106 is triggered to illuminate the object during rows 400 - 600 .
  • the image writing can start before triggering of the flash illumination 106 and the flash illumination 106 is only triggered to illuminate the region of interest 120 .
  • Such flash delay 124 ensures that the region of interest 120 is illuminated in the images 126 captured by the rolling shutter camera 108 .
  • the resulting image 126 generally includes a stripe of flash illumination 106 extending across the image 126 and over the eyes of the user (see, e.g., FIG. 9 ). Depending on the length of the exposure time, the height of the stripe of flash illumination 106 can be greater or smaller relative to the top and bottom of the image 126 .
  • the adaptive trigger module 112 can include a face finder algorithm that locates the face and/or features of the face in the scene.
  • the face finder algorithm can therefore locate the face identifies one or more eyes of the face as the region of interest 120 .
  • the region of interest 120 can be a unique identifier associated with a physical item that can be detected by an identifier finder algorithm of the adaptive trigger module 112 .
  • the identifier finder algorithm of the adaptive trigger module 112 can be used to detect, e.g., a barcode, a QR code, or the like, on a physical item, identifies such unique identifier as the region of interest 120 , and adjusts the flash delay 124 to ensure that the unique identifier is illuminated during capture of the image 126 .
  • the adaptive trigger module 112 can be configured to track movement of the object within the field-of-view or frame 110 of the rolling shutter camera 108 .
  • the adaptive trigger module 112 can be configured to continuously scan the scene in the frame 110 to detect the position of the region of interest 120 . If the user moves such that the region of interest 120 in the frame 110 is not centered or changes relative to the previously detected rows for the region of interest 120 , the adaptive trigger module 112 can determine the new rows corresponding to the region of interest 120 , stores the new rows in the database 122 , and adjusts the flash delay 124 to coincide with the new rows for the region of interest 120 .
  • an object moving in the frame 110 of the rolling shutter camera 108 can be tracked by the adaptive trigger module 112 to allow for the region of interest 120 of a moving object to be properly illuminated by the flash illumination 106 .
  • the system 100 includes a feedback module 128 configured to be executed by the processing device 114 and/or the central computing system 118 .
  • the rolling shutter camera 108 can include the feedback module 128 .
  • the feedback module 128 receives as input the images 126 captured by the rolling shutter camera 108 and analyzes the region of interest 120 captured in the image 126 to determine if the region of interest 120 is illuminated by the stripe of flash illumination 106 . If the region of interest 120 is illuminated by the stripe of flash illumination 106 , the feedback module 128 electronically outputs such findings to the adaptive trigger module 112 and no changes are made in the flash delay 124 .
  • the feedback module 128 electronically outputs such findings to the adaptive trigger module 112 .
  • the adaptive trigger module 112 analyzes the frame 110 to locate the region of interest 120 , determines the new rows associated with the region of interest 120 , and adjusts the flash pulse delay 124 to trigger the flash illumination 106 at the appropriate rows coinciding with the region of interest 120 .
  • the feedback loop can be continuously performed to ensure the region of interest 120 is always illuminated by the stripe of flash illumination 106 .
  • the system 100 can include a user interface 130 with a graphical user interface (GUI) 132 for receiving input from a user and for outputting information to the user.
  • GUI graphical user interface
  • the user interface 130 can be used to initiate the process performed by the system 100 .
  • the user interface 130 can output an audio and/or visual alarm to the user when the region of interest 120 is no longer centered in the frame 110 .
  • the system 100 can include a communication interface 134 configured to provide a means for electronic transmission between components of the system 100 , e.g., the illumination sources 102 , the processing device 114 , the rolling shutter camera 108 , the adaptive trigger module 112 , the database 122 , the user interface 130 , the central computing system 118 , and feedback module 128 , or the like.
  • a communication interface 134 configured to provide a means for electronic transmission between components of the system 100 , e.g., the illumination sources 102 , the processing device 114 , the rolling shutter camera 108 , the adaptive trigger module 112 , the database 122 , the user interface 130 , the central computing system 118 , and feedback module 128 , or the like.
  • FIG. 10 is a diagrammatic view of an exemplary biometric analysis system 200 (hereinafter “system 200 ”) of the present disclosure.
  • the system 200 includes an illumination source 202 that provides dim illumination to the subject 204 in the scene.
  • the illumination source 202 can provide 1 ms NIR pulses to the scene.
  • the illumination source 202 can be an NIR light source not connected to the camera 208 .
  • the illumination source 202 can be positioned a distance 206 from the subject 204 .
  • the illumination source 202 can also be configured to provide the flash illumination to the subject 204 .
  • the system 200 includes a camera 208 including a rolling shutter (e.g., a rolling shutter camera, such as a smartphone camera, or the like) configured to capture one or more images of the subject.
  • the camera 208 can have a field-of-view represented by area 216 , including the region of interest of the subject 204 .
  • the camera 208 includes an illumination source 210 (e.g., an integrated NIR light source) configured to provide flash illumination to a region of interest (e.g., the eyes of the subject 204 ).
  • area 212 indicates the coverage of the NIR flash illumination provided by the illumination source 210 , with the region of interest on the subject 204 being within the area 212 .
  • the camera 208 can be spaced by a distance 214 from the subject 204 .
  • the illumination source 202 appears separate from camera 202 , however, one skilled in the art shall appreciate that, in some embodiments, the illumination source 202 can be incorporated into the camera 208 assembly such that the camera 208 assembly is capable of providing both the dim illumination and flash illumination to the subject 204 .
  • FIG. 11 is a diagrammatic view of a frame action 300 of a rolling shutter camera of the exemplary biometric analysis systems disclosed herein.
  • the frame action 300 includes a total frame time 302 .
  • Ambient light 304 having a low illumination level 306 can be provided for all or part of the total frame time 302 by a dim illumination source and/or sunlight.
  • the total camera exposure time 308 indicates when the rolling shutter camera begins and ends image writing and read-out.
  • a pulse of flash illumination 310 (e.g., NIR from a light-emitting diode (LED) illumination source) is triggered to coincide with the rows of the region of interest.
  • the pulse of flash illumination 310 has a time or pulse width 312 that is significantly smaller than the camera exposure time 308 , and an illumination level 314 that is brighter or higher than the level 306 of the ambient light 304 .
  • the region of interest is thereby illuminated by the flash illumination 310 while the remaining rows of the captured image are only illuminated by the ambient light 304 .
  • a 1280 ⁇ 960 sensor can be used with the rolling shutter camera.
  • a 120 fps VGA (640 ⁇ 480) sensor can be used with the rolling shutter camera to allow for full frame exposure in 4 ms, and can be synchronized with an approximately 4 ms LED flash every 8 th frame.
  • the VGA output can provide (unscaled) 15 fps feedback to the user facing the monitor of the system.
  • the 4 ms can provide sufficient flash pulse while allowing for cost effective use of an LED and rolling shutter accommodation, and allowing for eye safety.
  • a bright light (e.g., approximately 850 nm LED, two LEDs) can be used for the flash pulse.
  • the systems can therefore have high pulsed current capabilities.
  • the rolling shutter camera can have a high quality camera lens and NIR filtering.
  • the NIR filter can be tighter than 100 nm around the LED center wavelength.
  • the system can include optics with a modulation transfer function (MTF) optimized for the spatial frequencies that enter the system.
  • the system can include preprocessing of iris images for a particular spectrum of spatial frequencies (e.g., low, medium, high, or the like).
  • FIG. 12 is a flowchart illustrating an exemplary process 400 of implementing the biometric analysis systems disclosed herein.
  • a scene and an object in the scene are illuminated with dim illumination from one or more illumination sources.
  • the scene is analyzed with an adaptive trigger module to detect the object (or a region of interest associated with the object) in the scene during dim illumination.
  • a position in a frame of a rolling shutter camera is determined with the adaptive trigger module, the position coinciding with the detected object (or region of interest associated with the object) in the scene.
  • a delay is arranged by the adaptive trigger module between a start of image writing by the rolling shutter camera and a trigger of the illumination sources such that a stripe of flash illumination provided by the illumination sources coincides with the detected object (or region of interest associated with the object) in the scene.
  • movement of the object is tracked within a field-of-view of the rolling shutter camera with the adaptive trigger module.
  • the delay between the start of image writing by the rolling shutter camera and the trigger of the illumination sources is modified with the adaptive trigger module based on detected movement of the object (or region of interest associated with the object) out of the field-of-view.
  • FIG. 13 is a flowchart illustrating an exemplary process 500 of implementing the biometric analysis systems disclosed herein.
  • the adaptive trigger module can set the flash delay at 0 ms, the flash duration equal to the time of the full frame, and the flash brightness to low.
  • a full frame image can be acquired with the rolling shutter camera under dim illumination.
  • the adaptive trigger module analyzes the image captured at step 504 and locates the region of interest in the image.
  • the adaptive trigger module sets the flash delay to a value greater than 0 ms, the flash duration to short (e.g., a pulse), and the flash brightness to high (e.g., higher than the dim illumination) to illuminate the region of interest.
  • the adaptive trigger module measures the position of the object of interest within the region of interest (e.g., eyes of the user located on the user's face).
  • the process 500 continues a loop to constantly (or at predetermined frequencies of time) ensure that the object is maintained within the region of interest.
  • the process 500 performs step 516 to correct the flash delay such that the object is centered within the region of interest.
  • the region of interest can thereby be maintained in the flash illumination stripe despite relative motion of the object and/or the camera.
  • the process can include a loop that analyzes the captures images to control the intensity of the flash illumination, thereby avoiding saturation of the object.
  • FIG. 14 is a block diagram of a computing device 600 in accordance with exemplary embodiments of the present disclosure.
  • the computing device 600 includes one or more non-transitory computer-readable media for storing one or more computer-executable instructions or software for implementing exemplary embodiments.
  • the non-transitory computer-readable media may include, but are not limited to, one or more types of hardware memory, non-transitory tangible media (for example, one or more magnetic storage disks, one or more optical disks, one or more flash drives), and the like.
  • memory 606 included in the computing device 600 may store computer-readable and computer-executable instructions or software for implementing exemplary embodiments of the present disclosure (e.g., instructions for operating the illumination sources, instructions for operating the processing device, instructions for operating the rolling shutter camera, instructions for operating the adaptive trigger module, instructions for operating the communication interface, instructions for operating the user interface, instructions for operating the central computing system, instructions for operating the feedback module, combinations thereof, or the like).
  • instructions for operating the illumination sources e.g., instructions for operating the illumination sources, instructions for operating the processing device, instructions for operating the rolling shutter camera, instructions for operating the adaptive trigger module, instructions for operating the communication interface, instructions for operating the user interface, instructions for operating the central computing system, instructions for operating the feedback module, combinations thereof, or the like.
  • the computing device 600 also includes configurable and/or programmable processor 602 and associated core 604 , and optionally, one or more additional configurable and/or programmable processor(s) 602 ′ and associated core(s) 604 ′ (for example, in the case of computer systems having multiple processors/cores), for executing computer-readable and computer-executable instructions or software stored in the memory 606 and other programs for controlling system hardware.
  • Processor 602 and processor(s) 602 ′ may each be a single core processor or multiple core ( 604 and 604 ′) processor.
  • Virtualization may be employed in the computing device 600 so that infrastructure and resources in the computing device 600 may be shared dynamically.
  • a virtual machine 614 may be provided to handle a process running on multiple processors so that the process appears to be using only one computing resource rather than multiple computing resources. Multiple virtual machines may also be used with one processor.
  • Memory 606 may include a computer system memory or random access memory, such as DRAM, SRAM, EDO RAM, and the like. Memory 606 may include other types of memory as well, or combinations thereof.
  • a user may interact with the computing device 600 through a visual display device 618 (e.g., a personal computer, a mobile smart device, or the like), such as a computer monitor, which may display one or more user interfaces 620 (e.g., a graphical user interface) that may be provided in accordance with exemplary embodiments.
  • the computing device 600 may include other I/O devices for receiving input from a user, for example, a camera, a sensor, a keyboard or any suitable multi-point touch interface 608 , a pointing device 610 (e.g., a mouse).
  • the keyboard 608 and the pointing device 610 may be coupled to the visual display device 618 .
  • the computing device 600 may include other suitable conventional I/O peripherals.
  • the computing device 600 may also include one or more storage devices 624 , such as a hard-drive, CD-ROM, eMMC (MultiMediaCard), SD (secure digital) card, flash drive, non-volatile storage media, or other computer readable media, for storing data and computer-readable instructions and/or software that implement exemplary embodiments of the biometric analysis systems described herein.
  • Exemplary storage device 624 may also store one or more databases 626 for storing any suitable information required to implement exemplary embodiments.
  • exemplary storage device 624 can store one or more databases 626 for storing information, such as data relating to captured images 126 under dim illumination 104 and flash illumination 106 , regions of interest 120 , flash delay 124 , combinations thereof, or the like, and computer-readable instructions and/or software that implement exemplary embodiments described herein.
  • the databases 626 may be updated by manually or automatically at any suitable time to add, delete, and/or update one or more items in the databases.
  • the computing device 600 can include a network interface 612 configured to interface via one or more network devices 622 with one or more networks, for example, Local Area Network (LAN), Wide Area Network (WAN) or the Internet through a variety of connections including, but not limited to, standard telephone lines, LAN or WAN links (for example, 802.11, T1, T3, 56 kb, X.25), broadband connections (for example, ISDN, Frame Relay, ATM), wireless connections, controller area network (CAN), or some combination of any or all of the above.
  • LAN Local Area Network
  • WAN Wide Area Network
  • the Internet through a variety of connections including, but not limited to, standard telephone lines, LAN or WAN links (for example, 802.11, T1, T3, 56 kb, X.25), broadband connections (for example, ISDN, Frame Relay, ATM), wireless connections, controller area network (CAN), or some combination of any or all of the above.
  • the network interface 612 may include a built-in network adapter, network interface card, PCMCIA network card, PCI/PCIe network adapter, SD adapter, Bluetooth adapter, card bus network adapter, wireless network adapter, USB network adapter, modem or any other device suitable for interfacing the computing device 600 to any type of network capable of communication and performing the operations described herein.
  • the computing device 600 may be any computer system, such as a workstation, desktop computer, server, laptop, handheld computer, tablet computer (e.g., the tablet computer), mobile computing or communication device (e.g., the smart phone communication device), an embedded computing platform, or other form of computing or telecommunications device that is capable of communication and that has sufficient processor power and memory capacity to perform the operations described herein.
  • the computing device 600 may run any operating system 616 , such as any of the versions of the Microsoft® Windows® operating systems, the different releases of the Unix and Linux operating systems, any version of the MacOS® for Macintosh computers, any embedded operating system, any real-time operating system, any open source operating system, any proprietary operating system, or any other operating system capable of running on the computing device and performing the operations described herein.
  • the operating system 616 may be run in native mode or emulated mode.
  • the operating system 616 may be run on one or more cloud machine instances.
  • FIG. 15 is a block diagram of an exemplary biometric analysis system environment 700 in accordance with exemplary embodiments of the present disclosure.
  • the environment 700 can include servers 702 , 704 configured to be in communication with one or more illumination sources 706 , one or more rolling shutter cameras 708 , one or more adaptive trigger modules 710 , a feedback module 712 , a user interface 714 , and a central computing system 716 via a communication platform 922 , which can be any network over which information can be transmitted between devices communicatively coupled to the network.
  • the communication platform 722 can be the Internet, Intranet, virtual private network (VPN), wide area network (WAN), local area network (LAN), and the like.
  • the communication platform 722 can be part of a cloud environment.
  • the environment 700 can include repositories or databases 718 , 720 , which can be in communication with the servers 702 , 904 , as well as the one or more illumination sources 706 , one or more rolling shutter cameras 708 , one or more adaptive trigger modules 710 , the feedback module 712 , the user interface 714 , and the central computing system 716 , via the communications platform 722 .
  • the servers 702 , 704 , one or more illumination sources 706 , one or more rolling shutter cameras 708 , one or more adaptive trigger modules 710 , the feedback module 712 , the user interface 714 , and the central computing system 716 can be implemented as computing devices (e.g., computing device 600 ).
  • the databases 718 , 720 can be incorporated into one or more of the servers 702 , 704 .
  • the database 718 can store data relating to captured images, regions of interest 120 , flash delay 124 , combinations thereof, or the like, can be distributed over multiple databases 718 , 720 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)
  • Image Input (AREA)
  • Exposure Control For Cameras (AREA)

Abstract

Exemplary embodiments are directed to biometric analysis systems including one or more illumination sources configured to provide dim illumination to a scene including an object and configured to provide flash illumination to the object in the scene. The biometric analysis systems include a rolling shutter camera configured to capture one or more images. The biometric analysis systems include an adaptive trigger module configured to analyze the scene to detect the object in the scene during dim illumination of the scene, determine a position in a frame of the rolling shutter camera that coincides with the detected object in the scene, and arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • The present application claims the benefit of priority to U.S. Provisional Patent Application No. 62/316,347, filed Mar. 31, 2016, which is hereby incorporated by reference in its entirety.
  • TECHNICAL FIELD
  • The present disclosure relates to systems and methods of biometric analysis and, in particular, to biometric analysis systems including an adaptive trigger configured to align an object (e.g., a subject, a barcode, or the like) within a camera field-of-view to improve application of flash illumination to the desired area of the object for capture and analysis.
  • BACKGROUND
  • Security is a concern in a variety of transactions involving private information. As an example, iris recognition is a well-accepted and accurate means of biometric identification used in government and commercial systems around the world that enables secure transactions and an added layer of security beyond keys and/or passwords. Due to the increased security provided by iris recognition systems, an increase in use of such systems has occurred around the world.
  • Traditional cameras used in biometric identification are generally expensive. Video cameras that use complementary metal-oxide-semiconductor (CMOS) or semiconductor charge-coupled device (CCD) image sensors typically use electronic shutters to determine the time period over which the sensor measures light. A trigger signal opens the shutter for a predetermined time during which each pixel within the sensor array collects (integrates) incoming light. At the end of the exposure, the signal collected (integrated) in each pixel during the exposure remains fixed and is then systematically read out, converted to a digital signal, and processed to become an image. Pixels are then cleared and readied for the next exposure to light.
  • There are different types of electronic shutters used in the industry. FIG. 1 is a diagram of a traditional global shutter camera schedule of events for signal collection, including the timeline for exposure of each row of an image sensor in the global shutter camera. Global shutters simultaneously expose an entire array of pixels, e.g., N rows by M columns, during signal collection. During a single exposure event (during texp), light is simultaneously collected in each pixel. When the global shutter closes, the light signal within each pixel represents the image during the period of the single exposure. All pixels integrate signal over exactly the same period of time. Global shutter cameras avoid flash timing issues incurred when using rolling shutter cameras. However, global shutter cameras are expensive options for biometric analysis, thereby increasing the overall costs associated with biometric analysis systems.
  • Rolling shutter cameras save cost and size in their sensor design. FIG. 2 is a diagram of a traditional rolling shutter schedule of events for signal collection, including the rolling shutter timeline. Rolling shutters expose an array of pixels differently from global shutters. A rolling shutter system exposes a first row of pixels for an exposure time (texp) and then commences to read-out the exposed row of pixels for digitization. The read-out process occupies a unique onboard resource for a period referred to as a read-out time during which no other row can be read-out. To minimize the duration of the total exposure including the read-out process, the rolling shutter exposes the second row of pixels during a time that is equal to but delayed from the first row by a read-out time. The second row is thereby exposed to light and ready to be read-out at the moment that the read-out process for the first row is complete. The third row is exposed to light for a time interval equal in length to that of the first two rows but delayed relative to the second row by a read-out time allowing for the required time to read-out the second row. The process “rolls” down the pixel array reading row-by-row in sequence taking a total time equal to the exposure time for a single row plus the read-out time interval, times the number of rows. The time interval during which a row is exposed and therefore the events captured by that row are different for each row for a rolling shutter sensor. This is a key difference from a global shutter sensor, especially when using a short flash.
  • As shown in FIGS. 1 and 2, the light collection time period for each row of a sensor with a global shutter is simultaneous while the time periods of light collection for each row of a sensor equipped with a rolling shutter are not simultaneous. Rather, light collection time periods for each row of a rolling shutter are offset from one another with a delay between rows equal to the row read-out time. The different exposure techniques result in image shearing. For example, FIG. 3 shows an image in which a moving fan blade was captured by an image sensor with a global shutter with no or little distortion as compared to the same moving fan captured by an image sensor with a rolling shutter shown in FIG. 4. Image shearing is an inevitable consequence of the row-by-row time delays built into a rolling shutter sensor in which each row “sees” the scene over a slightly different and offset time interval.
  • The row-by-row delay in exposure of a rolling shutter also has an effect on coordinating an exposure with flash illumination. As discussed herein, a flash refers to a short, intense period of illumination of a subject during which the light applied by the flash dominates other sources of light on the scene (e.g., the area surrounding the subject). FIG. 5 shows the exposure of each row during texp with the shaded region indicating the duration of time of the flash illumination occurring simultaneous during texp. In a global shutter, the exposure of all the pixels in a sensor can be coordinated with the application of the flash illumination to the scene. For example, if the period of the flash pulse is 1 ms, the global shutter can open simultaneously with the start of the pulse and close simultaneously with the end of the pulse 1 ms later. The flash illuminates the pixels during and only during their global exposure. Light forming the image is, by assumption, dominated by the light applied to the scene by the flash. If, for example, sunlight is present in the scene, the irradiance on the object from the flash is significantly brighter than that of the sunlight during the flash pulse when pixels are exposed by the global shutter.
  • Coordinating a flash pulse with a rolling shutter exposure is more complicated than with a global shutter. FIG. 6 shows the rolling shutter exposure during texp for each row extending diagonally across the diagram, and the time period for flash illumination illustrated as the vertical shaded region tp. After a flash delay, the flash pulse occurs between the start pulse and end pulse points of the diagram. Because the exposure period for each row is delayed from the previous row by a short read-out time interval, illumination of a full frame requires that a flash pulse provide illumination during a period when all rows are integrating light. Failure to meet this condition creates a situation in which some rows of pixels integrate light from the flash pulse while some do not, and perhaps some rows integrate light from only a portion of the flash pulse. In this case, the image is unevenly illuminated. As shown in FIG. 6, some rows are finished integrating before the flash starts and other rows do not start integrating until after the flash ends. In addition, other rows integrate a partial flash and some integrate the full flash. Thus, a subset of lines on the rolling shutter sensor receive adequate illumination, but outside of this set of lines, the other parts of the sensor remain largely dark.
  • One example of a short flash pulse can be considered with respect to FIG. 6, which, across the top horizontal line, four horizontal dashed lines, and bottom lines, respectively shows row numbers 0, 200, 450, 600, 850, and 1000. The flash pulse can start as row 200 of 1000 rows finishes integrating the signal and begins to read-out, indicated by the top dashed line. The flash pulse can end as row 850 of 1000 begins integrating the signal, indicated by the bottom dashed line. FIG. 6 shows that rows 450 through 599 receive the full illumination of the flash pulse, as bracketed by the middle two dashed lines. However, rows 200 to 449 and rows 600 to 849 only receive a portion of the flash illumination while rows outside of these ranges, e.g., rows 1 to 199 and 850 to 1000, receive no flash illumination. Assuming insignificant ambient light, the resulting image would show an illumination stripe surrounded by dim regions. The transition from bright to dim at the top and bottom of the stripe is due to rows that receive flash illumination over a fraction of the total pulse time. FIG. 7 shown a portion of an image acquired using a rolling shutter camera with a delayed flash pulse in which the recorded irradiance is plotted to show dark regions before and after the flash, ramp-up and ramp-down regions of partial illumination, and a plateau region of complete flash illumination. The plateau region of FIG. 7 is not flat because the flash itself was not uniform over the field-of-view. As another example, an image captured using a rolling shutter camera would include a horizontal stripe with a vertical height proportional to the duration of the flash illumination. In cases with bright ambient illumination, the un-flashed portion of the image would appear, but might be significantly dimmer if the flash illumination is brighter than the ambient illumination.
  • When an image of a particular object is desired with a rolling shutter camera, a trigger signal can be initiated by the sensor controller to fire the flash at a preset time relative to the start of image writing. For example, the flash can fire when the first line is written and can remain on for 50 of 1000 lines. The resultant image would be flash illuminated for the top 5% of the image and would be dark elsewhere. The same flash can be delayed until the 500th line of 1000 lines, resulting in an image with a stripe of illuminated content approximately halfway down the frame. With such an arrangement, the photographer would need to align the subject within the camera field-of-view such that the stripe of illumination detected by the sensor corresponds to the position of the desired object.
  • Traditionally, one solution to the problem of flash illuminating an image using a rolling shutter has been to use an extended period of illumination, e.g., a flash pulse that is started simultaneously with the beginning of the exposure of the first row of pixels and is not finished until the last row of pixels has been exposed. The extended period of illumination is needed to expose the entire image since image lines are written sequentially rather than all at once (as is the case with a camera including a more expensive and physically larger global shutter sensor). This technique necessitates a longer flash pulse compared to the global shutter case, and would show up in FIG. 6 as a shaded region covering all of the rows with a duration equal to the frame time. This technique would also illuminate the full frame shown in FIG. 7. Additional requirements on the flash in terms of power output, heating and reliability are needed based on the longer pulse for this technique. A longer pulse might also challenge requirements for eye-safety. For these reasons, full frame pulses with rolling shutters are considered impractical.
  • Thus, a need exists for improved biometric analysis systems including a rolling shutter that are capable of illuminating and capturing the desired area of an object for identification without an extended flash. These and other needs are addressed by the systems and methods of biometric analysis of the present disclosure.
  • SUMMARY
  • In accordance with embodiments of the present disclosure, an exemplary biometric analysis system is provided that includes one or more illumination sources configured to provide dim illumination to a scene including an object, and further configured to provide flash illumination to the object in the scene. In some embodiments, the dim illumination can be provided by an illumination source external and separate from the biometric analysis system, such as ambient light, sunlight, any other light source, or the like (e.g., one or more of the illumination sources can be ambient light). In some embodiments, the biometric analysis system can include a single illumination source that provides the dim illumination, with the same illumination source providing the flash illumination at the determined time period. In some embodiments, the biometric analysis system can include a first illumination source that provides the dim illumination, and a second (separate) illumination source that provides the flash illumination. In some embodiments, the first illumination source can continue to provide the dim illumination during the flash illumination from the second illumination source. In some embodiments, the first illumination source can be automatically actuated into a non-illuminating configuration during the flash illumination provided by the second illumination source, and automatically actuated into an illuminating configuration after the flash illumination is complete.
  • The biometric analysis system includes a rolling shutter camera configured to capture one or more images. The rolling shutter camera generally includes a frame with a field-of-view. The term “image” as used herein can include still frame images, video, combinations thereof, or the like. The biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device. The adaptive trigger module, when executed, can be configured to analyze the scene to detect the object in the scene during dim illumination of the scene. The adaptive trigger module, when executed, can be configured to determine a position in the frame of the rolling shutter camera that coincides with the detected object in the scene. The adaptive trigger module, when executed, can be configured to arrange a delay (a time delay) between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
  • In some embodiments, the adaptive trigger module, when executed, can be configured to track movement of the object within a field-of-view of the rolling shutter camera. In such embodiments, the adaptive trigger module can be configured to modify the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources based on detected movement of the object within the field-of-view.
  • In some embodiments, the adaptive trigger module can be configured to detect a region of interest of the object and arranges the delay such that the stripe of flash illumination coincides with the detected region of interest of the object. In one embodiment, the region of interest of the object can include one or both eyes of a person.
  • In some embodiments, the object can be a person. In such embodiments, the adaptive trigger module includes a face finder configured to detect a face (and/or features of the face) of the person. In some embodiments, the object can be a physical item. In such embodiments, the adaptive trigger module can include an identifier finder configured to detect a unique identifier (e.g., a barcode, a quick response (QR) code, combinations thereof, or the like) associated with the physical item.
  • In some embodiments, the one or more illumination sources can be configured to provide the flash illumination as a synchronized pulse of flash illumination. The flash illumination provided by the one or more illumination sources is brighter than the dim illumination provided by the one or more illumination sources. In one embodiment, the one or more illumination sources can be near infrared (NIR) illumination sources. In some embodiments, the adaptive trigger module can be configured to sweep an illuminated stripe down the frame as the rolling shutter camera captures the one or more images, analyze an illuminated section of the one or more images to identify a region of interest in the illuminated section, and stop sweeping of the illuminated stripe when the region of interest is identified.
  • In accordance with embodiments of the present disclosure, an exemplary biometric analysis system is provided that includes one or more illumination sources configured to provide dim illumination to a scene including a subject and configured to provide flash illumination to the subject in the scene. The biometric analysis system includes a rolling shutter camera configured to capture one or more images. The biometric analysis system includes an adaptive trigger module configured to be executed by a controller or processing device. The adaptive trigger module, when executed, can be configured to analyze the scene to detect eyes of the subject in the scene during dim illumination of the scene. The adaptive trigger module, when executed, can be configured to identify the eyes of the subject as a region of interest. The adaptive trigger module, when executed, can be configured to determine a position in a frame of the rolling shutter camera that coincides with the identified region of interest. The adaptive trigger module, when executed, can be configured to arrange a flash pulse delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the identified region of interest. In some embodiments, rather than or in addition to using dim illumination, the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected. Thus, rather than using a single frame (as in dim illumination), the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest. The region of interest can be tracked by the adaptive trigger module after being detected.
  • The flash pulse delay can ensure that the illuminated region of interest is substantially maintained in a center of the frame of the rolling shutter camera. The biometric analysis system can include a feedback module configured to be executed by the controller or processing device. The feedback module, when executed, can be configured to analyze a captured image of the region of interest and determine if the region of interest is illuminated by the stripe of the flash illumination. The adaptive trigger module, when executed, can be configured to adjust the flash pulse delay based on the determination of the feedback module to ensure that the region of interest is illuminated by the stripe of the flash illumination.
  • In accordance with embodiments of the present disclosure, an exemplary method of biometric analysis is provided. The method includes illuminating a scene and an object in the scene with dim illumination from one or more illumination sources. The method includes analyzing the scene with an adaptive trigger module to detect the object in scene during dim illumination. The method includes determining a position in a frame of a rolling shutter camera that coincides with the detected object in the scene. The method includes arranging a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
  • In some embodiments, the method can include tracking movement of the object within a field-of-view of the rolling shutter camera with the adaptive trigger module. In such embodiments, the method can include modifying the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources with the adaptive trigger module based on detected movement of the object within the field-of-view.
  • In accordance with embodiments of the present disclosure, an exemplary non-transitory computer-readable medium storing instructions for biometric analysis is provided. Execution of the instructions by a processing device (or controller) causes the processing device to illuminate a scene and an object in the scene with dim illumination from one or more illumination sources. Execution of the instructions by a processing device causes the processing device to analyze the scene with an adaptive trigger module to detect the object in scene during dim illumination. Execution of the instructions by a processing device causes the processing device to determine a position in a frame of a rolling shutter camera that coincides with the detected object in the scene. Execution of the instructions by a processing device causes the processing device to arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of flash illumination provided by the one or more illumination sources coincides with the detected object in the scene.
  • Other objects and features will become apparent from the following detailed description considered in conjunction with the accompanying drawings. It is to be understood, however, that the drawings are designed as an illustration only and not as a definition of the limits of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • To assist those of skill in the art in making and using the disclosed systems and methods of biometric analysis, reference is made to the accompanying figures, wherein:
  • FIG. 1 (prior art) is a diagram of a traditional global shutter camera schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out.
  • FIG. 2 (prior art) is a diagram of a traditional rolling shutter camera schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with delay between rows being equal to read-out time.
  • FIG. 3 (prior art) is an image of a moving fan illustrating minimal distortion as captured by a traditional global shutter camera.
  • FIG. 4 (prior art) is an image of a moving fan illustrating shearing distortion effects caused by cascading row delays as captured by a traditional rolling shutter camera.
  • FIG. 5 (prior art) is a diagram of a traditional global shutter schedule of events showing simultaneous exposure of each row between start and end of exposure followed by simultaneous read-out, with a flash firing during the same time interval as that in which the rows are exposed (texp).
  • FIG. 6 (prior art) is a diagram of a traditional rolling shutter schedule of events showing start of frame, cascading exposure of pixel rows, and rolling read-out, with a flash firing after the beginning of the frame, illuminating some but not all rows, and some rows partially but others completely, resulting in a ramp-up, plateau and ramp-down periods of flash illumination.
  • FIG. 7 (prior art) is a diagram of a vertical strip of an image (re-oriented horizontally to align with the time axis) showing dark, ramp-up, plateau, ramp-down, and another dark region due to timing of flash in a traditional rolling shutter frame.
  • FIG. 8 is a schematic block diagram of an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 9 is a representation of an image captured by a rolling shutter camera of the exemplary biometric analysis system including a stripe of flash illumination across a region of interest.
  • FIG. 10 is a diagrammatic view of an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 11 is a diagrammatic view of a frame action of a rolling shutter camera of an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 12 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 13 is a flowchart illustrating an exemplary process of implementing an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 14 is a block diagram of an exemplary computing device for implementing an exemplary biometric analysis system in accordance with the present disclosure.
  • FIG. 15 is a block diagram of an exemplary biometric analysis system environment in accordance with the present disclosure.
  • DESCRIPTION OF EXEMPLARY EMBODIMENTS
  • In accordance with embodiments of the present disclosure, exemplary biometric analysis systems are described herein for a rolling shutter camera that include an adaptive trigger for ensuring that the stripe of light from the flash illumination illuminates a region of interest of an object, such as the eyes of a person, during image writing. The biometric analysis systems generally include one or more illumination sources that initially provide dim illumination to a scene to detect the object and identify the region of interest of the object. Based on the identified region of interest, a delay between the start of image writing by the rolling shutter camera and the trigger of the illumination source is arranged or coordinated to ensure that the stripe of the flash illumination coincides with the region of interest of the object.
  • Cameras that include rolling shutter image sensors are typically cheaper than those with global shutter image sensors. The exemplary biometric analysis systems use flash illumination in combination with inexpensive rolling shutter cameras of the type found in low-cost cameras and smartphones. The biometric analysis systems address a rolling shutter-specific problem that occurs when the brightness of the ambient light in the scene may need an exposure time less than the full frame time to avoid saturation, and the exposure time is sufficiently short that the first rows of the image are finished integrating light and reading out before the last lines of the image have begun integration. In cases of short exposure times, a single, short flash pulse will not illuminate an entire frame, leaving portions of the frame relatively dark. Long exposure times can result in motion blur.
  • The exemplary biometric analysis systems balance the following parameters to ensure the captured image quality is sufficient for identification, and further ensuring that the region of interest is adequately illuminated by the flash: sensor frame rate, sensor exposure time, flash pulse intensity, flash pulse duration, flash pulse repetition, processed frame rate, and flash pulse offset. Increasing the sensor frame rate over the available ranges increases the speed that the reset and read lines move over the image. For a given sensor exposure time, increasing the frame rate effectively increases the number of exposed lines during that period. For example, with the exposure time remaining the same, the execution time of the internal clock of the camera can be increased to operate the camera at a faster rate, thereby capturing a greater number of lines during the exposure time. The drawbacks of increased sensor frame rates are additional data transfers and additional heating due to higher clock speed. Lower frame rates may be desirable for dark indoor lighting conditions and higher frame rates may be desirable for outdoor bright lighting conditions.
  • Increasing the sensor exposure time increases the distance between the reset and read times. This effectively increases the number of exposed lines during that period, allowing a short flash pulse to cover more lines. The drawback of increasing this period is that additional ambient light is accumulated over the entire period. During outdoor operation, it may be beneficial to keep the exposure time as short as possible, but allowing enough lines to be covered by the flash. Keeping the exposure time shorter is also important to avoid introducing motion blur which can degrade recognition performance.
  • The flash pulse needs to be bright enough to overpower direct sunlight for effective outdoor operation. The flash pulse is limited by available system power and eye safety limits. In general, eye safety can be roughly proportional to the total amount of energy absorbed by the eye over a period of time (e.g., t−0.75). Reducing the intensity, pulse duration, and pulse repetition all move the system into a safer standing. Longer flash pulses allow more lines to be exposed in a rolling shutter. However, longer pulses can lead to increased eye safety concerns forcing the system to use less intense pulses when longer durations are used.
  • Several short flash pulses may be used back-to-back to illuminate additional lines with minimal overlap between exposed lines. This technique can reduce the total eye exposure and power requirements while expanding the total number of lines that can be properly illuminated. Pulse rate, duration, and intensity can all impact the system power requirements. Increasing any of these parameters generally increases the demand for power.
  • In some embodiments, since eye safety depends on the total exposure over a period of time, one method to reduce the total exposure can be to selectively process frames. While images can be captured at 60 frames per second (fps), only one in four can be illuminated with the flash, which would reduce the exposure to an equivalent of 15 fps with the advantages of capturing at 60 fps. Such technique allows for improved line read speeds, bringing the rolling shutter camera closer to the global shutter camera timing.
  • In some embodiments, the systems can include a continuous feedback loop to a user who is attempting to align their eye with a small box on the screen (e.g., in situations involving a smaller screen, such as a smartphone). The feedback loop allows the system to predetermine which lines should be exposed. A low level of continuous, dim illumination can be provided to ensure the user has a reasonable image to complete this task. Depending on the available processing power and flash offset controls, the continuously illuminated frames can be processed to detect specular reflections, and further used to change the flash offset or delay. Such operation allows the user additional margin when attempting to place their eye in the box while keeping the exposure levels to a minimum.
  • The biometric analysis systems adjust the portion of the frame that is illuminated such that the important portion of the image (e.g., a region of interest) is well-lit and the unimportant portions of the images are left dim. The biometric analysis systems can be used in imaging in the near infrared (NIR) spectrum using a short, intense, eye-safe NIR flash with an NIR filter to freeze motion and emphasize the flash relative to ambient light (including bright sun). The biometric analysis systems including the NIR flash with a rolling shutter camera can be used in, e.g., iris biometrics, or the like, in which cost reduction realized by using a rolling shutter instead of a global shutter is a primary driver in component selection.
  • The biometric analysis systems include an adaptive step in which dim illumination allows a vision system (e.g., a sensor in a rolling shutter camera including a processor) to detect the position of the desired object in the field-of-view of the camera. In some embodiments, the dim illumination remains on for the entire process of capturing an image, and reveals an under-illuminated frame. Dim illumination, although dimmer than illumination by the flash, is bright enough to perform an automated scene analysis, e.g., face finding, object identification, or the like, that can output a recommendation of a position in the frame that would benefit from flash illumination. For example, if a face appeared between 20% and 40% of the frame height measured from the bottom, a controller or processor can arrange a delay between the start of image writing and the flash trigger such that a stripe of illumination coincides with the desired object. If the object moves in the field-of-view, biometric analysis systems can function dynamically to track movement of the object and change the delay accordingly. In some embodiments, rather than or in addition to using dim illumination, the adaptive trigger module can be executed to search and find the region of interest by systematically sweeping an illuminated stripe down the field-of-view and stopped when a region of interest is detected. Thus, rather than using a single frame (as in dim illumination), the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest. The region of interest can be tracked by the adaptive trigger module after being detected. Such arrangements to locate and track the object and indicate to the adaptive trigger module changes in the position of the object permits moving objects to be accurately imaged by the biometric analysis systems.
  • The biometric analysis systems disclosed herein can be used for rolling shutter photography (e.g., NIR photography) of objects while outdoors or indoors. The biometric analysis systems can acquire NIR images with controlled illumination in full sun or in dim ambient lighting. An image of a desired object that fills less than a full frame, e.g., a person's face with the entire person in the field-of-view, can be captured by the biometric analysis systems. Due to the low cost, small size and availability of rolling shutter sensors, incorporation of the biometric analysis systems into cost-sensitive or size-sensitive platforms can benefit cost-sensitive or size-sensitive users. As one example, the biometric analysis systems can be used in cargo identification in which infrared dye for markings signify authenticity. As another example, a box marked with writing can be imaged with a small and inexpensive handheld infrared camera that would look for the marking surrounding an NIR watermark.
  • The flash trigger can be adjusted to fire in order to illuminate the markings, leaving the rest of the field-of-view above and below the mark of interest in the dark. Handheld devices, including smartphones, barcode readers, and other mass produced electronic devices, can incorporate the biometric analysis systems at a small and low cost. Traditionally, a rolling shutter camera would reduce the cost for such systems, but would require an extended flash to illuminate an entire scene to guarantee capturing the subject of interest. The exemplary biometric analysis systems include a rolling shutter camera with an adaptive image content-based trigger that issues a shorter flash of a near infrared light that allows for use of smaller batteries, provides for a longer battery life, provides for improved eye safety, and allows for longer illumination source lifetimes.
  • The biometric analysis systems disclosed herein can provide flash illumination for illuminating an object sufficiently to outshine the sun, and in some embodiments can use strobing flash illumination as is disclosed in, e.g., U.S. Pat. Nos. 7,542,628; 7,627,147; 7,657,127; 9,131,141; U.S. Patent Application Publication No. 2015/0098630; U.S. Patent Application Publication No. 2016/0014121; U.S. Patent Application Publication No. 2016/0012218; and U.S. Patent Application Publication No. 2016/0012292, which are incorporated by reference herein.
  • The exemplary biometric analysis systems use a rolling shutter camera and a flash illumination that is triggered adaptively to illuminate the region of interest of the scene or object. The adaptive trigger can be configured to follow or track the subject, and adjusts the flash timing to maintain the stripe of flash illumination on the subject of interest. In one embodiment, a face finder can be used as an image analytic to trigger the flash. In one embodiment, a unique identifier, such as a bar code or QR code, can be detected and illuminated by the stripe of flash illumination. In either case, the biometric analysis system can locate the object in a relatively dimly lit scene, and then initiates the repetitive flash that illuminates the object. Flash timing can be adjusted to maintain the object in the illuminated region of the field-of-view of the rolling shutter camera.
  • With respect to the delayed flash, an assumption can be made that the start of integration of the first row of a rolling shutter sensor coincides with time t=0. A further assumption can be made that a flash pulse is started at some time later, e.g., delayed by a time Δt. Rows that are read-out (e.g., integrate light) before At miss any light from the flash. Rows that are in the process of integrating light when the flash starts—some nearly finished, some just starting—receive fractions of the flash and form the ramp-up region of partial illumination. Pixels in rows that start integrating at or after the start of the flash and that finish integrating light by the end of the flash are illuminated by the full flash. By adjusting the delay time At, the rows of full flash illumination can be adjusted from the top of the image frame to the bottom to focus on a region of interest located within the frame.
  • In particular, the biometric analysis systems can be used to detect the region of interest within the frame of the rolling shutter camera, and intentionally delay the flash illumination in a coordinated manner to ensure that the region of interest is illuminated by the flash during exposure and read-out of the corresponding rows. Proper adjustment of the delay time can cause the stripe of flash illumination due to the short flash pulse to brighten the region of interest. In some embodiments, the stripe can scan or search the image in successive frames of a photo or video stream. Such technique can be performed by systematically sweeping an illuminated stripe down the field-of-view and stopping when a region of interest is detected. Thus, rather than using a single frame (as in dim illumination), the illuminated stripe can be a fraction in width of the full frame and is advanced across the field-of-view, thereby using multiple frames to detect the region of interest. The region of interest can be tracked by the adaptive trigger module after being detected.
  • In some embodiments, a pre-flash (e.g., a single long, low intensity flash that illuminates the full frame and that is both eye-safe and within the operating range of the illumination hardware) can provide a single, dim image of the scene. Automated image analysis of the full scene that can tolerate the low light can detect and output a region of interest that is the width of the image and the height of the flash illumination stripe. With the region of interest defined, an automated delay signal can set the illumination stripe to the optimal position to illuminate the region of interest in succeeding frames of a photo or video stream using the full flash brightness over the narrowed region of interest. If the desired feature of the object begins to move out of the illumination stripe (as determined by video analysis performed at the frame rate), the delay can adapt using negative feedback to maintain the position of the desired feature centered within the well-lit region of the image.
  • In some embodiments, the biometric analysis systems can be used to capture NIR images of an object that is bathed in sunlight. While a fraction of the solar spectrum covers the NIR wavelengths, it can be assumed that filtered solar NIR irradiance (e.g., from 800 nm to 900 nm) is insufficient to illuminate a video frame that is adequately short in duration to suppress motion blur. It can be further assumed that using a long flash that covers the entire rolling shutter frame period is impractical, even with a short exposure time, due to hardware limitations or eye safety concerns. A short NIR flash pulse that can only illuminate a portion of the frame with an adaptive delay time to ensure that the region of interest is illuminated can therefore be used.
  • The biometric analysis systems initially detect and identify a region of interest of an object, such as the eyes of a subject. Eye finding can be accomplished using a number of different processes, such as full frame NIR illumination with a relatively dim, eye-safe (and hardware-safe) long NIR flash pulse; partial frame illumination due to a short, intense pulse that is automatically swept frame-by-frame over the full field-of-view until the eyes appear in the search; and/or image acquisition by a color (or other spectrum) auxiliary camera that produces a full frame image allowing eye-finding or face-finding, and then translation from the auxiliary camera frame position to a NIR camera position.
  • After the eyes are detected and their position located in the NIR camera frame, the biometric analysis systems automatically set the flash pulse delay to illuminate a stripe across the frame including the eyes (e.g., the region of interest). The delay that sets the vertical position of the illumination stripe can be automatically adjusted to maintain the eyes in substantially the center of the frame. While the portions of the frame outside of the illumination stripe remain unusably dark, the important information about the eyes remains well lit.
  • The competing solar illumination may dimly illuminate the rest of the frame. This is not an issue as long as the flash illumination stripe remains resolved and does not saturate due to the sum of the in-band solar irradiance and the applied NIR light. If the illumination stripe nears the saturation level, the intensity of the applied flash can be throttled back. It is preferable to decrease the NIR illumination irradiance on the object rather than reducing the exposure or pulse duration, since the later strategies reduce the portions of the image that are illuminated.
  • Any spectrum of light can be used as long as the camera uses a rolling shutter with an in-band flash illuminator. For example, visible light, NIR, short-wave infrared (SWIR), long-wave infrared (LWIR), ultraviolet (UV), or multi-spectral cameras with rolling shutters that use flash illumination. Any process of finding the region of interest can be used as long as the process returns a coordinate to the biometric analysis system that indicates the time delay between the start of frame and the start of the flash. For example, use of an eye finder that seeks corneal reflections of the flash could be useful if the region of interest contains one or two eyes of the subject. As a further example, a color face camera could be useful to find the eyes of a subject in iris recognition; but if nose-recognition were used, the region of interest would contain the subject's nose. If the same camera were used to find the region of interest and to acquire the information from the region of interest, a different illumination protocol could be used for each function. For example, a long, relatively dim pulse could illuminate the entire frame, followed by a short, bright pulse to illuminate the region of interest with better fidelity (e.g., improved signal-to-noise ratio). Any process that provides the information needed to set the delay time would also be useful.
  • In some embodiments, feedback can be used to maintain the region of interest in the illumination stripe. For example, negative feedback can be used to correct the position of the region of interest within the illumination stripe. If the object of interest begins to move or slip to the top or bottom of the region of interest illuminated by the flash, feedback can be used to correct the delay time. For example, if the stripe illuminating the region of interest needs to move up in the frame to follow the object of interest, the delay would need to be corrected in one direction. If the object of interest moves down in the frame, the sign of the delay correction would need to be in the opposite direction. If the delay reaches the minimum or maximum value or, in other words, the object of interest begins to leave the camera field-of-view, the controller or processing device automatically reorients the camera to maintain the object in the field-of-view. In some embodiments, a feedback signal (e.g., an audio alarm, a visual alarm, combinations thereof, or the like) can provide feedback to an operator regarding the object leaving the field-of-view.
  • The width of the flash illumination stripe relative to the frame height can be set by practical considerations. For example, rolling shutter cameras with very short read-out times generally have wider flash illumination stripes, all other things being equal. The longest flash pulse can be dictated by practical issues, including available power to the illumination source, operating conditions in power, and temperature of the illumination source. For example, it is known that flashing an illumination source too often can damage it, and reduces the safety of human eyes in and around the illumination source. A flash illumination stripe that is narrow relative to the frame height generally should be more accurately timed than a flash illumination stripe that is wider and that can tolerate more motion of the object of interest within the flash illumination stripe.
  • With reference to FIG. 8, a schematic block diagram of an exemplary biometric analysis system 100 (hereinafter “system 100”) is provided. The system 100 is configured to detect and identify a region of interest of an object, such as the eyes of a person, and generates a time delay of the flash illumination such that the flash coincides with the region of interest. A rolling shutter camera can therefore be used for capturing one or more images for biometric analysis with the delayed flash illumination ensuring that the region of interest will be properly illuminated in the captured images. The system 100 is configured to track the region of interest within the frame or field-of-view of the rolling shutter camera, and adjusts the time delay of the flash pulse appropriately based on movement of the region of interest in and out of the frame.
  • The system 100 generally includes one or more illumination sources 102. The illumination sources 102 are configured to initially provide dim illumination 104 to a scene including an object to detect and identify the object (or the region of interest of the object). In some embodiments, the dim illumination 104 can be separate from the system 100 and can be provided by, e.g., ambient light, any other illumination source, sunlight, combinations thereof, or the like. The illumination sources 102 are further configured to provide flash illumination 106 (e.g., a pulse of flash illumination) to illuminate the region of interest during exposure of the rows corresponding to the region of interest, thereby ensuring that the region of interest will be illuminated in the captured image. In some embodiments, a single illumination source 102 can provide both the dim illumination 104 and the flash illumination 106. In some embodiments, separate illumination sources 102 can provide the dim illumination 104 and the flash illumination 106. In one embodiment, the illumination sources 102 can be NIR illumination sources. It should be understood that the dim illumination 104 is less bright than the flash illumination 106, and provides sufficient ambient light for the system 100 to detect objects in a scene as viewed by the rolling shutter camera.
  • The system 100 includes a rolling shutter camera 108 including a frame 110 with a field-of-view. The rolling shutter camera 108 is configured to capture one or more images (e.g., still frame images, video images, combinations thereof, or the like). The rolling shutter camera 108 is configured to initially capture an image of the scene and the object in the scene under dim illumination 104 conditions. The system 100 includes an adaptive trigger module 112 configured to be executed by a processing device 114 having a processor 116 (e.g., a controller). Although illustrated as separate components from the rolling shutter camera 108, it should be understood that in some embodiments, the rolling shutter camera 108 can include the illumination sources 102, the adaptive trigger module 112, the processing device 114, or the like. In some embodiments, the processing device 114 can be part of a central computing system 118. When executed, the adaptive trigger module 112 can analyze the captured image of the scene illuminated by the dim illumination 104 and detects the region of interest 120 within the scene. For example, the adaptive trigger module 112 can detect and identify the region of interest 120 as one of more eyes of the user, and stores the rows of the frame 110 corresponding to the region of interest 120 in a database 122 (e.g., a local or remote database).
  • Based on the rows of the frame 110 corresponding to the region of interest 120, the adaptive trigger module 112 arranges a flash pulse delay 124 between the start of image writing by the rolling shutter camera 108 and a trigger of the flash illumination 106 such that the flash illumination 108 coincides with the identified region of interest 120. For example, if the eyes of the user are the region of interest 120 corresponding with rows 400-600 of the frame 110, the adaptive trigger module 112 arranges the flash pulse delay 124 such that the flash illumination 106 is triggered to illuminate the object during rows 400-600. Thus, rather than triggering the flash illumination 106 at the same time as the start of image writing by the rolling shutter camera 108, the image writing can start before triggering of the flash illumination 106 and the flash illumination 106 is only triggered to illuminate the region of interest 120. Such flash delay 124 ensures that the region of interest 120 is illuminated in the images 126 captured by the rolling shutter camera 108. The resulting image 126 generally includes a stripe of flash illumination 106 extending across the image 126 and over the eyes of the user (see, e.g., FIG. 9). Depending on the length of the exposure time, the height of the stripe of flash illumination 106 can be greater or smaller relative to the top and bottom of the image 126.
  • In some embodiments, the adaptive trigger module 112 can include a face finder algorithm that locates the face and/or features of the face in the scene. The face finder algorithm can therefore locate the face identifies one or more eyes of the face as the region of interest 120. Although discussed herein as eyes of a user, it should be understood that the region of interest 120 can be a unique identifier associated with a physical item that can be detected by an identifier finder algorithm of the adaptive trigger module 112. For example, the identifier finder algorithm of the adaptive trigger module 112 can be used to detect, e.g., a barcode, a QR code, or the like, on a physical item, identifies such unique identifier as the region of interest 120, and adjusts the flash delay 124 to ensure that the unique identifier is illuminated during capture of the image 126.
  • The adaptive trigger module 112 can be configured to track movement of the object within the field-of-view or frame 110 of the rolling shutter camera 108. In some embodiments, the adaptive trigger module 112 can be configured to continuously scan the scene in the frame 110 to detect the position of the region of interest 120. If the user moves such that the region of interest 120 in the frame 110 is not centered or changes relative to the previously detected rows for the region of interest 120, the adaptive trigger module 112 can determine the new rows corresponding to the region of interest 120, stores the new rows in the database 122, and adjusts the flash delay 124 to coincide with the new rows for the region of interest 120. Thus, an object moving in the frame 110 of the rolling shutter camera 108 can be tracked by the adaptive trigger module 112 to allow for the region of interest 120 of a moving object to be properly illuminated by the flash illumination 106.
  • The system 100 includes a feedback module 128 configured to be executed by the processing device 114 and/or the central computing system 118. In some embodiments, the rolling shutter camera 108 can include the feedback module 128. When executed, the feedback module 128 receives as input the images 126 captured by the rolling shutter camera 108 and analyzes the region of interest 120 captured in the image 126 to determine if the region of interest 120 is illuminated by the stripe of flash illumination 106. If the region of interest 120 is illuminated by the stripe of flash illumination 106, the feedback module 128 electronically outputs such findings to the adaptive trigger module 112 and no changes are made in the flash delay 124. If the region of interest 120 is not illuminated (or only partially illuminated) by the stripe of flash illumination 106, the feedback module 128 electronically outputs such findings to the adaptive trigger module 112. Based on the feedback from the feedback module 128, the adaptive trigger module 112 analyzes the frame 110 to locate the region of interest 120, determines the new rows associated with the region of interest 120, and adjusts the flash pulse delay 124 to trigger the flash illumination 106 at the appropriate rows coinciding with the region of interest 120. The feedback loop can be continuously performed to ensure the region of interest 120 is always illuminated by the stripe of flash illumination 106.
  • The system 100 can include a user interface 130 with a graphical user interface (GUI) 132 for receiving input from a user and for outputting information to the user. For example, the user interface 130 can be used to initiate the process performed by the system 100. As a further example, the user interface 130 can output an audio and/or visual alarm to the user when the region of interest 120 is no longer centered in the frame 110. The system 100 can include a communication interface 134 configured to provide a means for electronic transmission between components of the system 100, e.g., the illumination sources 102, the processing device 114, the rolling shutter camera 108, the adaptive trigger module 112, the database 122, the user interface 130, the central computing system 118, and feedback module 128, or the like.
  • FIG. 10 is a diagrammatic view of an exemplary biometric analysis system 200 (hereinafter “system 200”) of the present disclosure. The system 200 includes an illumination source 202 that provides dim illumination to the subject 204 in the scene. In some embodiments, the illumination source 202 can provide 1 ms NIR pulses to the scene. In some embodiments, the illumination source 202 can be an NIR light source not connected to the camera 208. The illumination source 202 can be positioned a distance 206 from the subject 204. In some embodiments, the illumination source 202 can also be configured to provide the flash illumination to the subject 204.
  • The system 200 includes a camera 208 including a rolling shutter (e.g., a rolling shutter camera, such as a smartphone camera, or the like) configured to capture one or more images of the subject. The camera 208 can have a field-of-view represented by area 216, including the region of interest of the subject 204. In some embodiments, the camera 208 includes an illumination source 210 (e.g., an integrated NIR light source) configured to provide flash illumination to a region of interest (e.g., the eyes of the subject 204). As an example, area 212 indicates the coverage of the NIR flash illumination provided by the illumination source 210, with the region of interest on the subject 204 being within the area 212. The camera 208 can be spaced by a distance 214 from the subject 204. In FIG. 10, the illumination source 202 appears separate from camera 202, however, one skilled in the art shall appreciate that, in some embodiments, the illumination source 202 can be incorporated into the camera 208 assembly such that the camera 208 assembly is capable of providing both the dim illumination and flash illumination to the subject 204.
  • FIG. 11 is a diagrammatic view of a frame action 300 of a rolling shutter camera of the exemplary biometric analysis systems disclosed herein. The frame action 300 includes a total frame time 302. Ambient light 304 having a low illumination level 306 can be provided for all or part of the total frame time 302 by a dim illumination source and/or sunlight. The total camera exposure time 308 indicates when the rolling shutter camera begins and ends image writing and read-out.
  • Based on the determination of the region of interest and the corresponding rows of the frame, a pulse of flash illumination 310 (e.g., NIR from a light-emitting diode (LED) illumination source) is triggered to coincide with the rows of the region of interest. The pulse of flash illumination 310 has a time or pulse width 312 that is significantly smaller than the camera exposure time 308, and an illumination level 314 that is brighter or higher than the level 306 of the ambient light 304. The region of interest is thereby illuminated by the flash illumination 310 while the remaining rows of the captured image are only illuminated by the ambient light 304.
  • In some embodiments, a 1280×960 sensor can be used with the rolling shutter camera. In some embodiments, a 120 fps VGA (640×480) sensor can be used with the rolling shutter camera to allow for full frame exposure in 4 ms, and can be synchronized with an approximately 4 ms LED flash every 8th frame. The VGA output can provide (unscaled) 15 fps feedback to the user facing the monitor of the system. The 4 ms can provide sufficient flash pulse while allowing for cost effective use of an LED and rolling shutter accommodation, and allowing for eye safety.
  • In some embodiments, a bright light (e.g., approximately 850 nm LED, two LEDs) can be used for the flash pulse. The systems can therefore have high pulsed current capabilities. The rolling shutter camera can have a high quality camera lens and NIR filtering. In some embodiments, the NIR filter can be tighter than 100 nm around the LED center wavelength. In some embodiments, the system can include optics with a modulation transfer function (MTF) optimized for the spatial frequencies that enter the system. In some embodiments, the system can include preprocessing of iris images for a particular spectrum of spatial frequencies (e.g., low, medium, high, or the like).
  • FIG. 12 is a flowchart illustrating an exemplary process 400 of implementing the biometric analysis systems disclosed herein. To begin, at step 402, a scene and an object in the scene are illuminated with dim illumination from one or more illumination sources. At step 404, the scene is analyzed with an adaptive trigger module to detect the object (or a region of interest associated with the object) in the scene during dim illumination. At step 406, a position in a frame of a rolling shutter camera is determined with the adaptive trigger module, the position coinciding with the detected object (or region of interest associated with the object) in the scene.
  • At step 408, a delay is arranged by the adaptive trigger module between a start of image writing by the rolling shutter camera and a trigger of the illumination sources such that a stripe of flash illumination provided by the illumination sources coincides with the detected object (or region of interest associated with the object) in the scene. At step 410, movement of the object is tracked within a field-of-view of the rolling shutter camera with the adaptive trigger module. At step 412, the delay between the start of image writing by the rolling shutter camera and the trigger of the illumination sources is modified with the adaptive trigger module based on detected movement of the object (or region of interest associated with the object) out of the field-of-view.
  • FIG. 13 is a flowchart illustrating an exemplary process 500 of implementing the biometric analysis systems disclosed herein. To begin, at step 502, the adaptive trigger module can set the flash delay at 0 ms, the flash duration equal to the time of the full frame, and the flash brightness to low. At step 504, a full frame image can be acquired with the rolling shutter camera under dim illumination. At step 506, the adaptive trigger module analyzes the image captured at step 504 and locates the region of interest in the image.
  • At step 508, the adaptive trigger module sets the flash delay to a value greater than 0 ms, the flash duration to short (e.g., a pulse), and the flash brightness to high (e.g., higher than the dim illumination) to illuminate the region of interest. At step 510, the adaptive trigger module measures the position of the object of interest within the region of interest (e.g., eyes of the user located on the user's face). At step 512, if the object of interest is within the region of interest, the process 500 continues a loop to constantly (or at predetermined frequencies of time) ensure that the object is maintained within the region of interest. At step 514, if the object of interest is near an edge of the frame or region of interest, the process 500 performs step 516 to correct the flash delay such that the object is centered within the region of interest. The region of interest can thereby be maintained in the flash illumination stripe despite relative motion of the object and/or the camera. In some embodiments, the process can include a loop that analyzes the captures images to control the intensity of the flash illumination, thereby avoiding saturation of the object.
  • FIG. 14 is a block diagram of a computing device 600 in accordance with exemplary embodiments of the present disclosure. The computing device 600 includes one or more non-transitory computer-readable media for storing one or more computer-executable instructions or software for implementing exemplary embodiments. The non-transitory computer-readable media may include, but are not limited to, one or more types of hardware memory, non-transitory tangible media (for example, one or more magnetic storage disks, one or more optical disks, one or more flash drives), and the like. For example, memory 606 included in the computing device 600 may store computer-readable and computer-executable instructions or software for implementing exemplary embodiments of the present disclosure (e.g., instructions for operating the illumination sources, instructions for operating the processing device, instructions for operating the rolling shutter camera, instructions for operating the adaptive trigger module, instructions for operating the communication interface, instructions for operating the user interface, instructions for operating the central computing system, instructions for operating the feedback module, combinations thereof, or the like). The computing device 600 also includes configurable and/or programmable processor 602 and associated core 604, and optionally, one or more additional configurable and/or programmable processor(s) 602′ and associated core(s) 604′ (for example, in the case of computer systems having multiple processors/cores), for executing computer-readable and computer-executable instructions or software stored in the memory 606 and other programs for controlling system hardware. Processor 602 and processor(s) 602′ may each be a single core processor or multiple core (604 and 604′) processor.
  • Virtualization may be employed in the computing device 600 so that infrastructure and resources in the computing device 600 may be shared dynamically. A virtual machine 614 may be provided to handle a process running on multiple processors so that the process appears to be using only one computing resource rather than multiple computing resources. Multiple virtual machines may also be used with one processor. Memory 606 may include a computer system memory or random access memory, such as DRAM, SRAM, EDO RAM, and the like. Memory 606 may include other types of memory as well, or combinations thereof.
  • A user may interact with the computing device 600 through a visual display device 618 (e.g., a personal computer, a mobile smart device, or the like), such as a computer monitor, which may display one or more user interfaces 620 (e.g., a graphical user interface) that may be provided in accordance with exemplary embodiments. The computing device 600 may include other I/O devices for receiving input from a user, for example, a camera, a sensor, a keyboard or any suitable multi-point touch interface 608, a pointing device 610 (e.g., a mouse). The keyboard 608 and the pointing device 610 may be coupled to the visual display device 618. The computing device 600 may include other suitable conventional I/O peripherals.
  • The computing device 600 may also include one or more storage devices 624, such as a hard-drive, CD-ROM, eMMC (MultiMediaCard), SD (secure digital) card, flash drive, non-volatile storage media, or other computer readable media, for storing data and computer-readable instructions and/or software that implement exemplary embodiments of the biometric analysis systems described herein. Exemplary storage device 624 may also store one or more databases 626 for storing any suitable information required to implement exemplary embodiments. For example, exemplary storage device 624 can store one or more databases 626 for storing information, such as data relating to captured images 126 under dim illumination 104 and flash illumination 106, regions of interest 120, flash delay 124, combinations thereof, or the like, and computer-readable instructions and/or software that implement exemplary embodiments described herein. The databases 626 may be updated by manually or automatically at any suitable time to add, delete, and/or update one or more items in the databases.
  • The computing device 600 can include a network interface 612 configured to interface via one or more network devices 622 with one or more networks, for example, Local Area Network (LAN), Wide Area Network (WAN) or the Internet through a variety of connections including, but not limited to, standard telephone lines, LAN or WAN links (for example, 802.11, T1, T3, 56 kb, X.25), broadband connections (for example, ISDN, Frame Relay, ATM), wireless connections, controller area network (CAN), or some combination of any or all of the above. The network interface 612 may include a built-in network adapter, network interface card, PCMCIA network card, PCI/PCIe network adapter, SD adapter, Bluetooth adapter, card bus network adapter, wireless network adapter, USB network adapter, modem or any other device suitable for interfacing the computing device 600 to any type of network capable of communication and performing the operations described herein. Moreover, the computing device 600 may be any computer system, such as a workstation, desktop computer, server, laptop, handheld computer, tablet computer (e.g., the tablet computer), mobile computing or communication device (e.g., the smart phone communication device), an embedded computing platform, or other form of computing or telecommunications device that is capable of communication and that has sufficient processor power and memory capacity to perform the operations described herein.
  • The computing device 600 may run any operating system 616, such as any of the versions of the Microsoft® Windows® operating systems, the different releases of the Unix and Linux operating systems, any version of the MacOS® for Macintosh computers, any embedded operating system, any real-time operating system, any open source operating system, any proprietary operating system, or any other operating system capable of running on the computing device and performing the operations described herein. In exemplary embodiments, the operating system 616 may be run in native mode or emulated mode. In an exemplary embodiment, the operating system 616 may be run on one or more cloud machine instances.
  • FIG. 15 is a block diagram of an exemplary biometric analysis system environment 700 in accordance with exemplary embodiments of the present disclosure. The environment 700 can include servers 702, 704 configured to be in communication with one or more illumination sources 706, one or more rolling shutter cameras 708, one or more adaptive trigger modules 710, a feedback module 712, a user interface 714, and a central computing system 716 via a communication platform 922, which can be any network over which information can be transmitted between devices communicatively coupled to the network. For example, the communication platform 722 can be the Internet, Intranet, virtual private network (VPN), wide area network (WAN), local area network (LAN), and the like. In some embodiments, the communication platform 722 can be part of a cloud environment.
  • The environment 700 can include repositories or databases 718, 720, which can be in communication with the servers 702, 904, as well as the one or more illumination sources 706, one or more rolling shutter cameras 708, one or more adaptive trigger modules 710, the feedback module 712, the user interface 714, and the central computing system 716, via the communications platform 722.
  • In exemplary embodiments, the servers 702, 704, one or more illumination sources 706, one or more rolling shutter cameras 708, one or more adaptive trigger modules 710, the feedback module 712, the user interface 714, and the central computing system 716 can be implemented as computing devices (e.g., computing device 600). Those skilled in the art will recognize that the databases 718, 720 can be incorporated into one or more of the servers 702, 704. In some embodiments, the database 718 can store data relating to captured images, regions of interest 120, flash delay 124, combinations thereof, or the like, can be distributed over multiple databases 718, 720.
  • While exemplary embodiments have been described herein, it is expressly noted that these embodiments should not be construed as limiting, but rather that additions and modifications to what is expressly described herein also are included within the scope of the invention. Moreover, it is to be understood that the features of the various embodiments described herein are not mutually exclusive and can exist in various combinations and permutations, even if such combinations or permutations are not made express herein, without departing from the spirit and scope of the invention.

Claims (21)

1. A biometric analysis system, comprising:
one or more illumination sources configured to provide dim illumination to a scene including an object and configured to provide flash illumination to the object in the scene;
a rolling shutter camera configured to capture one or more images; and
an adaptive trigger module configured to (i) analyze the scene to detect the object in the scene during dim illumination of the scene, (ii) determine a position in a frame of the rolling shutter camera that coincides with the detected object in the scene, and (iii) arrange a delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the detected object in the scene.
2. The biometric analysis system of claim 1, wherein the adaptive trigger module is configured to track movement of the object within a field-of-view of the rolling shutter camera.
3. The biometric analysis system of claim 2, wherein the adaptive trigger module is configured to modify the delay between the start of image writing by the rolling shutter camera and the trigger of the one or more illumination sources based on detected movement of the object within the field-of-view.
4. The biometric analysis system of claim 1, wherein the adaptive trigger module is configured to detect a region of interest of the object and arranges the delay such that the stripe of flash illumination coincides with the detected region of interest of the object.
5. The biometric analysis system of claim 4, wherein the region of interest of the object includes eyes of a person.
6. The biometric analysis system of claim 1, wherein the object is a person, and wherein the adaptive trigger module comprises a face finder configured to detect a face of the person.
7. The biometric analysis system of claim 1, wherein the object is a physical item, and the adaptive trigger module comprises an identifier finder configured to detect a unique identifier associated with the physical item.
8. The biometric analysis system of claim 7, wherein the unique identifier is a barcode or a quick response (QR) code.
9. The biometric analysis system of claim 1, wherein the one or more illumination sources are configured to provide the flash illumination as a synchronized pulse of flash illumination.
10. The biometric analysis system of claim 1, wherein the flash illumination provided by the one or more illumination sources is brighter than the dim illumination provided by the one or more illumination sources.
11. The biometric analysis system of claim 1, wherein the one or more illumination sources comprise a first illumination source configured to provide the dim illumination and a second illumination source configured to provide the flash illumination.
12. The biometric analysis system of claim 1, wherein the one or more illumination sources are near infrared illumination sources.
13. The biometric analysis system of claim 1, wherein the one or more illumination sources are ambient light.
14. The biometric analysis system of claim 1, wherein the adaptive trigger module is configured to sweep an illuminated stripe down the frame as the rolling shutter camera captures the one or more images, analyzes an illuminated section of the one or more images to identify a region of interest in the illuminated section, and stops sweeping of the illuminated stripe when the region of interest is identified.
15. The biometric analysis system of claim 1 provided as a smartphone having said one or more illumination sources, said rolling shutter camera, and said adaptive trigger module.
16. A biometric analysis system, comprising:
one or more illumination sources configured to provide dim illumination to a scene including a subject and configured to provide flash illumination to the subject in the scene;
a rolling shutter camera configured to capture one or more images; and
an adaptive trigger module configured to (i) analyze the scene to detect eyes of the subject in the scene during dim illumination of the scene, (ii) identify the eyes of the subject as a region of interest, (iii) determine a position in a frame of the rolling shutter camera that coincides with the identified region of interest, and (iv) arrange a flash pulse delay between a start of image writing by the rolling shutter camera and a trigger of the one or more illumination sources such that a stripe of the flash illumination coincides with the identified region of interest.
17. The biometric analysis system of claim 16, wherein the flash pulse delay ensures that the illuminated region of interest is maintained in a center of the frame of the rolling shutter camera.
18. The biometric analysis system of claim 16, comprising a feedback module configured to analyze a captured image of the region of interest and determine if the region of interest is illuminated by the stripe of the flash illumination.
19. The biometric analysis system of claim 18, wherein the adaptive trigger module is configured to adjust the flash pulse delay based on the determination of the feedback module to ensure that the region of interest is illuminated by the stripe of the flash illumination.
20. The biometric analysis system of claim 16 provided as a smartphone having said one or more illumination sources, said rolling shutter camera, and said adaptive trigger module.
21-24. (canceled)
US16/531,488 2016-03-31 2019-08-05 Systems and methods of biometric analysis with adaptive trigger Abandoned US20190356827A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/531,488 US20190356827A1 (en) 2016-03-31 2019-08-05 Systems and methods of biometric analysis with adaptive trigger

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201662316347P 2016-03-31 2016-03-31
US15/471,131 US10373008B2 (en) 2016-03-31 2017-03-28 Systems and methods of biometric analysis with adaptive trigger
US16/531,488 US20190356827A1 (en) 2016-03-31 2019-08-05 Systems and methods of biometric analysis with adaptive trigger

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US15/471,131 Continuation US10373008B2 (en) 2016-03-31 2017-03-28 Systems and methods of biometric analysis with adaptive trigger

Publications (1)

Publication Number Publication Date
US20190356827A1 true US20190356827A1 (en) 2019-11-21

Family

ID=59961086

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/471,131 Active 2037-07-20 US10373008B2 (en) 2016-03-31 2017-03-28 Systems and methods of biometric analysis with adaptive trigger
US16/531,488 Abandoned US20190356827A1 (en) 2016-03-31 2019-08-05 Systems and methods of biometric analysis with adaptive trigger

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/471,131 Active 2037-07-20 US10373008B2 (en) 2016-03-31 2017-03-28 Systems and methods of biometric analysis with adaptive trigger

Country Status (2)

Country Link
US (2) US10373008B2 (en)
WO (1) WO2017172695A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112040139A (en) * 2020-09-09 2020-12-04 绍兴埃瓦科技有限公司 Light supplementing method for camera imaging
US11816198B2 (en) 2021-04-06 2023-11-14 Bank Of America Corporation Systems and methods for geolocation security using biometric analysis

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10425814B2 (en) 2014-09-24 2019-09-24 Princeton Identity, Inc. Control of wireless communication device capability in a mobile device with a biometric key
EP3227816A4 (en) 2014-12-03 2018-07-04 Princeton Identity, Inc. System and method for mobile device biometric add-on
JP2019506694A (en) 2016-01-12 2019-03-07 プリンストン・アイデンティティー・インコーポレーテッド Biometric analysis system and method
WO2017173228A1 (en) 2016-03-31 2017-10-05 Princeton Identity, Inc. Biometric enrollment systems and methods
US10373008B2 (en) 2016-03-31 2019-08-06 Princeton Identity, Inc. Systems and methods of biometric analysis with adaptive trigger
JP2017211891A (en) * 2016-05-27 2017-11-30 ソニー株式会社 Information processing device, information processing method, and recording medium
US10607096B2 (en) 2017-04-04 2020-03-31 Princeton Identity, Inc. Z-dimension user feedback biometric system
JP2020529073A (en) 2017-07-26 2020-10-01 プリンストン・アイデンティティー・インコーポレーテッド Biosecurity systems and methods
CN109474788B (en) * 2017-09-08 2021-11-30 超威半导体公司 Illumination control techniques for color and IR camera sensors
US10691969B2 (en) * 2017-11-06 2020-06-23 EagleSens Systems Corporation Asynchronous object ROI detection in video mode
WO2019111785A1 (en) * 2017-12-08 2019-06-13 キヤノン株式会社 Identification device and sorting system
CN108197546B (en) * 2017-12-26 2020-12-11 深圳市友信长丰科技有限公司 Illumination processing method and device in face recognition, computer equipment and storage medium
NL2021044B1 (en) * 2018-06-01 2019-12-10 Eldolab Holding Bv LED driver and method of operating a camera
US10728447B2 (en) * 2018-10-26 2020-07-28 Alibaba Group Holding Limited Capturing images using sub-frame illumination
CN114009002B (en) * 2018-11-28 2022-11-22 亮锐有限责任公司 Method for obtaining digital image
EP3820134A1 (en) * 2019-11-06 2021-05-12 Koninklijke Philips N.V. A system for performing ambient light image correction
US11941897B2 (en) * 2021-12-09 2024-03-26 Fotonation Limited Vehicle occupant monitoring system including an image acquisition device with a rolling shutter image sensor

Family Cites Families (215)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3852592A (en) 1973-06-07 1974-12-03 Stanley Works Automatic door operator
US3993888A (en) 1974-10-29 1976-11-23 Calspan Corporation Scanning line filter
US4109237A (en) 1977-01-17 1978-08-22 Hill Robert B Apparatus and method for identifying individuals through their retinal vasculature patterns
US4641349A (en) 1985-02-20 1987-02-03 Leonard Flom Iris recognition system
US5291560A (en) 1991-07-15 1994-03-01 Iri Scan Incorporated Biometric personal identification system based on iris analysis
US6850252B1 (en) 1999-10-05 2005-02-01 Steven M. Hoffberg Intelligent electronic appliance system and method
US5337104A (en) 1992-11-09 1994-08-09 Smith Neil D Apparatus and method for redirecting camera flash light so as to avoid red-eye
JP2882266B2 (en) 1993-12-28 1999-04-12 株式会社日立製作所 Signal transmission device and circuit block
US5835616A (en) 1994-02-18 1998-11-10 University Of Central Florida Face detection using templates
US5481622A (en) 1994-03-01 1996-01-02 Rensselaer Polytechnic Institute Eye tracking apparatus and method employing grayscale threshold values
US6714665B1 (en) 1994-09-02 2004-03-30 Sarnoff Corporation Fully automated iris recognition system utilizing wide and narrow fields of view
US5572596A (en) 1994-09-02 1996-11-05 David Sarnoff Research Center, Inc. Automated, non-invasive iris recognition system and method
US5833188A (en) 1994-12-20 1998-11-10 Twofish Unlimited Accessory mounting apparatus
WO1997014873A1 (en) 1995-10-16 1997-04-24 Valentin Semenovich Gorelykh Method of converting energy and a device for applying the said method
EP0865637A4 (en) 1995-12-04 1999-08-18 Sarnoff David Res Center Wide field of view/narrow field of view recognition system and method
US6226016B1 (en) 1996-02-05 2001-05-01 Seiko Epson Corporation Display apparatus and method capable of rotating an image by 180 degrees
US5987459A (en) 1996-03-15 1999-11-16 Regents Of The University Of Minnesota Image and document management system for content-based retrieval
US6320610B1 (en) 1998-12-31 2001-11-20 Sensar, Inc. Compact imaging device incorporating rotatably mounted cameras
JP3436293B2 (en) 1996-07-25 2003-08-11 沖電気工業株式会社 Animal individual identification device and individual identification system
US5933515A (en) 1996-07-25 1999-08-03 California Institute Of Technology User identification through sequential input of fingerprints
GB2315858A (en) 1996-08-01 1998-02-11 Sharp Kk System for eye detection and gaze direction determination
AU727389B2 (en) 1996-08-25 2000-12-14 Sarnoff Corporation, The Apparatus for the iris acquiring images
US6123733A (en) 1996-11-27 2000-09-26 Voxel, Inc. Method and apparatus for rapidly evaluating digital data processing parameters
US6215891B1 (en) 1997-03-26 2001-04-10 Oki Electric Industry Co., Ltd. Eye image recognition method eye image selection method and system therefor
US6144754A (en) 1997-03-28 2000-11-07 Oki Electric Industry Co., Ltd. Method and apparatus for identifying individuals
US6204858B1 (en) 1997-05-30 2001-03-20 Adobe Systems Incorporated System and method for adjusting color data of pixels in a digital image
US6119096A (en) 1997-07-31 2000-09-12 Eyeticket Corporation System and method for aircraft passenger check-in and boarding using iris recognition
US6252976B1 (en) 1997-08-29 2001-06-26 Eastman Kodak Company Computer program product for redeye detection
EP0910986A1 (en) 1997-10-24 1999-04-28 BRITISH TELECOMMUNICATIONS public limited company Imaging apparatus
US6055322A (en) 1997-12-01 2000-04-25 Sensor, Inc. Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination
US5953440A (en) 1997-12-02 1999-09-14 Sensar, Inc. Method of measuring the focus of close-up images of eyes
EP0934986A1 (en) 1997-12-17 1999-08-11 Roche Diagnostics GmbH Dye-polysaccharide or -cyclosaccharide conjugates and their use as diagnostic agent
US6421462B1 (en) 1998-02-06 2002-07-16 Compaq Computer Corporation Technique for differencing an image
US6301370B1 (en) 1998-04-13 2001-10-09 Eyematic Interfaces, Inc. Face recognition from video images
US5966197A (en) 1998-04-21 1999-10-12 Visx, Incorporated Linear array eye tracker
JP3315648B2 (en) 1998-07-17 2002-08-19 沖電気工業株式会社 Iris code generation device and iris recognition system
JP3610234B2 (en) 1998-07-17 2005-01-12 株式会社メディア・テクノロジー Iris information acquisition device and iris identification device
US6424727B1 (en) 1998-11-25 2002-07-23 Iridian Technologies, Inc. System and method of animal identification and animal transaction authorization using iris patterns
JP2000189403A (en) 1998-12-25 2000-07-11 Oki Electric Ind Co Ltd Iris region extraction and individual identifying device
US6247813B1 (en) 1999-04-09 2001-06-19 Iritech, Inc. Iris identification system and method of identifying a person through iris recognition
JP3636927B2 (en) 1999-05-18 2005-04-06 三菱電機株式会社 Face image processing device
US6433326B1 (en) 1999-07-14 2002-08-13 Sarnoff Corporation CMOS/CCD line transfer imager with low dark current
CN1361894A (en) 1999-07-22 2002-07-31 瑞士电信流动电话公司 Transaction method and suitable devices therefor
US6525303B1 (en) 1999-09-23 2003-02-25 Mitutoyo Corporation Control systems and methods for synchronizing diffuse illumination with a camera
US6836554B1 (en) 2000-06-16 2004-12-28 International Business Machines Corporation System and method for distorting a biometric for transactions with enhanced security and privacy
JP3466173B2 (en) 2000-07-24 2003-11-10 株式会社ソニー・コンピュータエンタテインメント Image processing system, device, method and computer program
JP4469476B2 (en) 2000-08-09 2010-05-26 パナソニック株式会社 Eye position detection method and eye position detection apparatus
US7277561B2 (en) 2000-10-07 2007-10-02 Qritek Co., Ltd. Iris identification
US7155036B2 (en) 2000-12-04 2006-12-26 Sony Corporation Face detection under varying rotation
US7015955B2 (en) 2000-12-22 2006-03-21 Eastman Kodak Company Camera having verification display with viewer adaptation compensation for reference illuminants and method
JP3586431B2 (en) 2001-02-28 2004-11-10 松下電器産業株式会社 Personal authentication method and device
KR100374708B1 (en) 2001-03-06 2003-03-04 에버미디어 주식회사 Non-contact type human iris recognition method by correction of rotated iris image
US7095901B2 (en) 2001-03-15 2006-08-22 Lg Electronics, Inc. Apparatus and method for adjusting focus position in iris recognition system
JP2002330318A (en) 2001-04-27 2002-11-15 Matsushita Electric Ind Co Ltd Mobile terminal
US6937135B2 (en) 2001-05-30 2005-08-30 Hewlett-Packard Development Company, L.P. Face and environment sensing watch
US6895103B2 (en) 2001-06-19 2005-05-17 Eastman Kodak Company Method for automatically locating eyes in an image
US20030169334A1 (en) 2001-08-06 2003-09-11 Michael Braithwaite Iris capture device having expanded capture volume
US7167987B2 (en) 2001-08-29 2007-01-23 Hewlett-Packard Development Company, L.P. Use of biometrics to provide physical and logic access to computer devices
US7269737B2 (en) 2001-09-21 2007-09-11 Pay By Touch Checking Resources, Inc. System and method for biometric authorization for financial transactions
US8200980B1 (en) 2001-09-21 2012-06-12 Open Invention Network, Llc System and method for enrolling in a biometric system
KR100456619B1 (en) 2001-12-05 2004-11-10 한국전자통신연구원 A system for registering and authenticating human face using support vector machines and method thereof
DE10163583A1 (en) 2001-12-21 2003-07-03 Philips Intellectual Property Method and device for exposing x-rays
KR100854890B1 (en) 2001-12-28 2008-08-28 엘지전자 주식회사 Iris recording and recognition method using of several led for iris recognition system
US7715595B2 (en) 2002-01-16 2010-05-11 Iritech, Inc. System and method for iris identification using stereoscopic face recognition
US7118042B2 (en) 2002-01-18 2006-10-10 Microscan Systems Incorporated Method and apparatus for rapid image capture in an image system
US8819788B2 (en) 2002-10-21 2014-08-26 Clearone Communications Hong Kong, Limited Method and system for providing security data to security stations
AU2003298731A1 (en) 2002-11-26 2004-06-18 Digimarc Id Systems Systems and methods for managing and detecting fraud in image databases used with identification documents
TWI224287B (en) 2003-01-23 2004-11-21 Ind Tech Res Inst Iris extraction method
US7280678B2 (en) 2003-02-28 2007-10-09 Avago Technologies General Ip Pte Ltd Apparatus and method for detecting pupils
US7380938B2 (en) 2003-03-25 2008-06-03 Sarnoff Corporation Apparatus to detect and measure saccade and pupilary changes
US7599524B2 (en) 2003-04-04 2009-10-06 Sarnoff Corporation Method and apparatus for providing a robust object finder
KR20030066512A (en) 2003-07-04 2003-08-09 김재민 Iris Recognition System Robust to noises
US7379567B2 (en) 2003-07-17 2008-05-27 Matsushita Electric Industrial Co., Ltd. Iris code generation method, individual authentication method, iris code entry device, individual authentication device, and individual certification program
DE10335261A1 (en) 2003-08-01 2005-02-17 Daimlerchrysler Ag Compressor and / or turbine wheel for a secondary air conveyor
US6912298B1 (en) 2003-08-11 2005-06-28 Adobe Systems Incorporation Object detection using dynamic probability scans
US7333653B2 (en) 2003-08-29 2008-02-19 Hewlett-Packard Development Company, L.P. Detecting and correcting redeye in an image
KR100682889B1 (en) 2003-08-29 2007-02-15 삼성전자주식회사 Method and Apparatus for image-based photorealistic 3D face modeling
WO2005024698A2 (en) 2003-09-04 2005-03-17 Sarnoff Corporation Method and apparatus for performing iris recognition from an image
US8098901B2 (en) 2005-01-26 2012-01-17 Honeywell International Inc. Standoff iris recognition system
US8090157B2 (en) 2005-01-26 2012-01-03 Honeywell International Inc. Approaches and apparatus for eye detection in a digital image
US7593550B2 (en) 2005-01-26 2009-09-22 Honeywell International Inc. Distance iris recognition
US7310443B1 (en) 2003-09-17 2007-12-18 Sonic Solutions, Inc. Automatic red eye detection and correction in digital images
US7295686B2 (en) 2003-09-29 2007-11-13 Primax Electronics Ltd. Method of processing red eye in digital images
US20050210267A1 (en) 2004-03-18 2005-09-22 Jun Sugano User authentication method and system, information terminal device and service providing server, subject identification method and system, correspondence confirmation method and system, object confirmation method and system, and program products for them
JP4059224B2 (en) 2004-04-13 2008-03-12 株式会社デンソー Driver appearance recognition system
JP2005334402A (en) 2004-05-28 2005-12-08 Sanyo Electric Co Ltd Method and device for authentication
US7466308B2 (en) 2004-06-28 2008-12-16 Microsoft Corporation Disposing identifying codes on a user's hand to provide input to an interactive display application
JP4574277B2 (en) 2004-08-09 2010-11-04 興和株式会社 Ophthalmic imaging equipment
WO2007011375A1 (en) 2004-09-13 2007-01-25 Cdm Optics, Inc. Iris image capture devices and associated systems
US7856558B2 (en) 2004-10-21 2010-12-21 Honeywell International Inc. Biometric verification and duress detection system and method
US20060184243A1 (en) 2004-10-22 2006-08-17 Omer Yilmaz System and method for aligning an optic with an axis of an eye
US7565183B2 (en) 2004-11-15 2009-07-21 Sony Ericsson Mobile Communications Ab Mobile device with selectable camera position
US7428320B2 (en) 2004-12-07 2008-09-23 Aoptix Technologies, Inc. Iris imaging using reflection from the eye
KR101224408B1 (en) 2005-01-26 2013-01-22 허니웰 인터내셔널 인코포레이티드 A distance iris recognition system
US7568628B2 (en) * 2005-03-11 2009-08-04 Hand Held Products, Inc. Bar code reading device with global electronic shutter control
JP4702598B2 (en) 2005-03-15 2011-06-15 オムロン株式会社 Monitoring system, monitoring apparatus and method, recording medium, and program
US20060222212A1 (en) 2005-04-05 2006-10-05 Yingzi Du One-dimensional iris signature generation system and method
US7542628B2 (en) 2005-04-11 2009-06-02 Sarnoff Corporation Method and apparatus for providing strobed image capture
US7634114B2 (en) 2006-09-01 2009-12-15 Sarnoff Corporation Method and apparatus for iris biometric systems for use in an entryway
US20060274918A1 (en) 2005-06-03 2006-12-07 Sarnoff Corporation Method and apparatus for designing iris biometric systems for use in minimally constrained settings
US20070047770A1 (en) 2005-06-13 2007-03-01 Swope Guy G Multiple biometrics enrollment and verification system
JP2007011667A (en) 2005-06-30 2007-01-18 Matsushita Electric Ind Co Ltd Iris authentication device and iris authentication method
JP4664147B2 (en) 2005-07-29 2011-04-06 株式会社山武 Iris authentication device
US20100176802A1 (en) 2005-08-16 2010-07-15 Koninklijke Philips Electronics, N.V. Portable electronic device having a rotary unit
WO2007025258A2 (en) 2005-08-25 2007-03-01 Sarnoff Corporation Methods and systems for biometric identification
US7873490B2 (en) 2005-12-28 2011-01-18 Solmetric Corporation Solar access measurement device
US20070160266A1 (en) 2006-01-11 2007-07-12 Jones Michael J Method for extracting features of irises in images using difference of sum filters
US7583823B2 (en) 2006-01-11 2009-09-01 Mitsubishi Electric Research Laboratories, Inc. Method for localizing irises in images using gradients and textures
GB0603411D0 (en) 2006-02-21 2006-03-29 Xvista Ltd Method of processing an image of an eye
JP2007225763A (en) 2006-02-22 2007-09-06 Sony Corp Imaging apparatus, imaging method, and program
US7844084B2 (en) 2006-02-27 2010-11-30 Donald Martin Monro Rotation compensated iris comparison
DE602007007062D1 (en) 2006-03-03 2010-07-22 Honeywell Int Inc IRISER IDENTIFICATION SYSTEM WITH IMAGE QUALITY METERING
US7903168B2 (en) 2006-04-06 2011-03-08 Eastman Kodak Company Camera and method with additional evaluation image capture based on scene brightness changes
WO2007124020A2 (en) 2006-04-21 2007-11-01 Sick, Inc. Image quality analysis with test pattern
US7682026B2 (en) 2006-08-22 2010-03-23 Southwest Research Institute Eye location and gaze detection system and method
WO2008032329A2 (en) 2006-09-13 2008-03-20 Alon Atsmon Providing content responsive to multimedia signals
WO2008091401A2 (en) 2006-09-15 2008-07-31 Retica Systems, Inc Multimodal ocular biometric system and methods
US8121356B2 (en) 2006-09-15 2012-02-21 Identix Incorporated Long distance multimodal biometric system and method
EP2062197A4 (en) * 2006-09-15 2010-10-06 Retica Systems Inc Long distance multimodal biometric system and method
US7574021B2 (en) 2006-09-18 2009-08-11 Sarnoff Corporation Iris recognition for a secure facility
US20080121721A1 (en) 2006-11-25 2008-05-29 Chun-Hsien Chen Reflective Reading Structure For Portable Computer
WO2008073962A2 (en) 2006-12-12 2008-06-19 Rutgers, The State University Of New Jersey System and method for detecting and tracking features in images
US7825958B2 (en) * 2007-01-25 2010-11-02 Research In Motion Limited Method and apparatus for controlling a camera module to compensate for the light level of a white LED
KR100807940B1 (en) 2007-03-08 2008-02-28 박경진 Intraocular lens
US8023699B2 (en) 2007-03-09 2011-09-20 Jiris Co., Ltd. Iris recognition system, a method thereof, and an encryption system using the same
US9002073B2 (en) 2007-09-01 2015-04-07 Eyelock, Inc. Mobile identity platform
WO2009029757A1 (en) 2007-09-01 2009-03-05 Global Rainmakers, Inc. System and method for iris data acquisition for biometric identification
EP2215579A4 (en) 2007-11-29 2013-01-30 Wavefront Biometric Technologies Pty Ltd Biometric authentication using the eye
CN104866553A (en) 2007-12-31 2015-08-26 应用识别公司 Method, system, and computer program for identification and sharing of digital images with face signatures
CA2736609C (en) 2008-02-14 2016-09-20 Iristrac, Llc System and method for animal identification using iris images
US8930238B2 (en) 2008-02-21 2015-01-06 International Business Machines Corporation Pervasive symbiotic advertising system and methods therefor
US9131141B2 (en) 2008-05-12 2015-09-08 Sri International Image sensor with integrated region of interest calculation for iris capture, autofocus, and gain control
US8994877B2 (en) * 2008-07-30 2015-03-31 Semiconductor Components Industries, Llc Method and system for synchronizing a flash to an imager
US8213782B2 (en) 2008-08-07 2012-07-03 Honeywell International Inc. Predictive autofocusing system
US8081254B2 (en) 2008-08-14 2011-12-20 DigitalOptics Corporation Europe Limited In-camera based method of detecting defect eye with high accuracy
US9633261B2 (en) 2008-08-22 2017-04-25 International Business Machines Corporation Salting system and method for cancelable iris biometric
US20100082398A1 (en) 2008-09-29 2010-04-01 Yahoo! Inc. System for providing contextually relevant data
US20100278394A1 (en) 2008-10-29 2010-11-04 Raguin Daniel H Apparatus for Iris Capture
US8317325B2 (en) 2008-10-31 2012-11-27 Cross Match Technologies, Inc. Apparatus and method for two eye imaging for iris identification
CN101750697A (en) 2008-12-10 2010-06-23 深圳富泰宏精密工业有限公司 External mirror frame of electronic device
US8615596B1 (en) 2009-01-14 2013-12-24 Sprint Communications Company L.P. Communication method and system for providing content to a communication device according to a user preference
US8374404B2 (en) 2009-02-13 2013-02-12 Raytheon Company Iris recognition using hyper-spectral signatures
WO2010121227A1 (en) 2009-04-17 2010-10-21 Ioculi Image diversion systems & methods for image capture devices
US8387858B2 (en) 2009-06-01 2013-03-05 Synderesis Technologies, Inc. Consumer rewards systems and methods
US20100328420A1 (en) 2009-06-29 2010-12-30 Roman Kendyl A Optical adapters for mobile devices with a camera
US20110007205A1 (en) 2009-07-08 2011-01-13 Dechnia, LLC Rear to forward facing camera adapter
US8452131B2 (en) 2009-09-25 2013-05-28 International Business Machines Corporation Iris recognition system and method
US20110081946A1 (en) 2009-10-07 2011-04-07 Byron Singh N John Singh Apparatus and method for changing an image-capturing position of a mobile phone camera using a mirror device
US20110142297A1 (en) 2009-12-16 2011-06-16 Eye Controls, Llc Camera Angle Compensation in Iris Identification
US20110187878A1 (en) 2010-02-02 2011-08-04 Primesense Ltd. Synchronization of projected illumination with rolling shutter of image sensor
JP5003774B2 (en) 2010-02-15 2012-08-15 東京エレクトロン株式会社 Developing device, developing method, and storage medium
US10178290B2 (en) 2010-02-17 2019-01-08 Sri International Method and apparatus for automatically acquiring facial, ocular, and iris images from moving subjects at long-range
KR101046459B1 (en) 2010-05-13 2011-07-04 아이리텍 잉크 An iris recognition apparatus and a method using multiple iris templates
US8957956B2 (en) 2010-06-09 2015-02-17 Honeywell International Inc. Method and system for iris image capture
US8150255B2 (en) * 2010-06-25 2012-04-03 Apple Inc. Flash control for electronic rolling shutter
US20120086645A1 (en) 2010-10-11 2012-04-12 Siemens Corporation Eye typing system using a three-layer user interface
US8955001B2 (en) 2011-07-06 2015-02-10 Symphony Advanced Media Mobile remote media control platform apparatuses and methods
KR20120067761A (en) 2010-12-16 2012-06-26 한국전자통신연구원 Apparatus for measuring biometric information using user terminal and method thereof
US8254768B2 (en) 2010-12-22 2012-08-28 Michael Braithwaite System and method for illuminating and imaging the iris of a person
US8195576B1 (en) 2011-01-31 2012-06-05 Bank Of America Corporation Mobile transaction device security system
US8625847B2 (en) 2011-03-21 2014-01-07 Blackberry Limited Login method based on direction of gaze
US8639058B2 (en) 2011-04-28 2014-01-28 Sri International Method of generating a normalized digital image of an iris of an eye
US8682073B2 (en) 2011-04-28 2014-03-25 Sri International Method of pupil segmentation
US8854446B2 (en) 2011-04-28 2014-10-07 Iristrac, Llc Method of capturing image data for iris code based identification of vertebrates
US8755607B2 (en) 2011-04-28 2014-06-17 Sri International Method of normalizing a digital image of an iris of an eye
EP2710514A4 (en) 2011-05-18 2015-04-01 Nextgenid Inc Multi-biometric enrollment kiosk including biometric enrollment and verification, face recognition and fingerprint matching systems
CN103763972A (en) 2011-06-20 2014-04-30 Rhp多媒体公司 Case for portable electronic device
EP2731490B1 (en) 2011-07-14 2015-01-21 Koninklijke Philips N.V. System and method for remote measurement of optical focus
KR102024954B1 (en) 2011-08-22 2019-09-24 아이락 엘엘씨 Systems and methods for capturing artifact free images
US8473748B2 (en) 2011-09-27 2013-06-25 George P. Sampas Mobile device-based authentication
WO2013049778A1 (en) 2011-09-30 2013-04-04 Intuitive Medical Technologies, Llc Optical adapter for ophthalmological imaging apparatus
US20130089240A1 (en) 2011-10-07 2013-04-11 Aoptix Technologies, Inc. Handheld iris imager
US9241200B2 (en) 2011-10-11 2016-01-19 Verizon Patent And Licensing Inc. Targeted advertising
WO2013056001A1 (en) 2011-10-12 2013-04-18 Carnegie Mellon University System and method for the long range acquisition of iris images for stationary and mobile subjects
TW201324180A (en) 2011-12-07 2013-06-16 Wistron Corp Method for integrating a mobile communication device with a portable electronic device and portable computer module thereof
GB2497553B (en) 2011-12-13 2018-05-16 Irisguard Inc Improvements relating to iris cameras
US20130188943A1 (en) 2012-01-19 2013-07-25 Shing-Wai David Wu Camera Adapter
US9330294B2 (en) 2012-01-26 2016-05-03 Aware, Inc. System and method of capturing and producing biometric-matching quality fingerprints and other types of dactylographic images with a mobile device
US9373023B2 (en) 2012-02-22 2016-06-21 Sri International Method and apparatus for robustly collecting facial, ocular, and iris images using a single sensor
US9100825B2 (en) 2012-02-28 2015-08-04 Verizon Patent And Licensing Inc. Method and system for multi-factor biometric authentication based on different device capture modalities
US9138140B2 (en) 2012-03-20 2015-09-22 Lrs Identity, Inc. Compact iris imaging system
CN102708357A (en) 2012-04-12 2012-10-03 北京释码大华科技有限公司 Single-image sensor based double-eye iris recognition equipment
US8977560B2 (en) 2012-08-08 2015-03-10 Ebay Inc. Cross-browser, cross-machine recoverable user identifiers
US8437513B1 (en) 2012-08-10 2013-05-07 EyeVerify LLC Spoof detection for biometric authentication
US20140055337A1 (en) 2012-08-22 2014-02-27 Mobitv, Inc. Device eye tracking calibration
CN104781714A (en) 2012-09-12 2015-07-15 欧乐柯利普有限责任公司 Removable optical devices for mobile electronic devices
US9154677B2 (en) 2012-09-20 2015-10-06 Apple Inc. Camera accessory for angled camera viewing
KR102039725B1 (en) 2012-11-23 2019-11-04 엘지디스플레이 주식회사 Array substrate for liquid crystal display and manufacturing method of the same
US9195890B2 (en) 2012-12-10 2015-11-24 Sri International Iris biometric matching system
ES2718493T3 (en) 2012-12-18 2019-07-02 Eyesmatch Ltd Aspect capture and visualization procedure
US9019420B2 (en) 2012-12-19 2015-04-28 Covr Llc Image redirection system for mobile devices
CN103048848A (en) 2012-12-21 2013-04-17 吴刚 Corner lens applied to camera of mobile device and assembly of corner lens
CN103099624A (en) 2013-01-11 2013-05-15 北京释码大华科技有限公司 Iris distance measurement plate, iris recognition all-in-one machine and iris recognition method thereof
US9304376B2 (en) 2013-02-20 2016-04-05 Hand Held Products, Inc. Optical redirection adapter
US20140327815A1 (en) 2013-05-06 2014-11-06 Cvs Pharmacy, Inc. Smartphone scanner adapter
US9390327B2 (en) 2013-09-16 2016-07-12 Eyeverify, Llc Feature extraction and matching for biometric authentication
EP3074924A4 (en) 2013-10-08 2017-11-22 Princeton Identity, Inc. Iris biometric recognition module and access control assembly
US10042994B2 (en) 2013-10-08 2018-08-07 Princeton Identity, Inc. Validation of the right to access an object
US10025982B2 (en) 2013-10-08 2018-07-17 Princeton Identity, Inc. Collecting and targeting marketing data and information based upon iris identification
US10038691B2 (en) 2013-10-08 2018-07-31 Princeton Identity, Inc. Authorization of a financial transaction
US9300769B2 (en) 2013-11-01 2016-03-29 Symbol Technologies, Llc System for and method of adapting a mobile device having a camera to a reader for electro-optically reading targets
US10121073B2 (en) * 2014-02-12 2018-11-06 Samsung Electronics Co., Ltd Agile biometric camera with bandpass filter and variable light source
WO2016018488A2 (en) 2014-05-09 2016-02-04 Eyefluence, Inc. Systems and methods for discerning eye signals and continuous biometric identification
US9767358B2 (en) 2014-10-22 2017-09-19 Veridium Ip Limited Systems and methods for performing iris identification and verification using mobile devices
US9412169B2 (en) 2014-11-21 2016-08-09 iProov Real-time visual feedback for user positioning with respect to a camera and a display
EP3227816A4 (en) 2014-12-03 2018-07-04 Princeton Identity, Inc. System and method for mobile device biometric add-on
JP6535223B2 (en) 2015-05-28 2019-06-26 浜松ホトニクス株式会社 Blink measurement method, blink measurement apparatus, and blink measurement program
US20160364609A1 (en) 2015-06-12 2016-12-15 Delta ID Inc. Apparatuses and methods for iris based biometric recognition
TWI558202B (en) 2015-10-14 2016-11-11 緯創資通股份有限公司 Imaging device, imaging system and control method thereof
US20170124314A1 (en) 2015-11-03 2017-05-04 Tavita Laumea Device and Method for Keypad with an Integral Fingerprint Scanner
KR102402829B1 (en) 2015-11-10 2022-05-30 삼성전자 주식회사 Method for user authentication and electronic device implementing the same
JP2019506694A (en) 2016-01-12 2019-03-07 プリンストン・アイデンティティー・インコーポレーテッド Biometric analysis system and method
WO2017173228A1 (en) 2016-03-31 2017-10-05 Princeton Identity, Inc. Biometric enrollment systems and methods
US10373008B2 (en) 2016-03-31 2019-08-06 Princeton Identity, Inc. Systems and methods of biometric analysis with adaptive trigger
WO2018111860A1 (en) 2016-12-12 2018-06-21 Princeton Identity, Inc. Systems and methods of biometric analysis

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112040139A (en) * 2020-09-09 2020-12-04 绍兴埃瓦科技有限公司 Light supplementing method for camera imaging
US11816198B2 (en) 2021-04-06 2023-11-14 Bank Of America Corporation Systems and methods for geolocation security using biometric analysis

Also Published As

Publication number Publication date
WO2017172695A1 (en) 2017-10-05
US10373008B2 (en) 2019-08-06
US20170286792A1 (en) 2017-10-05

Similar Documents

Publication Publication Date Title
US10373008B2 (en) Systems and methods of biometric analysis with adaptive trigger
US11132771B2 (en) Bright spot removal using a neural network
US20210014411A1 (en) Method for image processing, electronic device, and computer readable storage medium
US9953428B2 (en) Digital camera unit with simultaneous structured and unstructured illumination
KR102270674B1 (en) Biometric camera
CN108805198B (en) Image processing method, image processing device, computer-readable storage medium and electronic equipment
US20180165537A1 (en) Systems And Methods Of Biometric Analysis
US20050212913A1 (en) Method and arrangement for recording regions of interest of moving objects
WO2021073140A1 (en) Monocular camera, and image processing system and image processing method
CN104137529A (en) Method and apparatus for enhanced automatic adjustment of focus, exposure and white balance in digital photography
CN110536068A (en) Focusing method and device, electronic equipment, computer readable storage medium
CN109862269A (en) Image-pickup method, device, electronic equipment and computer readable storage medium
CN108322651B (en) Photographing method and device, electronic equipment and computer readable storage medium
CN109327626B (en) Image acquisition method and device, electronic equipment and computer readable storage medium
US9489750B2 (en) Exposure metering based on background pixels
US20110081142A1 (en) Pulsed control of camera flash
US9195884B2 (en) Method, apparatus, and manufacture for smiling face detection
CN108600740A (en) Optical element detection method, device, electronic equipment and storage medium
CN110248101A (en) Focusing method and device, electronic equipment, computer readable storage medium
US20130286218A1 (en) Image recognition device that recognizes specific object area, method of controlling the device, and storage medium, as well as image pickup apparatus, and display device
US8442349B2 (en) Removal of artifacts in flash images
CN107040697A (en) Use the method for imaging and relevant camera system of gaze detection
US20200020086A1 (en) Information processing method and information processing apparatus
US20160350607A1 (en) Biometric authentication device
JP2023059952A (en) Image processing device, imaging device, image processing method, image processing program, and recording medium

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION