US20140355846A1 - Fingerprint Sensing Apparatus - Google Patents
Fingerprint Sensing Apparatus Download PDFInfo
- Publication number
- US20140355846A1 US20140355846A1 US14/463,062 US201414463062A US2014355846A1 US 20140355846 A1 US20140355846 A1 US 20140355846A1 US 201414463062 A US201414463062 A US 201414463062A US 2014355846 A1 US2014355846 A1 US 2014355846A1
- Authority
- US
- United States
- Prior art keywords
- light
- fingerprint
- sensor
- pixels
- scattered light
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/13—Sensors therefor
- G06V40/1318—Sensors therefor using electro-optical elements or layers, e.g. electroluminescent sensing
-
- G06K9/0004—
-
- G06K9/00087—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/1382—Detecting the live character of the finger, i.e. distinguishing from a fake or cadaver finger
- G06V40/1394—Detecting the live character of the finger, i.e. distinguishing from a fake or cadaver finger using acquisition arrangements
-
- H04N5/359—
Definitions
- the present invention relates to the field of fingerprint detection system.
- the present invention relates to a thin high contrast optical acquisition system for fingerprint recognition.
- a method for determining validity of a fingerprint includes determining a set of light sources from a plurality of light sources for emitting light to a fingerprint, determining a set of sensor zones from a plurality of sensor zones for sensing scattered light from the fingerprint, determining a minimum distance between a sensor zone and a light source for sensing selective scattered light from the fingerprint, emitting light from the set of light sources to generate the scattered light from the fingerprint, sensing the scattered light in the set of sensor zones, and determining a validity of the fingerprint using the scattered light sensed in the plurality of sensor zones.
- an apparatus for determining validity of a fingerprint includes a light refracting device (light refractor), a light source, a light collecting device, and a controller.
- the light refracting device can, for example, be an active matrix organic light emitting diode (AMOLED) panel structure with reverse current measurement and amplification circuitry, and includes an imaging surface and a viewing plane. Incident light from the light source is projected directly or indirectly onto the imaging surface to create an image of the patterned object from the projected light onto the viewing plane.
- the apparatus is configured to have a thin form factor, which may be flexible or conformable, compared to conventional optical fingerprint acquisition apparatuses.
- the AMOLED panel includes the light source panel as well as light collecting devices.
- the fingerprint sensing apparatus can be implemented as an in-cell structure.
- an apparatus for determining validity of a fingerprint includes a light refracting device (light refractor), a light source, a light collecting device, and a controller.
- the light refracting device can, for example, be a thin-film transistor (TFT) panel and includes an imaging surface, a light receiving surface, a viewing plane, and light collecting devices.
- the light source can be an individually addressable panel of discrete light sources, for example, a liquid crystal display (LCD) panel or an AMOLED panel. Incident light from the light source is projected through the light receiving surface and projected directly or indirectly onto the imaging surface to create an image of the patterned object from the projected light onto the viewing plane.
- the apparatus is configured to have a thin form factor, which may be flexible or conformable, compared to conventional optical fingerprint acquisition apparatuses.
- the TFT panel could be implemented as an add-on panel that is placed on top of the light source panel.
- FIG. 1A illustrates light characteristics based on illumination of a ridge of a fingerprint according to aspects of the present disclosure.
- FIG. 1B illustrates light characteristics based on illumination of a valley of a fingerprint according to aspects of the present disclosure.
- FIG. 2 illustrates a potential issue of interferences from two light sources according to aspects of the present disclosure.
- FIG. 3A illustrates an exemplary method of avoiding interferences from two light sources with respect to scattered light from ridges of a fingerprint according to aspects of the present disclosure.
- FIG. 3B illustrates an exemplary method of avoiding interferences from two light sources with respect to scattered light from a fingerprint according to aspects of the present disclosure.
- FIG. 3C illustrates another exemplary method of avoiding interferences from two light sources with respect to scattered light from a fingerprint according to aspects of the present disclosure.
- FIG. 3D illustrates a plot of light intensity overlap for the example of FIG. 3C according to aspects of the present disclosure.
- FIG. 4A illustrates a side view of the exemplary method of FIG. 3A according to aspects of the present disclosure.
- FIG. 4B illustrates a top view of the exemplary method of FIG. 3A according to aspects of the present disclosure.
- FIG. 5A illustrates an exemplary implementation of addressing a plurality of non-overlapping light sources and sensor zones according to aspects of the present disclosure.
- FIG. 5 B illustrates another exemplary implementation of addressing a plurality of non-overlapping light sources and sensor zones according to aspects of the present disclosure.
- FIG. 6A illustrates an exemplary implementation of scanning a fingerprint according to embodiments of the present disclosure.
- FIG. 6B illustrates another exemplary implementation of scanning a fingerprint according to embodiments of the present disclosure.
- FIG. 6C illustrates exemplary separations between adjacent sensor zones for avoiding interferences according to aspects of the present disclosure.
- FIG. 7A illustrates a method of capturing and validating a fingerprint image
- FIG. 7B illustrates a method of determining a set of light sources from a plurality of light sources for emitting light to a fingerprint
- FIG. 7C illustrates a method of sensing scattered light in a set of sensor zones
- FIG. 7D illustrates a method of determining a validity of the fingerprint using the scattered light sensed in a plurality of sensor zones
- FIG. 7E illustrates a further method of capturing and validating the fingerprint image of FIG. 7A according to aspects of the present disclosure.
- FIG. 8A illustrates an exemplary implementation of collecting sensor data according to aspects of the present disclosure.
- FIG. 8B illustrates another exemplary implementation of collecting sensor data according to aspects of the present disclosure.
- FIG. 8C illustrates yet another exemplary implementation of collecting sensor data according to aspects of the present disclosure.
- FIG. 8D illustrates yet another exemplary implementation of collecting sensor data according to aspects of the present disclosure.
- FIG. 8E illustrates yet another exemplary implementation of collecting sensor data according to aspects of the present disclosure.
- FIG. 9 illustrates effects of controlling a light emission cone on a corresponding active sensor zone according to aspects of the present disclosure.
- FIG. 10A illustrates a unidirectional OLED with fingerprint sensors according to aspects of the present disclosure
- FIG. 10B illustrates a bi-directional OLED with fingerprint sensors according to aspects of the present disclosure
- FIG. 10C illustrates a conformable OLED with fingerprint sensors according to aspects of the present disclosure
- FIG. 10D illustrates an exemplary OLED structure according to aspects of the present disclosure.
- FIG. 11A illustrates an exemplary subpixel circuit cell with forward bias according to aspects of the present disclosure
- FIG. 11B illustrates an exemplary subpixel circuit cell with reverse bias according to aspects of the present disclosure.
- FIG. 12 illustrates an exemplary pixel circuit cell with RGB subpixels according to aspects of the present disclosure.
- FIG. 13 illustrates a process of fingerprint acquisition using AMOLED according to aspects of the present disclosure.
- FIG. 14 illustrates an exemplary controller of a thin high contrast optical acquisition system for fingerprint recognition according to aspects of the present disclosure.
- FIG. 15 illustrates an exemplary light sensing panel using a thin film transistor (TFT) panel structure according to aspects of the present disclosure.
- TFT thin film transistor
- FIG. 16 illustrates an exemplary fingerprint detection system according to aspects of the present disclosure.
- FIG. 17 illustrates an exemplary panel structure according to aspects of the present disclosure.
- a procedure, computer-executed step, logic block, process, etc. is here conceived to be a self-consistent sequence of one or more steps or instructions leading to a desired result.
- the steps are those utilizing physical manipulations of physical quantities. These quantities can take the form of electrical, magnetic, or radio signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. These signals may be referred to at times as bits, values, elements, symbols, characters, terms, numbers, or the like.
- Each step may be performed by hardware, software, firmware, or combinations thereof.
- the drawing of the fingerprint acquisition apparatus shows incident light traveling from a light source component L in the light emission plane, through an optical structure to the imaging surface, and from the imaging surface to a light collecting (sensor) component S in the light emission plane.
- Incident light interacts with a finger placed on the imaging surface and is reflected, refracted, or scattered from the imaging surface.
- the pixel panel is configured so that there are multiple lighting pixels, to serve as discrete light sources, and multiple sensor pixels, to receive light carrying fingerprint image information.
- the light components used as light sources or sensors can be either pixels or subpixels that are arranged in an addressable pattern in the light emission plane.
- the pixel panel can be a display panel or non-display panel but is not limited LED or TFT-type display panels.
- the light emission plane can be considered as the same plane as the light collecting plane because the transparent areas of the cells of the TFT panel are aligned with the transparent areas of the cells of the light source panel.
- both the light emission plane and light collecting plane are one and the same.
- light is projected from one light source L onto the imaging surface at an incident angle that is less than the critical angle at the light refractor/air interface.
- incident light strikes the imaging surface where a finger is placed, it will get scattered by both ridges and valleys albeit to differing degrees and intensities. Ridges will scatter light across a wider range of angles compared to the valleys. More specifically, light scattered from ridges will hit the emission plane over a larger area compared to the area hit by light scattered from valleys.
- the sensor zone of the present disclosure can be determined in view of the characteristics of the light scattered, and it can be configured to differentiate the light scattered by the two topographical features.
- the difference in scattering range is due to the presence of air where there are valleys at the imaging surface. Since the angle of incident light is less than the critical angle, light passes out of the pixel panel, through the imaging surface into the air, strikes the valley, scatters, re-enters the pixel panel, and refracts. Because of the refraction, valleys scatter light onto the emission plane within a smaller area than that of the ridges (for example, within radius r 0 as shown in FIG. 4A ). Both valleys and ridges can scatter light into the same small area; there is no differentiation between light scattered from ridges and valleys, therefore, light is not collected in the small area close to the light source.
- the pixel panel provides a plurality of sensors across a controllable area to gather the well-differentiated light that is used to generate high-contrast fingerprint images.
- the light strikes where a ridge is touching the imaging surface. Part of the incident light is absorbed by the ridge while the rest of the light is scattered towards the light emission plane.
- a sensor S receives a ray of the scattered light. The intensity of light received by the sensor S is assigned to the position that corresponds to the point of illumination by incident light from the light source L. The image of ridges will appear bright.
- the light strikes the surface where there is a valley.
- the light passes through the surface and strikes the valley.
- Part of the light is absorbed by the valley while the rest of the light is scattered.
- Part of the scattered light re-enters the optical structure through the imaging surface and is refracted having an angle ⁇ R normal to the imaging surface.
- a ray of refracted light hits the light emission plane at a distance x from the light source. This distance x is limited by the critical angle at the light refractor/air interface; x also depends on the depth of the valley and the thickness of the optical structure (th).
- the optical imaging apparatus can be configured to address individual, multiple, or regions of pixels such that when L is activated. All sensors S may be located within a predetermined distance from L at x S , e.g. x S ⁇ x(max), can be ignored, not used, or turned off, and at least one sensor S may be located at a predetermined distance from L at x S , e.g. x S >x(max), can be turned on, and incoming light is collected; such a sensor is called an active sensor.
- All sensors S may be located within a predetermined distance from L at x S , e.g. x S ⁇ x(max), can be ignored, not used, or turned off, and at least one sensor S may be located at a predetermined distance from L at x S , e.g. x S >x(max), can be turned on, and incoming light is collected; such a sensor is called an active sensor.
- the active sensor will receive light scattered only from a ridge because no light scattered from a valley can be refracted by the optical structure beyond the critical angle, or beyond the distance x(max) except for negligible stray light.
- the following paragraphs describe a method to calculate a value for x(max) where there is a valley at the point of illumination.
- ⁇ R The largest possible value for ⁇ R is the critical angle at the light refractor/air interface, although ⁇ R is likely to be smaller than the critical angle in practice.
- the fingerprint acquisition apparatus can be designed to have a larger x(max) to reduce detection of stray light, which would reduce the sensitivity but would be offset by greater clarity and differentiation of fingerprint ridges and valleys. In some applications, however, a smaller x(max) can be chosen to boost sensitivity without lowering contrast to intolerable levels. Since the width of a valley can vary from one point of illumination to another, in some other applications, a large enough value of x(max) is estimated to cover the possible range of valley widths that may be encountered in such applications. If there are two or more active sensors, the intensity of light received by one sensor can be compared with that received by another sensor to calculate an overall brightness that corresponds to the point of illumination.
- the measured or calculated intensities of light received from various points of illumination can be assembled to map out the topography of the fingerprint corresponding to each point of illumination. Brighter areas indicate higher elevation (e.g. ridges) while darker areas indicate lower elevation (e.g. valleys).
- scattered light could be collected at a large distance beyond x (max). But it may not be practical to do so too far away for the following reasons: (a) the light intensity drops the farther away from the point of illumination, which would reduce sensitivity; (b) excessive redundant data would be generated, which would waste storage and processing resources; and/or (c) it may be advantageous to use a plurality of simultaneous light sources to speed up data collection. For faster scanning of a whole fingerprint, one approach is to illuminate more than one point from more than one light source at a time.
- light is projected simultaneously from two light sources, separated by a distance x L , onto the imaging surface, each having an incident angle that is less than the critical angle at the light refractor/air interface.
- Light source L 1 is located too close to light source L 2 causing interference: one sensor receives light from rays scattered from two points of illumination, such as two ridges.
- x L can be large enough so that an active sensor does not receive overlapping light from rays scattered from two separate points of illumination as shown in FIG. 3A and FIG. 3B .
- sensors located at a distance from L less than x(max) are ignored, not used, or turned off, and at least one sensor located at a distance from light source, between r 0 and r, where r 0 >x(max) and r>r 0 , is an active sensor ( FIG. 4A and FIG. 4B ).
- the value for r can be chosen as the distance where scattered light intensity hitting the emission plane decreases to a desired threshold value, such as ⁇ 20%, ⁇ 10%, ⁇ 5%, 1%, or ⁇ 1% of the scattered light intensity hitting the emission plane at r 0 .
- a desired threshold value such as ⁇ 20%, ⁇ 10%, ⁇ 5%, 1%, or ⁇ 1% of the scattered light intensity hitting the emission plane at r 0 .
- a buffer zone may be added beyond the active sensor(s) to avoid interference at the active sensor(s).
- r can be chosen as small as slightly larger than r 0 .
- x L (min) is referred to as the minimum distance between two light sources that can be simultaneously illuminated such that at least one active sensor will not receive interference from light reflected, refracted, or scattered from the point of illumination by the other light source.
- x L (min) can be estimated by using a number of models for attenuation of light intensity.
- x S is the distance from a sensor to its corresponding light source L. If assuming a hemispherical model for attenuation, the intensity will decrease by 1/x S 2 .
- the function determines how far x S can be where the intensity of light becomes negligible.
- the optical imaging apparatus can be configured so that light from two points of illumination do not overlap. To do so, x L (min) needs to be large enough but not too large as to make it impractical. x L (min) can be small enough that allow overlap of negligible light. A threshold can be determined for negligible light intensity.
- x B represents the width of a buffer zone, located between neighboring active sensor zones, in which scattered light is not sensed, the intensity of light becomes negligible, or scattered light from two points of illumination can overlap without causing interference that degrades the image.
- the threshold where the intensity of light becomes negligible can be determined for a particular design requirement. For example, the intensity may become negligible when it becomes ⁇ 30%, ⁇ 20%, ⁇ 10%, ⁇ 5%, 1%, or ⁇ 1% of the intensity of light at r 0 .
- the optical imaging apparatus can be configured so that light scattered from two points of illumination and collected in their respective active sensor zones do not overlap.
- x L (min) needs to be large enough but not too large as to make it impractical.
- x L (min) can be small enough to allow a tolerable overlap of negligible light for certain applications (such as shown in FIG. 5A ).
- FIG. 3C illustrates another exemplary method of avoiding interferences from two light sources with respect to scattered light from a fingerprint according to aspects of the present disclosure.
- FIG. 3D illustrates a plot of light intensity overlap for the example of FIG. 3C according to aspects of the present disclosure.
- An active sensor zone may include one light source and at least one active sensor that will not receive interference from a point of illumination from another light source.
- the active sensor zone can be a circular region surrounding one light source L circumscribed by a radius r.
- This zone includes at least one active sensor located inside of a circular ring having an inner radius of r 0 >x(max) and an outer radius of r>r 0 ( FIG. 4A and FIG. 4B ).
- multiple, non-overlapping active sensor zones can be activated simultaneously.
- the buffer zones may overlap if the interference due to the overlap can be tolerated by the given application. Examples of multiple active sensor zones are shown in arrays of different orientations in FIG. 5A and FIG. 5B . After one set of zones is illuminated and light information is collected, that set is deactivated, and another set of zones is activated in increments in space and time such as in a series of horizontal and vertical scans, or a combination of the two different scanning methods ( FIG. 6A and FIG. 6B ). This scanning process is repeated until the desired amount of information of the fingerprint is collected to assemble an image.
- the image quality and scanning/processing speed can be optimized by adjusting the arrangement of zones.
- the activated sensor zones such as in FIG. 5A can be arranged across the entire imaging area in rows, columns, or other conformations that allow for efficient addressing of the pixels and data collection.
- FIG. 5B shows an example of tighter packing of sensor zones for improved efficiency of fingerprint scanning. Space can be added between activated sensor zones such as in FIG. 6C to reduce leakage of stray light from one zone to another zone that may cause interference or introduce artifacts.
- the number of activated sensors within an active sensor zone may be reduced to form a line as in FIG. 8A or a cross such as in FIG. 8D .
- Removing sensors that are close to neighboring active sensor zones can also reduce possible interference from the neighboring zones.
- the line or cross shaped zones can be arranged into arrays such as in FIGS. 8B , 8 C, and 8 E that can help to further reduce the possibility of overlap and interference while allowing for tighter packing for improved efficiency of fingerprint scanning.
- the thickness of the optical structure can be reduced. This can reduce the active sensor zone radius, r, which allows for either a more compact array of zones which results in faster scanning or a larger gap between zones which produces a higher scanning quality.
- the light emitted from a light source may leave the pixel at an emission angle ranging from zero to ⁇ from the normal of the light emission plane.
- a cone of light shown as ⁇ can strike the imaging surface, effectively illuminating a circular area.
- the angle of incidence at the imaging surface, ⁇ i would range from zero to ⁇ , creating a circular area of illumination.
- the light scattered by a ridge in this circular area would widen the radius of an active sensor zone from r to r′ and possibly the x(max) so that r 0 ′>r 0 .
- the active sensor zones can be spaced further apart to account for a non-zero value of beta.
- the optical apparatus can be configured to reduce ⁇ and reduce the thickness of the optical structure, either of which can help to reduce the area of illumination.
- ⁇ may not be greater than the critical angle at the light refractor/air interface. If ⁇ cannot be reduced, then the material of the light refractor can be selected to increase the critical angle. If the rings of active sensor zones are widened due to a non-zero ⁇ , then it would preferable to activate more sensors within each ring to increase sensitivity. Alternatively, when generating the fingerprint image, the light received by sensors closer to the inner-most ring near r 0 can be used, while the light received by the remaining sensors can be discarded or ignored. Using the methodologies described above, the image of a fingerprint can have a high contrast, which is important for accurate fingerprint recognition.
- An add-on type of light refractor can be flexible, conformable, or transparent.
- One example of such an add-on light refractor can be a TFT panel structure including light sensing components on the bottom of the panel, beneath a substrate.
- the substrate can be, for example, glass, plastic, or polymer. Together with the substrate, the TFT panel structure may be configured to function as a light sensing panel.
- FIG. 15 illustrates an exemplary light sensing panel using a thin film transistor (TFT) panel structure according to aspects of the present disclosure.
- TFT thin film transistor
- Each cell of the TFT panel structure can be an addressable light sensing component, referred to as a sensing pixel.
- capture sensor 1500 includes a passivation layer 1518 , which can be formed of SiNx.
- passivation layer 1518 On top of passivation layer 1518 , a storage capacitor layer is formed including first electrode 1515 .
- This storage capacitor layer is preferably formed from indium tin oxide (ITO), which is conductive and transparent.
- ITO indium tin oxide
- an insulating layer 1517 is formed, preferably of SiNx.
- a second electrode 1514 is formed, preferably of tin oxide.
- First electrode 1515 , insulating layer 1517 and second electrode 1514 together form the storage capacitor.
- another insulating layer 1516 is formed, which can be formed from SiNx.
- a layer of glass layer 1511 is placed over insulating layer 1516 .
- a fingerprint to be imaged is placed on glass layer 1511 , which may be referred to herein as the imaging surface.
- a light sensing unit 1512 which is preferably a thin-film transistor, and a switching unit 1513 , which is also preferably a thin-film transistor, are horizontally arranged on a passivation layer 1518 .
- a back light 1520 irradiates light upward to be passed through the fingerprint capture sensor 1500 .
- back light 1520 can be separated from a lower, exposed surface of pas sivation layer 1518 . It is also considered, however, that backlight 1520 be placed against lower surface of passivation layer 1518 .
- Backlight 1520 can be an LED or any other type of light source.
- a source electrode 1512 -S of the light sensing unit 1512 and a drain electrode 1513 -D of the switching unit 1513 are electrically connected through second electrode 1514 .
- a gate electrode 1512 -G of the light sensing unit 1512 is connected to first electrode 1515 .
- a first light shielding layer 1513 - sh is placed between insulating layer 1517 and passivation layer 1518 at switching unit 1513 . As detailed below, first light shielding layer 1513 - sh blocks light from backlight 1520 from reaching switching unit 1513 .
- second light shielding layer 1522 is positioned between glass layer 1511 and insulating layer 1516 at switching unit 1513 to shield switching unit 1513 from light passing through or reflected from glass layer 1511 .
- a photosensitive layer 1512 -P such as amorphous silicon (a-Si:H) is formed between the drain electrode 1512 -D and source electrode 1512 -S of the light sensing unit 1512 .
- photosensitive layer 1512 -P allows current to flow in response to a predetermined amount of light striking a surface of photosensitive layer 1512 -P. In this way, when more than a predetermined quantity of light is received at a surface of photosensitive layer 1512 -P, current flows through the drain electrode 1512 -D and the source electrode 1512 -S.
- a second light shielding layer 1522 is first placed on glass layer 1511 via evaporation, sputtering or any other method.
- Glass layer 1511 is preferably between about 5 and 10 um, though may be either thicker or thinner.
- Light shielding layer 1522 is preferably formed from a metal such as aluminum, but may be formed from any suitable light blocking material.
- insulating layer 1516 is formed on top of glass layer 1511 and second light shielding layer 1522 . As noted above, insulating layer 1516 is preferably formed from SiNx. Photosensitive layer 1512 -P is then formed over insulating layer 116 .
- photosensitive layer 1512 -P is preferably formed from a-Si:H.
- Source electrode 1512 -D of light sensing unit 1512 , second electrode 1514 and drain electrode 1513 -D of switching unit 1513 are next formed over insulating layer 1516 .
- Source electrode 1512 -D, second electrode 1514 and drain electrode 1513 -D are each preferably formed of ITO, but may be formed of any suitable conductor.
- insulating layer 1517 is formed and over insulating layer 1517 first electrode 1515 is formed. Insulating layer 1517 is preferably formed from SiNx and first electrode 1515 is preferably formed of ITO but may be formed of any suitable conductor.
- gate electrode 1512 -G of light sensing unit 1512 and light shield 113 - sh are formed.
- gate electrode 1512 -G and light shielding layer 1513 - sh are each formed of ITO, but may be formed of any suitable material and light shielding layer 1513 - sh does not need to be formed from the same material as gate electrode 1512 -G.
- passivation layer 1518 which is preferably formed from SiNx, is formed over first electrode 1515 , gate electrode 1512 -G and light shielding layer 1513 - sh .
- backlight 1520 can either be attached to the lower, exposed surface of passivation layer 1518 or separately supported.
- an image capture sensor can have substantially the same structure as capture sensor shown in FIG. 15 except that conductive ITO layer is placed beneath glass layer and an insulating layer, which can be formed of SiNx, is placed below ITO layer. Because ITO layer is conductive, electrostatic charge built up on glass layer can be discharged by connecting ITO layer to a ground. This can prevent damage to capture sensor.
- Image capture sensor can be fabricated in substantially the same manner as image capture sensor except that ITO layer is formed over glass layer and insulating layer is formed over ITO layer prior to forming light shielding layer over insulating layer.
- an image capture sensor can have substantially the same structure as capture sensor shown in FIG. 15 .
- the capture sensor includes a light sensing unit, which is substantially the same and light sensing unit, and switching unit, which is substantially the same as switching unit, formed between an insulating layer and a passivation layer.
- above insulating layer capture sensor includes a substrate layer having a plurality of fiber-optic strands running in a direction perpendicular to a surface of substrate layer.
- the diameter of the fiber-optic strands 330 a forming substrate layer is from about 4 um to about 8 um in diameter and more preferably about 6 um in diameter, though larger or smaller diameters can also be used.
- Substrate layer can be formed from glass fiber optic strands 330 a or fiber optic strands of other substantially transparent materials including polymers. Fiber optic sheets can be used to form the substrate layer.
- a light sensor panel can be implemented as an add-on panel that is placed on top of a light source panel.
- the light source panel can be, for example, an LCD panel or an AMOLED panel.
- FIG. 16 illustrates an exemplary fingerprint detection system, in which a TFT type light sensor panel is placed on top of an LCD display panel structure as an add-on panel.
- a TFT type light sensor panel is placed on top of an LCD panel structure as an add-on panel.
- the sensing pixels of the TFT type light sensing panel can be individually addressable and can be activated according to a designated sensor zone pattern.
- FIG. 17 illustrates an exemplary TFT light sensor panel aligned with an LCD panel structure, wherein non-transparent components of the TFT light sensor panel are aligned with the black matrix areas of the LCD display panel structure.
- the TFT light sensor panel is aligned with the LCD panel structure.
- the non-transparent components on the TFT light sensor panel are aligned with the black matrix area on the LCD display panel structure.
- the black matrix areas of the LCD display panel are non-transparent and therefore would block the transmission of the display backlight.
- the light sensor panel can be designed so that its non-transparent areas can be aligned with the black matrix areas of the LCD panel. When the LCD display emits light through the transparent areas of the LCD display, this light can be used as the light source for the light sensor panel.
- the LCD display can individually control cells (individually addressable) to emit light as discrete light sources that are projected into the light refractor according to a designated illumination pattern.
- the light refracting device can, for example, also be a thin-film transistor (TFT) add-on panel placed on top of an LCD or AMOLED display panel structure that acts as a panel of light sources. Incident light from the light source panel is projected through the light receiving surface and projected directly or indirectly onto the imaging surface to create an image of the patterned object from the projected light onto the viewing plane.
- TFT thin-film transistor
- This fingerprint sensing apparatus can be also used as a touch sensor when implemented in a mobile device.
- both top emission and bottom emission type OLED structures can be used as the main component of a fingerprint acquisition apparatus.
- OLED devices such as small molecule OLED, polymer OLED, or solution based OLED, may be utilized as main OLED device structures.
- Both transparent and non-transparent OLED panels can be used as the main component of a fingerprint acquisition apparatus.
- FIG. 10B Both thin panel and flexible or conformable types of OLED panels can be used as the main component of a fingerprint acquisition apparatus.
- FIG. 10C Both thin panel and flexible or conformable types of OLED panels can be used as the main component of a fingerprint acquisition apparatus.
- An active matrix OLED (AMOLED) panel can be used as the main component of a fingerprint acquisition apparatus.
- An AMOLED panel may include subpixel areas (red, green, and blue subpixels) and a driving circuit area (thin film transistor and capacitor).
- the brightness of each subpixel can be adjusted by the driving and switching transistors and capacitors and by controlling the amount of current injected to the OLED subpixels.
- the dimension of subpixels can be formed using OLED material deposition techniques. For instance, the size and position of subpixels can be set by using shadow masks during the OLED material evaporation process.
- An OLED may have a layered structure with the following sequence: anode/hole injection layer/hole transport layer/emissive layer/electron transport layer/electron injection layer/cathode.
- ITO and other transparent conducting materials having high work function can be used for anode materials, and metals such as aluminum and magnesium can be used for cathode materials.
- FIG. 10D shows the structure of a bottom emission OLED. In this example, the imaging surface would be at the bottom of the substrate, and the light emission plane would be the cathode layer.
- the optical structure may include the transparent layers between the substrate and the cathode.
- FIG. 11A shows an exemplary AMOLED subpixel unit cell circuit (2D-driving TFT circuit with subpixels).
- the driving area may include a driving transistor, switching transistor, holding capacitor, and reverse current sensor.
- FIG. 11B shows the reverse current read and amplified in the OLED circuit structure.
- an AMOLED panel has a three-subpixel structure.
- a blue subpixel can be used as a light source while the neighboring green or red subpixels may be used as a sensor because the band gap of blue subpixels is larger than that of the green or red subpixels.
- FIG. 12 shows an exemplary R/G/B pixel structure where the blue subpixel is the light source, and the green or red subpixel is the sensor. The reverse voltage can be biased in the sensor subpixel when the lighting subpixel is turned on.
- the I-V curves correspond with subpixel structures in FIG. 12 .
- the amount of reverse current in the sensor subpixel under reverse bias is increased when light is reflected, refracted, or scattered from a fingerprint to the sensor subpixel.
- the amount of reverse current can be measured using current sensing circuits in the driving circuit area.
- the reverse current signal can be amplified using an amplification circuit, and/or a signal processor.
- the amplified current signal can then be processed to generate a fingerprint image by a signal processing algorithm ( FIG. 13 , algorithm flow chart).
- the OLED panel resolution can be controlled by varying the size and density of each subpixel and by setting the subpixel structure of the OLED panel.
- an OLED panel may have a larger lighting component (e.g. blue subpixels) and a smaller sensor component (e.g. green and/or red subpixels).
- subpixel structures can have different sizes.
- the subpixel density can be enhanced by changing pixel shape from stripe type to circular or diamond shape.
- an OLED subpixel structure can have different shapes, such as square, rectangle, circle, diamond, etc.
- the patterning of the subpixel structure can be fabricated by using fine metal mask processes, ink-jet printing, or laser transfer technologies.
- a mobile device is usually equipped with a touch sensor. If a mobile device was equipped with the fingerprint sensing apparatus of the present disclosure, then the touch sensor would not be required, as the fingerprint sensing apparatus may also be used as a touch sensor.
- a mobile device can be configured to include a thin high contrast optical acquisition system as a fingerprint sensing apparatus for fingerprint recognition.
- the mobile device may comprise a wireless transceiver which is capable of transmitting and receiving wireless signals via wireless antenna over a wireless communication network.
- Wireless transceiver may be connected to a bus by a wireless transceiver bus interface.
- the wireless transceiver bus interface may, in some embodiments be at least partially integrated with wireless transceiver.
- Some embodiments may include multiple wireless transceivers and wireless antennas to enable transmitting and/or receiving signals according to a corresponding multiple wireless communication standards such as, for example, versions of IEEE Std. 802.11, CDMA, WCDMA, LTE, UMTS, GSM, AMPS, Zigbee and Bluetooth®, etc.
- a corresponding multiple wireless communication standards such as, for example, versions of IEEE Std. 802.11, CDMA, WCDMA, LTE, UMTS, GSM, AMPS, Zigbee and Bluetooth®, etc.
- the mobile device may also comprise a SPS receiver capable of receiving and acquiring SPS signals via a SPS antenna.
- the SPS receiver may also process, in whole or in part, acquired SPS signals for estimating a location of the mobile device.
- processor(s), memory, DSP(s) and/or specialized processors may also be utilized to process acquired SPS signals, in whole or in part, and/or calculate an estimated location of the mobile device, in conjunction with the SPS receiver. Storage of SPS or other signals for use in performing positioning operations may be performed in memory or registers (not shown).
- the mobile device may comprise digital signal processor(s) (DSP(s)) connected to the bus by a bus interface, processor(s) connected to the bus by a bus interface and memory.
- DSP(s) digital signal processor
- the bus interface may be integrated with the DSP(s), processor(s) and memory.
- functions may be performed in response execution of one or more machine-readable instructions stored in memory such as on a computer-readable storage medium, such as RAM, ROM, FLASH, or disc drive, just to name a few example.
- the one or more instructions may be executable by processor(s), specialized processors, or DSP(s).
- the memory may comprise a non-transitory processor-readable memory and/or a computer-readable memory that stores software code (programming code, instructions, etc.) that are executable by processor(s) and/or DSP(s) to perform functions described herein.
- the wireless transceiver may communicate with processor(s) and/or DSP(s) through the bus to enable the mobile device to be configured as a wireless station as discussed above.
- Processor(s) and/or DSP(s) may execute instructions to execute one or more aspects of processes/methods discussed above in connection with FIG. 1 to FIG. 6A-6C and FIG. 8A-8E to FIG. 17 .
- a user interface may comprise any one of several devices such as, for example, a speaker, microphone, display device, vibration device, keyboard, touch screen, etc.
- the user interface may enable a user to interact with one or more applications hosted on the mobile device.
- devices of user interface may store analog or digital signals on the memory to be further processed by DSP(s) or processor in response to action from a user.
- applications hosted on the mobile device may store analog or digital signals on the memory to present an output signal to a user.
- the mobile device may optionally include a dedicated audio input/output (I/O) device comprising, for example, a dedicated speaker, microphone, digital to analog circuitry, analog to digital circuitry, amplifiers and/or gain control.
- I/O audio input/output
- the mobile device may comprise touch sensors responsive to touching or pressure on a keyboard or touch screen device.
- the mobile device may also comprise a dedicated camera device for capturing still or moving imagery.
- the dedicated camera device may comprise, for example an imaging sensor (e.g., charge coupled device or CMOS imager), lens, analog to digital circuitry, frame buffers, etc.
- additional processing, conditioning, encoding or compression of signals representing captured images may be performed at the processor(s) or DSP(s).
- a dedicated video processor may perform conditioning, encoding, compression or manipulation of signals representing captured images.
- the dedicated video processor may decode/decompress stored image data for presentation on a display device on the mobile device.
- the mobile device may also comprise sensors coupled to the bus which may include, for example, inertial sensors and environment sensors.
- Inertial sensors may comprise, for example accelerometers (e.g., collectively responding to acceleration of the mobile device in three dimensions), one or more gyroscopes or one or more magnetometers (e.g., to support one or more compass applications).
- Environment sensors of the mobile device may comprise, for example, temperature sensors, barometric pressure sensors, ambient light sensors, and camera imagers, microphones, just to name few examples.
- the sensors may generate analog or digital signals that may be stored in memory and processed by DPS(s) or processor(s) in support of one or more applications such as, for example, applications directed to positioning or navigation operations.
- the mobile device may comprise a dedicated modem processor capable of performing baseband processing of signals received and down-converted at a wireless transceiver or SPS receiver.
- the dedicated modem processor may perform baseband processing of signals to be up-converted for transmission by the wireless transceiver.
- baseband processing may be performed by a processor or DSP (e.g., processor(s) or DSP(s)).
- FIG. 7A illustrates a method of capturing and validating a fingerprint image according to aspects of the present disclosure.
- the method determines a set of light sources from a plurality of light sources for emitting light to a fingerprint.
- the method determines a set of sensor zones from a plurality of sensor zones for sensing scattered light from the fingerprint.
- the method emits light from the set of light sources to generate the scattered light from the fingerprint.
- the method senses the scattered light in the set of sensor zones.
- the method may optionally determine a validity of the fingerprint using the scattered light sensed in the plurality of sensor zones. The method may further determine a minimum distance between a light source in the set of light sources and a sensor in the set of sensor zones.
- FIG. 7B illustrates a method of determining a set of light sources from a plurality of light sources for emitting light to a fingerprint according to aspects of the present disclosure.
- the method determines a minimum distance between the set of light sources to avoid interference in the set of sensor zones.
- a sensor zone in the set of sensor zones is determined by a region between an inner radius and an outer radius; where the inner radius is a minimum distance away from the light source where the scattered light is sensed, and the outer radius is a maximum distance away from the light source where the scattered light is sensed.
- the methods performed in block 712 may further include the methods performed in block 714 and block 716 .
- the method determines a separation distance of a buffer zone between the set of light sources beyond the outer radius, wherein intensity of light in the buffer zone is below a predetermined threshold value. In block 714 , the method computes the minimum distance to be twice the outer radius plus the separation distance of the buffer zone.
- the light from the plurality of light sources are emitted from a plurality of pixels or sub-pixels of a pixel panel. The scattered light from the fingerprint is sensed by a plurality of pixels or sub-pixels of a pixel panel.
- FIG. 7C illustrates a method of sensing scattered light in a set of sensor zones according to aspects of the present disclosure.
- the method sets one or more unit cells of the pixel panel in the sensor zone under a reversed bias condition.
- the method detects a leakage current corresponding to the scattered light sensed in the one or more of the pixels or sub-pixels in the one or more unit cells of the pixel panel.
- the method amplifies a signal of the leakage current detected from the one or more of the pixels or sub-pixels.
- the one or more of the pixels or sub-pixels in the sensor zone can be arranged in the form of a line, the one or more of the pixels or sub-pixels in the sensor zone are arranged in the form of a cross, or any combination thereof.
- FIG. 7D illustrates a method of determining a validity of the fingerprint using the scattered light sensed in a plurality of sensor zones according to aspects of the present disclosure.
- the method gathers intensities of the scattered light sensed in the plurality of sensor zones.
- the method determines topography of the fingerprint based on the intensities of the scattered light sensed in the plurality of sensor zones.
- the method compares the topography of the fingerprint to a database of fingerprints to determine whether a match of the fingerprint is found in the database.
- FIG. 7E illustrates a further method of capturing and validating the fingerprint image of FIG. 7A according to aspects of the present disclosure.
- the method determines a next set of light sources from a plurality of light sources for emitting light to a fingerprint.
- the method determines a next set of sensor zones from a plurality of sensor zones for sensing a scattered light from the fingerprint.
- the method emits light from the next set of light sources to generate the scattered light from the fingerprint.
- the method senses the scattered light in the next set of sensor zones.
- the method may repeat the methods performed in block 742 to block 748 until an area under the fingerprint is covered.
- the invention can be implemented in any suitable form, including hardware, software, firmware, or any combination of these.
- the invention may optionally be implemented partly as computer software running on one or more data processors and/or digital signal processors.
- the elements and components of an embodiment of the invention may be physically, functionally, and logically implemented in any suitable way. Indeed, the functionality may be implemented in a single unit, in a plurality of units, or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units and processors.
Landscapes
- Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Image Input (AREA)
Abstract
Description
- This application claims the benefit of U.S. provisional application No. 62/008,242, “Fingerprint Sensing Apparatus,” filed Jun. 5, 2014, assigned to the assignee hereof. The aforementioned United States application is hereby incorporated by reference in its entirety.
- The present invention relates to the field of fingerprint detection system. In particular, the present invention relates to a thin high contrast optical acquisition system for fingerprint recognition.
- Various fingerprint recognition devices have been used for commercial and military applications. Many of such conventional fingerprint recognition devices focus on matching patterns of a captured fingerprint to an existing database of fingerprint samples. If a match is found, the fingerprint is deemed to be valid, but if a match is not found, the fingerprint is deemed to be invalid.
- Applications developed for mobile devices and platforms usually require secure electronic transactions, and fingerprint recognition is starting to play an important role for making such transactions more secure as well as convenient. One of the drawbacks of using conventional fingerprint recognition devices for mobile applications is that their physical form factor is not suitable for mobile devices and mobile platforms. Either the device must be attached to the mobile device or embedded within the device. Conventional optical fingerprint acquisition devices can be bulky and it can be challenging to seamlessly integrate with mobile devices.
- Therefore, it is desirable to have a thin high contrast optical acquisition system for fingerprint recognition that can address the above issues of conventional fingerprint recognition devices.
- Methods and systems are provided for a fingerprint sensing apparatus. In one embodiment, a method for determining validity of a fingerprint includes determining a set of light sources from a plurality of light sources for emitting light to a fingerprint, determining a set of sensor zones from a plurality of sensor zones for sensing scattered light from the fingerprint, determining a minimum distance between a sensor zone and a light source for sensing selective scattered light from the fingerprint, emitting light from the set of light sources to generate the scattered light from the fingerprint, sensing the scattered light in the set of sensor zones, and determining a validity of the fingerprint using the scattered light sensed in the plurality of sensor zones.
- In another embodiment, an apparatus for determining validity of a fingerprint includes a light refracting device (light refractor), a light source, a light collecting device, and a controller. The light refracting device can, for example, be an active matrix organic light emitting diode (AMOLED) panel structure with reverse current measurement and amplification circuitry, and includes an imaging surface and a viewing plane. Incident light from the light source is projected directly or indirectly onto the imaging surface to create an image of the patterned object from the projected light onto the viewing plane. The apparatus is configured to have a thin form factor, which may be flexible or conformable, compared to conventional optical fingerprint acquisition apparatuses. The AMOLED panel includes the light source panel as well as light collecting devices. The fingerprint sensing apparatus can be implemented as an in-cell structure.
- In another embodiment, an apparatus for determining validity of a fingerprint includes a light refracting device (light refractor), a light source, a light collecting device, and a controller. The light refracting device can, for example, be a thin-film transistor (TFT) panel and includes an imaging surface, a light receiving surface, a viewing plane, and light collecting devices. The light source can be an individually addressable panel of discrete light sources, for example, a liquid crystal display (LCD) panel or an AMOLED panel. Incident light from the light source is projected through the light receiving surface and projected directly or indirectly onto the imaging surface to create an image of the patterned object from the projected light onto the viewing plane. The apparatus is configured to have a thin form factor, which may be flexible or conformable, compared to conventional optical fingerprint acquisition apparatuses. The TFT panel could be implemented as an add-on panel that is placed on top of the light source panel.
- The aforementioned features and advantages of the disclosure, as well as additional features and advantages thereof, will be more clearly understandable after reading detailed descriptions of embodiments of the disclosure in conjunction with the non-limiting and non-exhaustive aspects of following drawings. Like numbers are used throughout the figures.
-
FIG. 1A illustrates light characteristics based on illumination of a ridge of a fingerprint according to aspects of the present disclosure.FIG. 1B illustrates light characteristics based on illumination of a valley of a fingerprint according to aspects of the present disclosure. -
FIG. 2 illustrates a potential issue of interferences from two light sources according to aspects of the present disclosure. -
FIG. 3A illustrates an exemplary method of avoiding interferences from two light sources with respect to scattered light from ridges of a fingerprint according to aspects of the present disclosure.FIG. 3B illustrates an exemplary method of avoiding interferences from two light sources with respect to scattered light from a fingerprint according to aspects of the present disclosure.FIG. 3C illustrates another exemplary method of avoiding interferences from two light sources with respect to scattered light from a fingerprint according to aspects of the present disclosure.FIG. 3D illustrates a plot of light intensity overlap for the example ofFIG. 3C according to aspects of the present disclosure. -
FIG. 4A illustrates a side view of the exemplary method ofFIG. 3A according to aspects of the present disclosure.FIG. 4B illustrates a top view of the exemplary method ofFIG. 3A according to aspects of the present disclosure. -
FIG. 5A illustrates an exemplary implementation of addressing a plurality of non-overlapping light sources and sensor zones according to aspects of the present disclosure. FIG. 5B illustrates another exemplary implementation of addressing a plurality of non-overlapping light sources and sensor zones according to aspects of the present disclosure. -
FIG. 6A illustrates an exemplary implementation of scanning a fingerprint according to embodiments of the present disclosure.FIG. 6B illustrates another exemplary implementation of scanning a fingerprint according to embodiments of the present disclosure.FIG. 6C illustrates exemplary separations between adjacent sensor zones for avoiding interferences according to aspects of the present disclosure. -
FIG. 7A illustrates a method of capturing and validating a fingerprint image;FIG. 7B illustrates a method of determining a set of light sources from a plurality of light sources for emitting light to a fingerprint;FIG. 7C illustrates a method of sensing scattered light in a set of sensor zones;FIG. 7D illustrates a method of determining a validity of the fingerprint using the scattered light sensed in a plurality of sensor zones; andFIG. 7E illustrates a further method of capturing and validating the fingerprint image ofFIG. 7A according to aspects of the present disclosure. -
FIG. 8A illustrates an exemplary implementation of collecting sensor data according to aspects of the present disclosure.FIG. 8B illustrates another exemplary implementation of collecting sensor data according to aspects of the present disclosure.FIG. 8C illustrates yet another exemplary implementation of collecting sensor data according to aspects of the present disclosure.FIG. 8D illustrates yet another exemplary implementation of collecting sensor data according to aspects of the present disclosure.FIG. 8E illustrates yet another exemplary implementation of collecting sensor data according to aspects of the present disclosure. -
FIG. 9 illustrates effects of controlling a light emission cone on a corresponding active sensor zone according to aspects of the present disclosure. -
FIG. 10A illustrates a unidirectional OLED with fingerprint sensors according to aspects of the present disclosure;FIG. 10B illustrates a bi-directional OLED with fingerprint sensors according to aspects of the present disclosure;FIG. 10C illustrates a conformable OLED with fingerprint sensors according to aspects of the present disclosure;FIG. 10D illustrates an exemplary OLED structure according to aspects of the present disclosure. -
FIG. 11A illustrates an exemplary subpixel circuit cell with forward bias according to aspects of the present disclosure;FIG. 11B illustrates an exemplary subpixel circuit cell with reverse bias according to aspects of the present disclosure. -
FIG. 12 illustrates an exemplary pixel circuit cell with RGB subpixels according to aspects of the present disclosure. -
FIG. 13 illustrates a process of fingerprint acquisition using AMOLED according to aspects of the present disclosure. -
FIG. 14 illustrates an exemplary controller of a thin high contrast optical acquisition system for fingerprint recognition according to aspects of the present disclosure. -
FIG. 15 illustrates an exemplary light sensing panel using a thin film transistor (TFT) panel structure according to aspects of the present disclosure. -
FIG. 16 illustrates an exemplary fingerprint detection system according to aspects of the present disclosure. -
FIG. 17 illustrates an exemplary panel structure according to aspects of the present disclosure. - Methods and systems are provided for fingerprint sensing apparatus. The following descriptions are presented to enable any person skilled in the art to make and use the disclosure. Descriptions of specific embodiments and applications are provided only as examples. Various modifications and combinations of the examples described herein will be readily apparent to those skilled in the art, and the general principles defined herein may be applied to other examples and applications without departing from the scope of the disclosure. Thus, the present disclosure is not intended to be limited to the examples described and shown, but is to be accorded the scope consistent with the principles and features disclosed herein. The word “exemplary” or “example” is used herein to mean “serving as an example, instance, or illustration.” Any aspect or embodiment described herein as “exemplary” or as an “example” in not necessarily to be construed as preferred or advantageous over other aspects or embodiments.
- Some portions of the detailed description that follows are presented in terms of flowcharts, logic blocks, and other symbolic representations of operations on information that can be performed on a computer system. A procedure, computer-executed step, logic block, process, etc., is here conceived to be a self-consistent sequence of one or more steps or instructions leading to a desired result. The steps are those utilizing physical manipulations of physical quantities. These quantities can take the form of electrical, magnetic, or radio signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. These signals may be referred to at times as bits, values, elements, symbols, characters, terms, numbers, or the like. Each step may be performed by hardware, software, firmware, or combinations thereof.
- The drawing of the fingerprint acquisition apparatus, as a pixel panel, in
FIG. 1A shows incident light traveling from a light source component L in the light emission plane, through an optical structure to the imaging surface, and from the imaging surface to a light collecting (sensor) component S in the light emission plane. Incident light interacts with a finger placed on the imaging surface and is reflected, refracted, or scattered from the imaging surface. The pixel panel is configured so that there are multiple lighting pixels, to serve as discrete light sources, and multiple sensor pixels, to receive light carrying fingerprint image information. The light components used as light sources or sensors can be either pixels or subpixels that are arranged in an addressable pattern in the light emission plane. The pixel panel can be a display panel or non-display panel but is not limited LED or TFT-type display panels. - For a TFT type of light refractor implemented as an add-on panel, although the light source components are located in the light source panel below the light refractor, the light emission plane can be considered as the same plane as the light collecting plane because the transparent areas of the cells of the TFT panel are aligned with the transparent areas of the cells of the light source panel. For an LED type of light refractor implemented in-cell, both the light emission plane and light collecting plane are one and the same.
- The following sections describe how fingerprint images can be captured according to aspects of the present disclosure. In one approach, light is projected from one light source L onto the imaging surface at an incident angle that is less than the critical angle at the light refractor/air interface. When incident light strikes the imaging surface where a finger is placed, it will get scattered by both ridges and valleys albeit to differing degrees and intensities. Ridges will scatter light across a wider range of angles compared to the valleys. More specifically, light scattered from ridges will hit the emission plane over a larger area compared to the area hit by light scattered from valleys. The sensor zone of the present disclosure can be determined in view of the characteristics of the light scattered, and it can be configured to differentiate the light scattered by the two topographical features.
- The difference in scattering range is due to the presence of air where there are valleys at the imaging surface. Since the angle of incident light is less than the critical angle, light passes out of the pixel panel, through the imaging surface into the air, strikes the valley, scatters, re-enters the pixel panel, and refracts. Because of the refraction, valleys scatter light onto the emission plane within a smaller area than that of the ridges (for example, within radius r0 as shown in
FIG. 4A ). Both valleys and ridges can scatter light into the same small area; there is no differentiation between light scattered from ridges and valleys, therefore, light is not collected in the small area close to the light source. But the ridges can scatter light outside of the small area; therefore, light is collected outside of the small area. The pixel panel provides a plurality of sensors across a controllable area to gather the well-differentiated light that is used to generate high-contrast fingerprint images. - In the example shown in
FIG. 1A , the light strikes where a ridge is touching the imaging surface. Part of the incident light is absorbed by the ridge while the rest of the light is scattered towards the light emission plane. A sensor S receives a ray of the scattered light. The intensity of light received by the sensor S is assigned to the position that corresponds to the point of illumination by incident light from the light source L. The image of ridges will appear bright. - In the example shown in
FIG. 1B , the light strikes the surface where there is a valley. The light passes through the surface and strikes the valley. Part of the light is absorbed by the valley while the rest of the light is scattered. Part of the scattered light re-enters the optical structure through the imaging surface and is refracted having an angle αR normal to the imaging surface. A ray of refracted light hits the light emission plane at a distance x from the light source. This distance x is limited by the critical angle at the light refractor/air interface; x also depends on the depth of the valley and the thickness of the optical structure (th). Beyond a certain distance from the light source, x(max), no light will be refracted onto the light emission plane because the angle of refraction is limited by the critical angle. Stray light might reach beyond x(max) but it can be ignored as its intensity would be negligible. The intensity of light received by the sensor S is assigned to the position that corresponds to the point of illumination by incident light from the light source L; therefore, a sensor located beyond x(max) would register an image of the valley as being dark. - To distinguish between a fingerprint ridge or valley using a single light source L, the optical imaging apparatus can be configured to address individual, multiple, or regions of pixels such that when L is activated. All sensors S may be located within a predetermined distance from L at xS, e.g. xS<x(max), can be ignored, not used, or turned off, and at least one sensor S may be located at a predetermined distance from L at xS, e.g. xS>x(max), can be turned on, and incoming light is collected; such a sensor is called an active sensor. The active sensor will receive light scattered only from a ridge because no light scattered from a valley can be refracted by the optical structure beyond the critical angle, or beyond the distance x(max) except for negligible stray light. The following paragraphs describe a method to calculate a value for x(max) where there is a valley at the point of illumination.
- After incident light passes through the imaging surface, it strikes the valley in air and scatters. Some of this scattered light re-enters the optical structure at a distance x(air) from the axis of the original incident light and refracts in the optical structure, making an angle, αR, normal to the imaging surface. X(max) can be computed as: x(max)=x(air)+thickness (th)*tan(αR).
- The largest possible value for αR is the critical angle at the light refractor/air interface, although αR is likely to be smaller than the critical angle in practice. The largest value for x(air) is approximately the width of the valley w. Therefore, x(max) may be estimated as with the following expression: x(max)=w+thickness*tan(critical angle).
- The fingerprint acquisition apparatus can be designed to have a larger x(max) to reduce detection of stray light, which would reduce the sensitivity but would be offset by greater clarity and differentiation of fingerprint ridges and valleys. In some applications, however, a smaller x(max) can be chosen to boost sensitivity without lowering contrast to intolerable levels. Since the width of a valley can vary from one point of illumination to another, in some other applications, a large enough value of x(max) is estimated to cover the possible range of valley widths that may be encountered in such applications. If there are two or more active sensors, the intensity of light received by one sensor can be compared with that received by another sensor to calculate an overall brightness that corresponds to the point of illumination.
- The measured or calculated intensities of light received from various points of illumination can be assembled to map out the topography of the fingerprint corresponding to each point of illumination. Brighter areas indicate higher elevation (e.g. ridges) while darker areas indicate lower elevation (e.g. valleys).
- According to aspects of the present disclosure, scattered light could be collected at a large distance beyond x (max). But it may not be practical to do so too far away for the following reasons: (a) the light intensity drops the farther away from the point of illumination, which would reduce sensitivity; (b) excessive redundant data would be generated, which would waste storage and processing resources; and/or (c) it may be advantageous to use a plurality of simultaneous light sources to speed up data collection. For faster scanning of a whole fingerprint, one approach is to illuminate more than one point from more than one light source at a time.
- In
FIG. 2 , light is projected simultaneously from two light sources, separated by a distance xL, onto the imaging surface, each having an incident angle that is less than the critical angle at the light refractor/air interface. Light source L1 is located too close to light source L2 causing interference: one sensor receives light from rays scattered from two points of illumination, such as two ridges. To avoid the problems of imaging artifacts and low contrast caused by interference, xL can be large enough so that an active sensor does not receive overlapping light from rays scattered from two separate points of illumination as shown inFIG. 3A andFIG. 3B . In these figures, sensors located at a distance from L less than x(max) are ignored, not used, or turned off, and at least one sensor located at a distance from light source, between r0 and r, where r0>x(max) and r>r0, is an active sensor (FIG. 4A andFIG. 4B ). - The value for r can be chosen as the distance where scattered light intensity hitting the emission plane decreases to a desired threshold value, such as <20%, <10%, <5%, 1%, or <1% of the scattered light intensity hitting the emission plane at r0. Beyond the distance r, lower intensity scattered light could contribute to interference if xL(min) is not large enough. In such applications, a buffer zone may be added beyond the active sensor(s) to avoid interference at the active sensor(s). For some other applications, if such lower intensity light is negligible, then interference may be tolerated. For some other applications, r can be chosen as small as slightly larger than r0. According to aspects of the present disclosure, xL(min) is referred to as the minimum distance between two light sources that can be simultaneously illuminated such that at least one active sensor will not receive interference from light reflected, refracted, or scattered from the point of illumination by the other light source.
- According to aspects of the present disclosure, xL(min) can be estimated by using a number of models for attenuation of light intensity. In one approach, the minimum distance may be computed as xL(min)=2*width of ridge+a function(1/xS 2). xS is the distance from a sensor to its corresponding light source L. If assuming a hemispherical model for attenuation, the intensity will decrease by 1/xS 2. αS is the angle that a ray of scattered light received by sensor S makes with the normal to the light emission plane. If the angle of incidence to the normal of imaging surface, αi, is 0 degrees, then xS=th*tan(αs). The function determines how far xS can be where the intensity of light becomes negligible. To improve the contrast of the fingerprint image, the optical imaging apparatus can be configured so that light from two points of illumination do not overlap. To do so, xL(min) needs to be large enough but not too large as to make it impractical. xL(min) can be small enough that allow overlap of negligible light. A threshold can be determined for negligible light intensity.
- In another approach, to avoid the worst case scenario of having interference from light scattered by two ridges, the minimum distance may be computed as xL(min)=2*r+xB. xB represents the width of a buffer zone, located between neighboring active sensor zones, in which scattered light is not sensed, the intensity of light becomes negligible, or scattered light from two points of illumination can overlap without causing interference that degrades the image. The threshold where the intensity of light becomes negligible can be determined for a particular design requirement. For example, the intensity may become negligible when it becomes <30%, <20%, <10%, <5%, 1%, or <1% of the intensity of light at r0. To improve the contrast of the fingerprint image, the optical imaging apparatus can be configured so that light scattered from two points of illumination and collected in their respective active sensor zones do not overlap. To do so, xL(min) needs to be large enough but not too large as to make it impractical. xL(min) can be small enough to allow a tolerable overlap of negligible light for certain applications (such as shown in
FIG. 5A ).FIG. 3C illustrates another exemplary method of avoiding interferences from two light sources with respect to scattered light from a fingerprint according to aspects of the present disclosure.FIG. 3D illustrates a plot of light intensity overlap for the example ofFIG. 3C according to aspects of the present disclosure. - An active sensor zone may include one light source and at least one active sensor that will not receive interference from a point of illumination from another light source. The active sensor zone can be a circular region surrounding one light source L circumscribed by a radius r. This zone includes at least one active sensor located inside of a circular ring having an inner radius of r0>x(max) and an outer radius of r>r0 (
FIG. 4A andFIG. 4B ). - To reduce the time needed to obtain a complete image of a fingerprint, multiple, non-overlapping active sensor zones can be activated simultaneously. The buffer zones may overlap if the interference due to the overlap can be tolerated by the given application. Examples of multiple active sensor zones are shown in arrays of different orientations in
FIG. 5A andFIG. 5B . After one set of zones is illuminated and light information is collected, that set is deactivated, and another set of zones is activated in increments in space and time such as in a series of horizontal and vertical scans, or a combination of the two different scanning methods (FIG. 6A andFIG. 6B ). This scanning process is repeated until the desired amount of information of the fingerprint is collected to assemble an image. The image quality and scanning/processing speed can be optimized by adjusting the arrangement of zones. - The activated sensor zones such as in
FIG. 5A can be arranged across the entire imaging area in rows, columns, or other conformations that allow for efficient addressing of the pixels and data collection.FIG. 5B shows an example of tighter packing of sensor zones for improved efficiency of fingerprint scanning. Space can be added between activated sensor zones such as inFIG. 6C to reduce leakage of stray light from one zone to another zone that may cause interference or introduce artifacts. - To further reduce the data collection and processing time, the number of activated sensors within an active sensor zone may be reduced to form a line as in
FIG. 8A or a cross such as inFIG. 8D . Removing sensors that are close to neighboring active sensor zones can also reduce possible interference from the neighboring zones. The line or cross shaped zones can be arranged into arrays such as inFIGS. 8B , 8C, and 8E that can help to further reduce the possibility of overlap and interference while allowing for tighter packing for improved efficiency of fingerprint scanning. - To decrease the active sensor zone size without compromising the high contrast and image clarity, the thickness of the optical structure can be reduced. This can reduce the active sensor zone radius, r, which allows for either a more compact array of zones which results in faster scanning or a larger gap between zones which produces a higher scanning quality.
- In
FIG. 9 , the light emitted from a light source may leave the pixel at an emission angle ranging from zero to β from the normal of the light emission plane. As shown inFIG. 9 , a cone of light shown as β can strike the imaging surface, effectively illuminating a circular area. The angle of incidence at the imaging surface, αi, would range from zero to β, creating a circular area of illumination. The light scattered by a ridge in this circular area would widen the radius of an active sensor zone from r to r′ and possibly the x(max) so that r0′>r0. The active sensor zones can be spaced further apart to account for a non-zero value of beta. Overlapping reflected, refracted, or scattered light from different areas of a topographical feature may degrade the clarity of the fingerprint image. To reduce this degradation, the optical apparatus can be configured to reduce β and reduce the thickness of the optical structure, either of which can help to reduce the area of illumination. - To maintain high contrast, β may not be greater than the critical angle at the light refractor/air interface. If β cannot be reduced, then the material of the light refractor can be selected to increase the critical angle. If the rings of active sensor zones are widened due to a non-zero β, then it would preferable to activate more sensors within each ring to increase sensitivity. Alternatively, when generating the fingerprint image, the light received by sensors closer to the inner-most ring near r0 can be used, while the light received by the remaining sensors can be discarded or ignored. Using the methodologies described above, the image of a fingerprint can have a high contrast, which is important for accurate fingerprint recognition.
- An add-on type of light refractor can be flexible, conformable, or transparent. One example of such an add-on light refractor can be a TFT panel structure including light sensing components on the bottom of the panel, beneath a substrate. The substrate can be, for example, glass, plastic, or polymer. Together with the substrate, the TFT panel structure may be configured to function as a light sensing panel.
-
FIG. 15 illustrates an exemplary light sensing panel using a thin film transistor (TFT) panel structure according to aspects of the present disclosure. Each cell of the TFT panel structure can be an addressable light sensing component, referred to as a sensing pixel. In the example shown inFIG. 15 ,capture sensor 1500 includes apassivation layer 1518, which can be formed of SiNx. On top ofpassivation layer 1518, a storage capacitor layer is formed includingfirst electrode 1515. This storage capacitor layer is preferably formed from indium tin oxide (ITO), which is conductive and transparent. On top offirst electrode 1515, an insulatinglayer 1517 is formed, preferably of SiNx. Over insulatinglayer 1517, asecond electrode 1514 is formed, preferably of tin oxide.First electrode 1515, insulatinglayer 1517 andsecond electrode 1514 together form the storage capacitor. Oversecond electrode 1514, another insulatinglayer 1516 is formed, which can be formed from SiNx. A layer ofglass layer 1511 is placed over insulatinglayer 1516. A fingerprint to be imaged is placed onglass layer 1511, which may be referred to herein as the imaging surface. - A
light sensing unit 1512, which is preferably a thin-film transistor, and aswitching unit 1513, which is also preferably a thin-film transistor, are horizontally arranged on apassivation layer 1518. Underpassivation layer 1518, a back light 1520 irradiates light upward to be passed through thefingerprint capture sensor 1500. As shown inFIG. 15 , back light 1520 can be separated from a lower, exposed surface ofpas sivation layer 1518. It is also considered, however, thatbacklight 1520 be placed against lower surface ofpassivation layer 1518.Backlight 1520 can be an LED or any other type of light source. A source electrode 1512-S of thelight sensing unit 1512 and a drain electrode 1513-D of theswitching unit 1513 are electrically connected throughsecond electrode 1514. A gate electrode 1512-G of thelight sensing unit 1512 is connected tofirst electrode 1515. Additionally, a first light shielding layer 1513-sh is placed between insulatinglayer 1517 andpassivation layer 1518 at switchingunit 1513. As detailed below, first light shielding layer 1513-sh blocks light frombacklight 1520 from reachingswitching unit 1513. Additionally, secondlight shielding layer 1522 is positioned betweenglass layer 1511 and insulatinglayer 1516 at switchingunit 1513 to shield switchingunit 1513 from light passing through or reflected fromglass layer 1511. - In the above structure, a photosensitive layer 1512-P such as amorphous silicon (a-Si:H) is formed between the drain electrode 1512-D and source electrode 1512-S of the
light sensing unit 1512. Note that photosensitive layer 1512-P allows current to flow in response to a predetermined amount of light striking a surface of photosensitive layer 1512-P. In this way, when more than a predetermined quantity of light is received at a surface of photosensitive layer 1512-P, current flows through the drain electrode 1512-D and the source electrode 1512-S. - According to aspects of the present disclosure, in a method of fabricating
capture sensor 1500, a secondlight shielding layer 1522 is first placed onglass layer 1511 via evaporation, sputtering or any other method.Glass layer 1511 is preferably between about 5 and 10 um, though may be either thicker or thinner.Light shielding layer 1522 is preferably formed from a metal such as aluminum, but may be formed from any suitable light blocking material. Next, insulatinglayer 1516 is formed on top ofglass layer 1511 and secondlight shielding layer 1522. As noted above, insulatinglayer 1516 is preferably formed from SiNx. Photosensitive layer 1512-P is then formed over insulating layer 116. As discussed above, photosensitive layer 1512-P is preferably formed from a-Si:H. Source electrode 1512-D oflight sensing unit 1512,second electrode 1514 and drain electrode 1513-D of switchingunit 1513 are next formed overinsulating layer 1516. Source electrode 1512-D,second electrode 1514 and drain electrode 1513-D are each preferably formed of ITO, but may be formed of any suitable conductor. Next, insulatinglayer 1517 is formed and over insulatinglayer 1517first electrode 1515 is formed. Insulatinglayer 1517 is preferably formed from SiNx andfirst electrode 1515 is preferably formed of ITO but may be formed of any suitable conductor. Next, gate electrode 1512-G oflight sensing unit 1512 and light shield 113-sh are formed. Preferably, gate electrode 1512-G and light shielding layer 1513-sh are each formed of ITO, but may be formed of any suitable material and light shielding layer 1513-sh does not need to be formed from the same material as gate electrode 1512-G. Next,passivation layer 1518, which is preferably formed from SiNx, is formed overfirst electrode 1515, gate electrode 1512-G and light shielding layer 1513-sh. As discussed above,backlight 1520 can either be attached to the lower, exposed surface ofpassivation layer 1518 or separately supported. - In another implementation, an image capture sensor can have substantially the same structure as capture sensor shown in
FIG. 15 except that conductive ITO layer is placed beneath glass layer and an insulating layer, which can be formed of SiNx, is placed below ITO layer. Because ITO layer is conductive, electrostatic charge built up on glass layer can be discharged by connecting ITO layer to a ground. This can prevent damage to capture sensor. Image capture sensor can be fabricated in substantially the same manner as image capture sensor except that ITO layer is formed over glass layer and insulating layer is formed over ITO layer prior to forming light shielding layer over insulating layer. - In yet another implementation, an image capture sensor can have substantially the same structure as capture sensor shown in
FIG. 15 . Specifically, the capture sensor includes a light sensing unit, which is substantially the same and light sensing unit, and switching unit, which is substantially the same as switching unit, formed between an insulating layer and a passivation layer. However, above insulating layer capture sensor includes a substrate layer having a plurality of fiber-optic strands running in a direction perpendicular to a surface of substrate layer. Preferably, the diameter of the fiber-optic strands 330 a forming substrate layer is from about 4 um to about 8 um in diameter and more preferably about 6 um in diameter, though larger or smaller diameters can also be used. Substrate layer can be formed from glass fiber optic strands 330 a or fiber optic strands of other substantially transparent materials including polymers. Fiber optic sheets can be used to form the substrate layer. - A light sensor panel can be implemented as an add-on panel that is placed on top of a light source panel. The light source panel can be, for example, an LCD panel or an AMOLED panel.
FIG. 16 illustrates an exemplary fingerprint detection system, in which a TFT type light sensor panel is placed on top of an LCD display panel structure as an add-on panel. In this example, a TFT type light sensor panel is placed on top of an LCD panel structure as an add-on panel. The sensing pixels of the TFT type light sensing panel can be individually addressable and can be activated according to a designated sensor zone pattern. - If there are non-transparent areas in the light sensor panel, these areas can be aligned with the non-transparent areas of the light source panel.
FIG. 17 illustrates an exemplary TFT light sensor panel aligned with an LCD panel structure, wherein non-transparent components of the TFT light sensor panel are aligned with the black matrix areas of the LCD display panel structure. In this example, the TFT light sensor panel is aligned with the LCD panel structure. The non-transparent components on the TFT light sensor panel are aligned with the black matrix area on the LCD display panel structure. - The black matrix areas of the LCD display panel are non-transparent and therefore would block the transmission of the display backlight. The light sensor panel can be designed so that its non-transparent areas can be aligned with the black matrix areas of the LCD panel. When the LCD display emits light through the transparent areas of the LCD display, this light can be used as the light source for the light sensor panel. The LCD display can individually control cells (individually addressable) to emit light as discrete light sources that are projected into the light refractor according to a designated illumination pattern.
- As described above, the light refracting device can, for example, also be a thin-film transistor (TFT) add-on panel placed on top of an LCD or AMOLED display panel structure that acts as a panel of light sources. Incident light from the light source panel is projected through the light receiving surface and projected directly or indirectly onto the imaging surface to create an image of the patterned object from the projected light onto the viewing plane. This fingerprint sensing apparatus can be also used as a touch sensor when implemented in a mobile device.
- According to aspects of the present disclosure, both top emission and bottom emission type OLED structures (
FIG. 10A ) can be used as the main component of a fingerprint acquisition apparatus. Several different types of OLED devices, such as small molecule OLED, polymer OLED, or solution based OLED, may be utilized as main OLED device structures. Both transparent and non-transparent OLED panels can be used as the main component of a fingerprint acquisition apparatus. (FIG. 10B ) Both thin panel and flexible or conformable types of OLED panels can be used as the main component of a fingerprint acquisition apparatus. (FIG. 10C ) - An active matrix OLED (AMOLED) panel can be used as the main component of a fingerprint acquisition apparatus. An AMOLED panel may include subpixel areas (red, green, and blue subpixels) and a driving circuit area (thin film transistor and capacitor). The brightness of each subpixel can be adjusted by the driving and switching transistors and capacitors and by controlling the amount of current injected to the OLED subpixels. The dimension of subpixels can be formed using OLED material deposition techniques. For instance, the size and position of subpixels can be set by using shadow masks during the OLED material evaporation process.
- An OLED may have a layered structure with the following sequence: anode/hole injection layer/hole transport layer/emissive layer/electron transport layer/electron injection layer/cathode. ITO and other transparent conducting materials having high work function can be used for anode materials, and metals such as aluminum and magnesium can be used for cathode materials.
FIG. 10D shows the structure of a bottom emission OLED. In this example, the imaging surface would be at the bottom of the substrate, and the light emission plane would be the cathode layer. The optical structure may include the transparent layers between the substrate and the cathode. - The reliability of such a fingerprint acquisition apparatus, i.e. the OLED panel lifetime, can be improved by using various sealing techniques and materials, such as desiccant, frit glass sealing, and thin film encapsulation. Various types of substrates such as sapphire, glass, and plastic materials can be used for OLED carriers in order to control the light travel path (refractive index control), to enhance/improve signal to noise ratio of image sensing, and to improve the reliability and lifetime of fingerprint apparatus.
FIG. 11A shows an exemplary AMOLED subpixel unit cell circuit (2D-driving TFT circuit with subpixels). The driving area may include a driving transistor, switching transistor, holding capacitor, and reverse current sensor.FIG. 11B shows the reverse current read and amplified in the OLED circuit structure. - In some embodiments, an AMOLED panel has a three-subpixel structure. In the subpixel structure, for example, a blue subpixel can be used as a light source while the neighboring green or red subpixels may be used as a sensor because the band gap of blue subpixels is larger than that of the green or red subpixels.
FIG. 12 shows an exemplary R/G/B pixel structure where the blue subpixel is the light source, and the green or red subpixel is the sensor. The reverse voltage can be biased in the sensor subpixel when the lighting subpixel is turned on. InFIG. 11B , the I-V curves correspond with subpixel structures inFIG. 12 . The amount of reverse current in the sensor subpixel under reverse bias is increased when light is reflected, refracted, or scattered from a fingerprint to the sensor subpixel. The amount of reverse current can be measured using current sensing circuits in the driving circuit area. The reverse current signal can be amplified using an amplification circuit, and/or a signal processor. The amplified current signal can then be processed to generate a fingerprint image by a signal processing algorithm (FIG. 13 , algorithm flow chart). - The OLED panel resolution can be controlled by varying the size and density of each subpixel and by setting the subpixel structure of the OLED panel. For example, an OLED panel may have a larger lighting component (e.g. blue subpixels) and a smaller sensor component (e.g. green and/or red subpixels). According to aspects of the present disclosure, subpixel structures can have different sizes. The subpixel density can be enhanced by changing pixel shape from stripe type to circular or diamond shape. In addition, an OLED subpixel structure can have different shapes, such as square, rectangle, circle, diamond, etc. The patterning of the subpixel structure can be fabricated by using fine metal mask processes, ink-jet printing, or laser transfer technologies.
- According to aspects of the present disclosure, a mobile device is usually equipped with a touch sensor. If a mobile device was equipped with the fingerprint sensing apparatus of the present disclosure, then the touch sensor would not be required, as the fingerprint sensing apparatus may also be used as a touch sensor. As described herein, a mobile device can be configured to include a thin high contrast optical acquisition system as a fingerprint sensing apparatus for fingerprint recognition. In some implementations, the mobile device may comprise a wireless transceiver which is capable of transmitting and receiving wireless signals via wireless antenna over a wireless communication network. Wireless transceiver may be connected to a bus by a wireless transceiver bus interface. The wireless transceiver bus interface may, in some embodiments be at least partially integrated with wireless transceiver. Some embodiments may include multiple wireless transceivers and wireless antennas to enable transmitting and/or receiving signals according to a corresponding multiple wireless communication standards such as, for example, versions of IEEE Std. 802.11, CDMA, WCDMA, LTE, UMTS, GSM, AMPS, Zigbee and Bluetooth®, etc.
- The mobile device may also comprise a SPS receiver capable of receiving and acquiring SPS signals via a SPS antenna. The SPS receiver may also process, in whole or in part, acquired SPS signals for estimating a location of the mobile device. In some embodiments, processor(s), memory, DSP(s) and/or specialized processors (not shown) may also be utilized to process acquired SPS signals, in whole or in part, and/or calculate an estimated location of the mobile device, in conjunction with the SPS receiver. Storage of SPS or other signals for use in performing positioning operations may be performed in memory or registers (not shown).
- In addition, the mobile device may comprise digital signal processor(s) (DSP(s)) connected to the bus by a bus interface, processor(s) connected to the bus by a bus interface and memory. The bus interface may be integrated with the DSP(s), processor(s) and memory. In various embodiments, functions may be performed in response execution of one or more machine-readable instructions stored in memory such as on a computer-readable storage medium, such as RAM, ROM, FLASH, or disc drive, just to name a few example. The one or more instructions may be executable by processor(s), specialized processors, or DSP(s). The memory may comprise a non-transitory processor-readable memory and/or a computer-readable memory that stores software code (programming code, instructions, etc.) that are executable by processor(s) and/or DSP(s) to perform functions described herein. In a particular implementation, the wireless transceiver may communicate with processor(s) and/or DSP(s) through the bus to enable the mobile device to be configured as a wireless station as discussed above. Processor(s) and/or DSP(s) may execute instructions to execute one or more aspects of processes/methods discussed above in connection with
FIG. 1 toFIG. 6A-6C andFIG. 8A-8E toFIG. 17 . - According to aspects of the present disclosure, a user interface may comprise any one of several devices such as, for example, a speaker, microphone, display device, vibration device, keyboard, touch screen, etc. In a particular implementation, the user interface may enable a user to interact with one or more applications hosted on the mobile device. For example, devices of user interface may store analog or digital signals on the memory to be further processed by DSP(s) or processor in response to action from a user. Similarly, applications hosted on the mobile device may store analog or digital signals on the memory to present an output signal to a user. In another implementation, the mobile device may optionally include a dedicated audio input/output (I/O) device comprising, for example, a dedicated speaker, microphone, digital to analog circuitry, analog to digital circuitry, amplifiers and/or gain control. In another implementation, the mobile device may comprise touch sensors responsive to touching or pressure on a keyboard or touch screen device.
- The mobile device may also comprise a dedicated camera device for capturing still or moving imagery. The dedicated camera device may comprise, for example an imaging sensor (e.g., charge coupled device or CMOS imager), lens, analog to digital circuitry, frame buffers, etc. In one implementation, additional processing, conditioning, encoding or compression of signals representing captured images may be performed at the processor(s) or DSP(s). Alternatively, a dedicated video processor may perform conditioning, encoding, compression or manipulation of signals representing captured images. Additionally, the dedicated video processor may decode/decompress stored image data for presentation on a display device on the mobile device.
- The mobile device may also comprise sensors coupled to the bus which may include, for example, inertial sensors and environment sensors. Inertial sensors may comprise, for example accelerometers (e.g., collectively responding to acceleration of the mobile device in three dimensions), one or more gyroscopes or one or more magnetometers (e.g., to support one or more compass applications). Environment sensors of the mobile device may comprise, for example, temperature sensors, barometric pressure sensors, ambient light sensors, and camera imagers, microphones, just to name few examples. The sensors may generate analog or digital signals that may be stored in memory and processed by DPS(s) or processor(s) in support of one or more applications such as, for example, applications directed to positioning or navigation operations.
- In a particular implementation, the mobile device may comprise a dedicated modem processor capable of performing baseband processing of signals received and down-converted at a wireless transceiver or SPS receiver. Similarly, the dedicated modem processor may perform baseband processing of signals to be up-converted for transmission by the wireless transceiver. In alternative implementations, instead of having a dedicated modem processor, baseband processing may be performed by a processor or DSP (e.g., processor(s) or DSP(s)).
-
FIG. 7A illustrates a method of capturing and validating a fingerprint image according to aspects of the present disclosure. In the example shown inFIG. 7A , inblock 702, the method determines a set of light sources from a plurality of light sources for emitting light to a fingerprint. Inblock 704, the method determines a set of sensor zones from a plurality of sensor zones for sensing scattered light from the fingerprint. Inblock 706, the method emits light from the set of light sources to generate the scattered light from the fingerprint. Inblock 708, the method senses the scattered light in the set of sensor zones. Inblock 710, the method may optionally determine a validity of the fingerprint using the scattered light sensed in the plurality of sensor zones. The method may further determine a minimum distance between a light source in the set of light sources and a sensor in the set of sensor zones. -
FIG. 7B illustrates a method of determining a set of light sources from a plurality of light sources for emitting light to a fingerprint according to aspects of the present disclosure. In the example shown inFIG. 7B , inblock 712, the method determines a minimum distance between the set of light sources to avoid interference in the set of sensor zones. In some approaches, a sensor zone in the set of sensor zones is determined by a region between an inner radius and an outer radius; where the inner radius is a minimum distance away from the light source where the scattered light is sensed, and the outer radius is a maximum distance away from the light source where the scattered light is sensed. According to aspects of the present disclosure, the methods performed inblock 712 may further include the methods performed inblock 714 and block 716. Inblock 714, the method determines a separation distance of a buffer zone between the set of light sources beyond the outer radius, wherein intensity of light in the buffer zone is below a predetermined threshold value. Inblock 714, the method computes the minimum distance to be twice the outer radius plus the separation distance of the buffer zone. In some implementations, the light from the plurality of light sources are emitted from a plurality of pixels or sub-pixels of a pixel panel. The scattered light from the fingerprint is sensed by a plurality of pixels or sub-pixels of a pixel panel. -
FIG. 7C illustrates a method of sensing scattered light in a set of sensor zones according to aspects of the present disclosure. In this example, inblock 722, the method sets one or more unit cells of the pixel panel in the sensor zone under a reversed bias condition. Inblock 724, the method detects a leakage current corresponding to the scattered light sensed in the one or more of the pixels or sub-pixels in the one or more unit cells of the pixel panel. Inblock 726, the method amplifies a signal of the leakage current detected from the one or more of the pixels or sub-pixels. Note that in some implementations, the one or more of the pixels or sub-pixels in the sensor zone can be arranged in the form of a line, the one or more of the pixels or sub-pixels in the sensor zone are arranged in the form of a cross, or any combination thereof. -
FIG. 7D illustrates a method of determining a validity of the fingerprint using the scattered light sensed in a plurality of sensor zones according to aspects of the present disclosure. In the exemplary method shown inFIG. 7D , inblock 732, the method gathers intensities of the scattered light sensed in the plurality of sensor zones. Inblock 734, the method determines topography of the fingerprint based on the intensities of the scattered light sensed in the plurality of sensor zones. Inblock 736, the method compares the topography of the fingerprint to a database of fingerprints to determine whether a match of the fingerprint is found in the database. -
FIG. 7E illustrates a further method of capturing and validating the fingerprint image ofFIG. 7A according to aspects of the present disclosure. As shown inFIG. 7E , inblock 742, the method determines a next set of light sources from a plurality of light sources for emitting light to a fingerprint. Inblock 744, the method determines a next set of sensor zones from a plurality of sensor zones for sensing a scattered light from the fingerprint. Inblock 746, the method emits light from the next set of light sources to generate the scattered light from the fingerprint. Inblock 748, the method senses the scattered light in the next set of sensor zones. Inblock 750, the method may repeat the methods performed inblock 742 to block 748 until an area under the fingerprint is covered. - It will be appreciated that the above descriptions for clarity have described embodiments of the invention with reference to different functional units and processors. However, it will be apparent that any suitable distribution of functionality between different functional units or processors may be used without detracting from the invention. For example, functionality illustrated to be performed by separate processors or controllers may be performed by the same processors or controllers. Hence, references to specific functional units are to be seen as references to suitable means for providing the described functionality rather than indicative of a strict logical or physical structure or organization.
- The invention can be implemented in any suitable form, including hardware, software, firmware, or any combination of these. The invention may optionally be implemented partly as computer software running on one or more data processors and/or digital signal processors. The elements and components of an embodiment of the invention may be physically, functionally, and logically implemented in any suitable way. Indeed, the functionality may be implemented in a single unit, in a plurality of units, or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units and processors.
- One skilled in the relevant art will recognize that many possible modifications and combinations of the disclosed embodiments may be used, while still employing the same basic underlying mechanisms and methodologies. The foregoing description, for purposes of explanation, has been written with references to specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described to explain the principles of the invention and their practical applications, and to enable others skilled in the art to best utilize the invention and various embodiments with various modifications as suited to the particular use contemplated.
Claims (23)
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/463,062 US8917387B1 (en) | 2014-06-05 | 2014-08-19 | Fingerprint sensing apparatus |
EP14893840.0A EP3152700B1 (en) | 2014-06-05 | 2014-12-04 | Fingerprint sensing apparatus |
PCT/US2014/068646 WO2015187197A1 (en) | 2014-06-05 | 2014-12-04 | Fingerprint sensing apparatus |
KR1020167033133A KR101900173B1 (en) | 2014-06-05 | 2014-12-04 | Fingerprint sensing apparatus |
CN201480078347.6A CN106233305B (en) | 2014-06-05 | 2014-12-04 | Finger sensor apparatus |
TW103143721A TWI550523B (en) | 2014-06-05 | 2014-12-15 | Fingerprint sensing apparatus |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462008242P | 2014-06-05 | 2014-06-05 | |
US14/463,062 US8917387B1 (en) | 2014-06-05 | 2014-08-19 | Fingerprint sensing apparatus |
Publications (2)
Publication Number | Publication Date |
---|---|
US20140355846A1 true US20140355846A1 (en) | 2014-12-04 |
US8917387B1 US8917387B1 (en) | 2014-12-23 |
Family
ID=51985160
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/463,062 Active US8917387B1 (en) | 2014-06-05 | 2014-08-19 | Fingerprint sensing apparatus |
Country Status (6)
Country | Link |
---|---|
US (1) | US8917387B1 (en) |
EP (1) | EP3152700B1 (en) |
KR (1) | KR101900173B1 (en) |
CN (1) | CN106233305B (en) |
TW (1) | TWI550523B (en) |
WO (1) | WO2015187197A1 (en) |
Cited By (46)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160224816A1 (en) * | 2015-02-02 | 2016-08-04 | Synaptics Incorporated | Optical sensor using collimator |
CN105893950A (en) * | 2016-03-30 | 2016-08-24 | 宁波三博电子科技有限公司 | Adaptive fingerprint identification method and system based on redundancy error sequence ranking algorithm |
CN105913021A (en) * | 2016-01-21 | 2016-08-31 | 友达光电股份有限公司 | Fingerprint sensing device |
US9454259B2 (en) * | 2016-01-04 | 2016-09-27 | Secugen Corporation | Multi-level command sensing apparatus |
WO2017091245A1 (en) * | 2015-11-24 | 2017-06-01 | Secugen Corporation | Fingerprint sensing and calibration apparatus |
US9741290B1 (en) * | 2016-06-30 | 2017-08-22 | Secugen Corporation | Multi-mode display |
US20170344798A1 (en) * | 2016-05-30 | 2017-11-30 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for controlling unlocking and mobile terminal |
WO2018023722A1 (en) * | 2016-08-05 | 2018-02-08 | 上海箩箕技术有限公司 | Display module |
CN107690653A (en) * | 2017-08-18 | 2018-02-13 | 深圳市汇顶科技股份有限公司 | Obtain the method, apparatus and terminal device of fingerprint image |
US9934418B2 (en) | 2015-12-03 | 2018-04-03 | Synaptics Incorporated | Display integrated optical fingerprint sensor with angle limiting reflector |
US10043050B2 (en) | 2016-03-07 | 2018-08-07 | Microsoft Technology Licensing, Llc | Image sensing with a display |
EP3361363A1 (en) * | 2017-01-19 | 2018-08-15 | Samsung Electronics Co., Ltd. | Apparatus and method for recognizing fingerprint |
US10147757B2 (en) | 2015-02-02 | 2018-12-04 | Synaptics Incorporated | Image sensor structures for fingerprint sensing |
US10146258B2 (en) * | 2015-09-30 | 2018-12-04 | Synaptics Incorporated | Optical image sensor for display integration |
WO2018231856A1 (en) * | 2017-06-12 | 2018-12-20 | Synaptics Incorporated | Systems and methods for optical sensing using point-based illumination |
US10169630B2 (en) | 2015-12-03 | 2019-01-01 | Synaptics Incorporated | Optical sensor for integration over a display backplane |
US10176355B2 (en) | 2015-12-03 | 2019-01-08 | Synaptics Incorporated | Optical sensor for integration in a display |
WO2019009986A1 (en) * | 2017-07-06 | 2019-01-10 | Synaptics Incorporated | Optical fingerprint sensor with scattered light image detection |
US10181070B2 (en) | 2015-02-02 | 2019-01-15 | Synaptics Incorporated | Low profile illumination in an optical fingerprint sensor |
US10282582B2 (en) | 2015-09-30 | 2019-05-07 | Apple Inc. | Finger biometric sensor for generating three dimensional fingerprint ridge data and related methods |
CN109923553A (en) * | 2016-11-30 | 2019-06-21 | 株式会社Lg化学 | Method for controlling digital device |
US20190213948A1 (en) * | 2018-01-05 | 2019-07-11 | Boe Technology Group Co., Ltd. | Oled display panel, driving method thereof and display device |
US10360431B2 (en) | 2015-03-25 | 2019-07-23 | Apple Inc. | Electronic device including pin hole array mask above optical image sensor and related methods |
US10387710B2 (en) | 2016-03-07 | 2019-08-20 | Microsoft Technology Licensing, Llc | Image sensing with a waveguide display |
US10388490B1 (en) * | 2018-03-14 | 2019-08-20 | Applied Materials Israel Ltd. | Detection unit, scanning charged particle beam device and a method |
US10430635B2 (en) | 2017-04-18 | 2019-10-01 | Boe Technology Group Co., Ltd. | Fingerprint identification sensor, fingerprint identification method and electronic device |
US10438044B2 (en) | 2016-08-04 | 2019-10-08 | Boe Technology Group Co., Ltd. | Surface texture identification display device |
CN110309794A (en) * | 2019-07-04 | 2019-10-08 | 北京集创北方科技股份有限公司 | Physical characteristics collecting method, collection apparatus equipment and computer readable storage medium |
CN111052141A (en) * | 2019-08-02 | 2020-04-21 | 深圳市汇顶科技股份有限公司 | Fingerprint detection device and electronic equipment |
CN111213152A (en) * | 2019-06-05 | 2020-05-29 | 深圳市汇顶科技股份有限公司 | Optical image acquisition unit, optical image acquisition system, display screen and electronic equipment |
WO2020119135A1 (en) * | 2018-12-14 | 2020-06-18 | Boe Technology Group Co., Ltd. | A driving method and an apparatus for a fingerprint recognition device |
CN111695390A (en) * | 2019-03-15 | 2020-09-22 | 京东方科技集团股份有限公司 | Method for driving grain recognition device and grain recognition device |
CN112036224A (en) * | 2019-06-04 | 2020-12-04 | 联詠科技股份有限公司 | Fingerprint sensing device and method with three-dimensional sensing mechanism |
US10885299B2 (en) | 2016-05-23 | 2021-01-05 | Apple Inc. | Electronic device including pin hole array mask above optical image sensor and laterally adjacent light source and related methods |
WO2021051276A1 (en) | 2019-09-17 | 2021-03-25 | Boe Technology Group Co., Ltd. | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus |
US11080507B2 (en) * | 2019-08-27 | 2021-08-03 | Samsung Display Co., Ltd. | Fingerprint sensing method, fingerprint sensor, and display device including the same |
US11093080B2 (en) | 2018-12-27 | 2021-08-17 | Novatek Microelectronics Corp. | Electronic device and fingerprint sensing control method thereof |
US11126813B2 (en) * | 2019-04-30 | 2021-09-21 | Egis Technology Inc. | Fingerprint sensing device |
US11132527B2 (en) * | 2018-12-14 | 2021-09-28 | Boe Technology Group Co., Ltd. | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus |
US20210326560A1 (en) * | 2018-12-28 | 2021-10-21 | Vivo Mobile Communication Co., Ltd. | Mobile terminal and recognition method |
US11239275B2 (en) | 2016-05-23 | 2022-02-01 | Apple Inc. | Electronic device including processing circuitry for sensing images from spaced apart sub-arrays and related methods |
US20220052128A1 (en) * | 2018-09-10 | 2022-02-17 | Sharp Kabushiki Kaisha | Display device |
US11301707B2 (en) * | 2019-04-30 | 2022-04-12 | Boe Technology Group Co., Ltd. | Texture recognition device and driving method of texture recognition device |
US20220115443A1 (en) * | 2020-10-08 | 2022-04-14 | Guangzhou Tyrafos Semiconductor Technologies Co., Ltd. | Display device and image sensing method thereof |
US11450088B2 (en) | 2019-10-01 | 2022-09-20 | Innolux Corporation | Method of detecting biometric feature |
US11531430B2 (en) | 2019-07-12 | 2022-12-20 | Shenzhen GOODIX Technology Co., Ltd. | Fingerprint detection apparatus and electronic device |
Families Citing this family (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108734074B (en) * | 2017-04-18 | 2022-02-18 | 金佶科技股份有限公司 | Fingerprint identification method and fingerprint identification device |
US10437974B2 (en) | 2015-06-18 | 2019-10-08 | Shenzhen GOODIX Technology Co., Ltd. | Optical sensing performance of under-screen optical sensor module for on-screen fingerprint sensing |
US10726240B2 (en) * | 2015-07-09 | 2020-07-28 | Gingy Technology Inc. | Image capturing apparatus |
DE102015115484C5 (en) | 2015-09-14 | 2019-11-21 | JENETRIC GmbH | Device and method for optical direct recording of living skin areas |
US10115000B2 (en) | 2015-12-11 | 2018-10-30 | Synaptics Incorporated | Method and system for optical imaging using patterned illumination |
KR101924916B1 (en) * | 2016-01-31 | 2018-12-04 | 선전 구딕스 테크놀로지 컴퍼니, 리미티드 | Under-screen optical sensor module for on-screen fingerprint sensing |
US10108841B2 (en) | 2016-03-31 | 2018-10-23 | Synaptics Incorporated | Biometric sensor with diverging optical element |
US10380395B2 (en) | 2016-09-30 | 2019-08-13 | Synaptics Incorporated | Optical sensor with angled reflectors |
US10055637B2 (en) | 2016-12-07 | 2018-08-21 | Synaptics Incorporated | Optical sensor with substrate light filter |
US10311276B2 (en) | 2017-02-22 | 2019-06-04 | Synaptics Incorporated | Under display optical fingerprint sensor arrangement for mitigating moiré effects |
CN106933415B (en) | 2017-03-07 | 2019-11-26 | Oppo广东移动通信有限公司 | A kind of sensor module and terminal |
CN107122742B (en) * | 2017-04-27 | 2019-12-03 | 上海天马微电子有限公司 | Display device, fingerprint identification method thereof and electronic equipment |
CN107145856B (en) | 2017-04-28 | 2021-04-02 | 北京小米移动软件有限公司 | Organic light emitting diode display module and control method thereof |
US11734944B2 (en) | 2017-08-03 | 2023-08-22 | Himax Technologies Limited | Display device with embedded biometric detection function in active region |
KR102452251B1 (en) | 2017-08-04 | 2022-10-11 | 삼성디스플레이 주식회사 | Display apparatus |
US10955603B2 (en) * | 2017-10-17 | 2021-03-23 | Synaptics Incorporated | Method and system for optical imaging using point source illumination |
CN108229394B (en) * | 2018-01-03 | 2020-08-25 | 上海天马微电子有限公司 | Display panel and display device |
KR20200142641A (en) | 2019-06-12 | 2020-12-23 | 삼성디스플레이 주식회사 | Display device |
TWI691909B (en) * | 2019-06-27 | 2020-04-21 | 大陸商北京集創北方科技股份有限公司 | Optical fingerprint collection method, fingerprint sensing device and information processing device |
KR20210018680A (en) | 2019-08-08 | 2021-02-18 | 삼성디스플레이 주식회사 | Display device |
US20220335746A1 (en) * | 2019-08-16 | 2022-10-20 | Egis Technology Inc. | Fingerprint sensing device |
TWI767285B (en) * | 2019-11-29 | 2022-06-11 | 神盾股份有限公司 | Fingerprint identification device, electronic device for identificating fingerprint image and fingerprint identification method |
TWI732400B (en) * | 2019-12-25 | 2021-07-01 | 大陸商北京集創北方科技股份有限公司 | Biological characteristic collection circuit and method, information processing device with the circuit, and information processing device using the method |
US11967172B2 (en) | 2020-11-20 | 2024-04-23 | Image Match Design Inc. | Biometric sensor device with in-glass fingerprint sensor |
US11610426B2 (en) | 2020-11-20 | 2023-03-21 | Image Match Design Inc. | In-glass fingerprint sensor and sensing circuit thereof |
CN112711150B (en) * | 2020-12-22 | 2022-11-22 | 业泓科技(成都)有限公司 | Display device |
TWI805362B (en) * | 2022-05-11 | 2023-06-11 | 速博思股份有限公司 | Wireless fingerprint sensing system |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040252867A1 (en) * | 2000-01-05 | 2004-12-16 | Je-Hsiung Lan | Biometric sensor |
US8254641B2 (en) * | 2007-07-25 | 2012-08-28 | Sony Corportion | Biometrics authentication system |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5929845A (en) * | 1996-09-03 | 1999-07-27 | Motorola, Inc. | Image scanner and display apparatus |
JP3898330B2 (en) * | 1998-03-12 | 2007-03-28 | カシオ計算機株式会社 | Reader |
US6469289B1 (en) * | 2000-01-21 | 2002-10-22 | Symagery Microsystems Inc. | Ambient light detection technique for an imaging array |
US7289649B1 (en) * | 2000-08-10 | 2007-10-30 | Avago Technologies Ecbu Ip (Singapore) Pte. Ltd. | Fingerprint imager |
EP1523807A1 (en) * | 2003-06-30 | 2005-04-20 | Mobisol | Pointing device having fingerprint image recognition function, fingerprint image recognition and pointing method, and method for providing portable terminal service using thereof |
JP2007206991A (en) * | 2006-02-02 | 2007-08-16 | Hitachi Ltd | Bioinformation processor and bioinformation processing program |
US7859526B2 (en) * | 2006-05-01 | 2010-12-28 | Konicek Jeffrey C | Active matrix emissive display and optical scanner system, methods and applications |
US8660324B2 (en) * | 2010-03-29 | 2014-02-25 | Raytheon Company | Textured pattern sensing using partial-coherence speckle interferometry |
US9632344B2 (en) * | 2010-07-09 | 2017-04-25 | Lester F. Ludwig | Use of LED or OLED array to implement integrated combinations of touch screen tactile, touch gesture sensor, color image display, hand-image gesture sensor, document scanner, secure optical data exchange, and fingerprint processing capabilities |
TWI536272B (en) * | 2012-09-27 | 2016-06-01 | 光環科技股份有限公司 | Bio-characteristic verification device and method |
-
2014
- 2014-08-19 US US14/463,062 patent/US8917387B1/en active Active
- 2014-12-04 WO PCT/US2014/068646 patent/WO2015187197A1/en active Application Filing
- 2014-12-04 EP EP14893840.0A patent/EP3152700B1/en active Active
- 2014-12-04 KR KR1020167033133A patent/KR101900173B1/en active IP Right Grant
- 2014-12-04 CN CN201480078347.6A patent/CN106233305B/en active Active
- 2014-12-15 TW TW103143721A patent/TWI550523B/en active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040252867A1 (en) * | 2000-01-05 | 2004-12-16 | Je-Hsiung Lan | Biometric sensor |
US8254641B2 (en) * | 2007-07-25 | 2012-08-28 | Sony Corportion | Biometrics authentication system |
Cited By (73)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11372143B2 (en) | 2015-02-02 | 2022-06-28 | Will Semiconductor (Shanghai) Co. Ltd. | Optical fingerprint sensor |
US10147757B2 (en) | 2015-02-02 | 2018-12-04 | Synaptics Incorporated | Image sensor structures for fingerprint sensing |
US10181070B2 (en) | 2015-02-02 | 2019-01-15 | Synaptics Incorporated | Low profile illumination in an optical fingerprint sensor |
US20160224816A1 (en) * | 2015-02-02 | 2016-08-04 | Synaptics Incorporated | Optical sensor using collimator |
US10705272B2 (en) | 2015-02-02 | 2020-07-07 | Will Semiconductor (Shanghai) Co., Ltd. | Optical fingerprint sensor |
US9829614B2 (en) * | 2015-02-02 | 2017-11-28 | Synaptics Incorporated | Optical sensor using collimator |
US10360431B2 (en) | 2015-03-25 | 2019-07-23 | Apple Inc. | Electronic device including pin hole array mask above optical image sensor and related methods |
US10146258B2 (en) * | 2015-09-30 | 2018-12-04 | Synaptics Incorporated | Optical image sensor for display integration |
US10282582B2 (en) | 2015-09-30 | 2019-05-07 | Apple Inc. | Finger biometric sensor for generating three dimensional fingerprint ridge data and related methods |
CN108292360A (en) * | 2015-11-24 | 2018-07-17 | 斯科竣公司 | Fingerprint sensing and calibration equipment |
EP3380989A4 (en) * | 2015-11-24 | 2019-05-15 | Secugen Corporation | Fingerprint sensing and calibration apparatus |
WO2017091245A1 (en) * | 2015-11-24 | 2017-06-01 | Secugen Corporation | Fingerprint sensing and calibration apparatus |
US11475692B2 (en) | 2015-12-03 | 2022-10-18 | Fingerprint Cards Anacatum Ip Ab | Optical sensor for integration over a display backplane |
US9934418B2 (en) | 2015-12-03 | 2018-04-03 | Synaptics Incorporated | Display integrated optical fingerprint sensor with angle limiting reflector |
US10169630B2 (en) | 2015-12-03 | 2019-01-01 | Synaptics Incorporated | Optical sensor for integration over a display backplane |
US10176355B2 (en) | 2015-12-03 | 2019-01-08 | Synaptics Incorporated | Optical sensor for integration in a display |
US9830009B2 (en) * | 2016-01-04 | 2017-11-28 | Secugen Corporation | Apparatus and method for detecting hovering commands |
US9606672B2 (en) | 2016-01-04 | 2017-03-28 | Secugen Corporation | Methods and apparatuses for user authentication |
US20170010753A1 (en) * | 2016-01-04 | 2017-01-12 | Secugen Corporation | Apparatus and Method for Detecting Hovering Commands |
US9454259B2 (en) * | 2016-01-04 | 2016-09-27 | Secugen Corporation | Multi-level command sensing apparatus |
TWI614694B (en) * | 2016-01-21 | 2018-02-11 | 友達光電股份有限公司 | Fingerprint sensing device |
CN105913021A (en) * | 2016-01-21 | 2016-08-31 | 友达光电股份有限公司 | Fingerprint sensing device |
US10043050B2 (en) | 2016-03-07 | 2018-08-07 | Microsoft Technology Licensing, Llc | Image sensing with a display |
US10817695B2 (en) | 2016-03-07 | 2020-10-27 | Microsoft Technology Licensing, Llc | Image sensing with a display |
US10043051B2 (en) | 2016-03-07 | 2018-08-07 | Microsoft Technology Licensing, Llc | Triggered image sensing with a display |
US10387710B2 (en) | 2016-03-07 | 2019-08-20 | Microsoft Technology Licensing, Llc | Image sensing with a waveguide display |
US10387711B2 (en) | 2016-03-07 | 2019-08-20 | Microsoft Technology Licensing, Llc | Pixel having a photoemitter and a photodetector triggered by a pixel selector signal bus |
CN105893950A (en) * | 2016-03-30 | 2016-08-24 | 宁波三博电子科技有限公司 | Adaptive fingerprint identification method and system based on redundancy error sequence ranking algorithm |
US10885299B2 (en) | 2016-05-23 | 2021-01-05 | Apple Inc. | Electronic device including pin hole array mask above optical image sensor and laterally adjacent light source and related methods |
US11239275B2 (en) | 2016-05-23 | 2022-02-01 | Apple Inc. | Electronic device including processing circuitry for sensing images from spaced apart sub-arrays and related methods |
US20170344798A1 (en) * | 2016-05-30 | 2017-11-30 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for controlling unlocking and mobile terminal |
US10460148B2 (en) * | 2016-05-30 | 2019-10-29 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for controlling unlocking and mobile terminal |
US9741290B1 (en) * | 2016-06-30 | 2017-08-22 | Secugen Corporation | Multi-mode display |
US10438044B2 (en) | 2016-08-04 | 2019-10-08 | Boe Technology Group Co., Ltd. | Surface texture identification display device |
US20190286869A1 (en) * | 2016-08-05 | 2019-09-19 | Shanghai Oxi Technology Co., Ltd | Display Module |
WO2018023722A1 (en) * | 2016-08-05 | 2018-02-08 | 上海箩箕技术有限公司 | Display module |
US10627937B2 (en) * | 2016-08-05 | 2020-04-21 | Shanghai Oxi Technology Co., Ltd | Display module |
CN109923553A (en) * | 2016-11-30 | 2019-06-21 | 株式会社Lg化学 | Method for controlling digital device |
US10586094B2 (en) | 2017-01-19 | 2020-03-10 | Samsung Electronics Co., Ltd. | Apparatus and method for recognizing fingerprint |
EP3361363A1 (en) * | 2017-01-19 | 2018-08-15 | Samsung Electronics Co., Ltd. | Apparatus and method for recognizing fingerprint |
US10430635B2 (en) | 2017-04-18 | 2019-10-01 | Boe Technology Group Co., Ltd. | Fingerprint identification sensor, fingerprint identification method and electronic device |
WO2018231856A1 (en) * | 2017-06-12 | 2018-12-20 | Synaptics Incorporated | Systems and methods for optical sensing using point-based illumination |
US10796123B2 (en) | 2017-06-12 | 2020-10-06 | Will Semiconductor (Shanghai) Co. Ltd. | Systems and methods for optical sensing using point-based illumination |
CN110753999A (en) * | 2017-06-12 | 2020-02-04 | 辛纳普蒂克斯公司 | System and method for optical sensing using point-based illumination |
US10664676B2 (en) | 2017-06-12 | 2020-05-26 | Will Semiconductor (Shanghai) Co. Ltd. | Systems and methods for reducing unwanted reflections in display systems incorporating an under display biometric sensor |
WO2019009986A1 (en) * | 2017-07-06 | 2019-01-10 | Synaptics Incorporated | Optical fingerprint sensor with scattered light image detection |
US10354116B2 (en) | 2017-07-06 | 2019-07-16 | Synaptics Incorporated | Optical fingerprint sensor with scattered light image detection |
CN107690653A (en) * | 2017-08-18 | 2018-02-13 | 深圳市汇顶科技股份有限公司 | Obtain the method, apparatus and terminal device of fingerprint image |
US20190213948A1 (en) * | 2018-01-05 | 2019-07-11 | Boe Technology Group Co., Ltd. | Oled display panel, driving method thereof and display device |
US10713999B2 (en) * | 2018-01-05 | 2020-07-14 | Boe Technology Group Co., Ltd. | OLED display panel, driving method thereof and display device |
US10388490B1 (en) * | 2018-03-14 | 2019-08-20 | Applied Materials Israel Ltd. | Detection unit, scanning charged particle beam device and a method |
US20220052128A1 (en) * | 2018-09-10 | 2022-02-17 | Sharp Kabushiki Kaisha | Display device |
WO2020119135A1 (en) * | 2018-12-14 | 2020-06-18 | Boe Technology Group Co., Ltd. | A driving method and an apparatus for a fingerprint recognition device |
US11132527B2 (en) * | 2018-12-14 | 2021-09-28 | Boe Technology Group Co., Ltd. | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus |
US11288485B2 (en) * | 2018-12-14 | 2022-03-29 | Boe Technology Group Co., Ltd. | Driving method and an apparatus for a fingerprint recognition device |
US11093080B2 (en) | 2018-12-27 | 2021-08-17 | Novatek Microelectronics Corp. | Electronic device and fingerprint sensing control method thereof |
US20210326560A1 (en) * | 2018-12-28 | 2021-10-21 | Vivo Mobile Communication Co., Ltd. | Mobile terminal and recognition method |
CN111695390A (en) * | 2019-03-15 | 2020-09-22 | 京东方科技集团股份有限公司 | Method for driving grain recognition device and grain recognition device |
US11188772B2 (en) * | 2019-03-15 | 2021-11-30 | Boe Technology Group Co., Ltd. | Drive method for texture recognition device and texture recognition device |
US11301707B2 (en) * | 2019-04-30 | 2022-04-12 | Boe Technology Group Co., Ltd. | Texture recognition device and driving method of texture recognition device |
US11126813B2 (en) * | 2019-04-30 | 2021-09-21 | Egis Technology Inc. | Fingerprint sensing device |
CN112036224A (en) * | 2019-06-04 | 2020-12-04 | 联詠科技股份有限公司 | Fingerprint sensing device and method with three-dimensional sensing mechanism |
CN111213152A (en) * | 2019-06-05 | 2020-05-29 | 深圳市汇顶科技股份有限公司 | Optical image acquisition unit, optical image acquisition system, display screen and electronic equipment |
CN110309794A (en) * | 2019-07-04 | 2019-10-08 | 北京集创北方科技股份有限公司 | Physical characteristics collecting method, collection apparatus equipment and computer readable storage medium |
US11531430B2 (en) | 2019-07-12 | 2022-12-20 | Shenzhen GOODIX Technology Co., Ltd. | Fingerprint detection apparatus and electronic device |
CN111052141A (en) * | 2019-08-02 | 2020-04-21 | 深圳市汇顶科技股份有限公司 | Fingerprint detection device and electronic equipment |
US11776301B2 (en) | 2019-08-02 | 2023-10-03 | Shenzhen GOODIX Technology Co., Ltd. | Fingerprint detection apparatus and electronic device |
US11080507B2 (en) * | 2019-08-27 | 2021-08-03 | Samsung Display Co., Ltd. | Fingerprint sensing method, fingerprint sensor, and display device including the same |
US11488412B2 (en) | 2019-08-27 | 2022-11-01 | Samsung Display Co., Ltd. | Fingerprint sensing method, fingerprint sensor, and display device including the same |
WO2021051276A1 (en) | 2019-09-17 | 2021-03-25 | Boe Technology Group Co., Ltd. | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus |
EP4032008A4 (en) * | 2019-09-17 | 2022-09-07 | BOE Technology Group Co., Ltd. | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus |
US11450088B2 (en) | 2019-10-01 | 2022-09-20 | Innolux Corporation | Method of detecting biometric feature |
US20220115443A1 (en) * | 2020-10-08 | 2022-04-14 | Guangzhou Tyrafos Semiconductor Technologies Co., Ltd. | Display device and image sensing method thereof |
Also Published As
Publication number | Publication date |
---|---|
TW201546722A (en) | 2015-12-16 |
TWI550523B (en) | 2016-09-21 |
KR101900173B1 (en) | 2018-09-18 |
WO2015187197A1 (en) | 2015-12-10 |
CN106233305B (en) | 2019-10-01 |
US8917387B1 (en) | 2014-12-23 |
KR20160147002A (en) | 2016-12-21 |
EP3152700A1 (en) | 2017-04-12 |
EP3152700A4 (en) | 2018-01-10 |
EP3152700B1 (en) | 2024-04-03 |
CN106233305A (en) | 2016-12-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8917387B1 (en) | Fingerprint sensing apparatus | |
US9342733B2 (en) | Fingerprint sensing and calibration apparatus | |
KR102144205B1 (en) | Multi-level command detection device | |
US9965669B2 (en) | Optical device | |
US9891746B2 (en) | Display apparatus capable of image scanning and driving method thereof | |
US9679182B2 (en) | Display apparatus capable of image scanning and driving method thereof | |
KR101957913B1 (en) | Fingerprint Image Scanning Panel Adaptable To Under-Glass Type Structure and Display Apparatus Comprising The Same | |
TWI522995B (en) | Display device and method for driving display device | |
US10019112B2 (en) | Touch panels with dynamic zooming and low profile bezels | |
KR101515868B1 (en) | Display apparatus | |
CN110941981B (en) | Mobile fingerprint identification method and apparatus using display | |
US20210019491A1 (en) | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus | |
WO2020206825A1 (en) | Display panel and electronic device | |
KR20160056760A (en) | Flexible display apparatus able to image scan and driving method thereof | |
US11200397B2 (en) | Fingerprint identification assembly, display substrate, display panel and fingerprint identification method | |
CN111898397A (en) | Grain recognition device | |
WO2021051276A1 (en) | Photo-sensing detection apparatus, display apparatus, method of fingerprint detection, and method of operating display apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SECUGEN CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEE, DONG WON;AHN, WINNIE;KIM, JAE HO;REEL/FRAME:033563/0951 Effective date: 20140818 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551) Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 8 |