WO2011149577A2 - Systems and methods for spatially controlled scene illumination - Google Patents
Systems and methods for spatially controlled scene illumination Download PDFInfo
- Publication number
- WO2011149577A2 WO2011149577A2 PCT/US2011/026736 US2011026736W WO2011149577A2 WO 2011149577 A2 WO2011149577 A2 WO 2011149577A2 US 2011026736 W US2011026736 W US 2011026736W WO 2011149577 A2 WO2011149577 A2 WO 2011149577A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- camera
- light sources
- brightness
- person
- view
- Prior art date
Links
- 238000005286 illumination Methods 0.000 title claims abstract description 60
- 238000000034 method Methods 0.000 title claims description 68
- 238000012545 processing Methods 0.000 claims abstract description 36
- 238000005070 sampling Methods 0.000 claims description 21
- 238000012544 monitoring process Methods 0.000 claims description 14
- 230000003213 activating effect Effects 0.000 claims 2
- 206010041349 Somnolence Diseases 0.000 claims 1
- 230000007423 decrease Effects 0.000 abstract description 5
- 238000012804 iterative process Methods 0.000 abstract description 2
- 230000001276 controlling effect Effects 0.000 description 8
- 230000001815 facial effect Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 210000001747 pupil Anatomy 0.000 description 6
- 230000000694 effects Effects 0.000 description 5
- 210000003128 head Anatomy 0.000 description 5
- 230000008859 change Effects 0.000 description 3
- 230000003247 decreasing effect Effects 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 229920006395 saturated elastomer Polymers 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 230000007123 defense Effects 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 230000005670 electromagnetic radiation Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000000737 periodic effect Effects 0.000 description 2
- 238000002310 reflectometry Methods 0.000 description 2
- 230000001105 regulatory effect Effects 0.000 description 2
- 210000003786 sclera Anatomy 0.000 description 2
- 241000251468 Actinopterygii Species 0.000 description 1
- WGCNASOHLSPBMP-UHFFFAOYSA-N Glycolaldehyde Chemical compound OCC=O WGCNASOHLSPBMP-UHFFFAOYSA-N 0.000 description 1
- 244000141359 Malus pumila Species 0.000 description 1
- 230000001154 acute effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 235000021016 apples Nutrition 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000001627 detrimental effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000004424 eye movement Effects 0.000 description 1
- 210000000720 eyelash Anatomy 0.000 description 1
- 210000000744 eyelid Anatomy 0.000 description 1
- 235000013305 food Nutrition 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000007373 indentation Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 230000006996 mental state Effects 0.000 description 1
- 230000003387 muscular Effects 0.000 description 1
- 230000001681 protective effect Effects 0.000 description 1
- 230000005180 public health Effects 0.000 description 1
- 238000003908 quality control method Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 210000003491 skin Anatomy 0.000 description 1
- 230000002747 voluntary effect Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/74—Circuitry for compensating brightness variation in the scene by influencing the scene brightness using illuminating means
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C11/00—Non-optical adjuncts; Attachment thereof
- G02C11/10—Electronic devices other than hearing aids
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/10—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths
- H04N23/11—Cameras or camera modules comprising electronic image sensors; Control thereof for generating image signals from different wavelengths for generating image signals from visible and infrared light wavelengths
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/56—Cameras or camera modules comprising electronic image sensors; Control thereof provided with illuminating means
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
- H04N23/611—Control of cameras or camera modules based on recognised objects where the recognised objects include parts of the human body
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/71—Circuitry for evaluating the brightness variation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
- H04N7/183—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a single remote source
Definitions
- the present invention relates to apparatus and methods for controlling illumination of objects, and, more particularly, to apparatus and methods for producing spatially controlled brightness, e.g., substantially uniform or varying brightness, within video camera images of objects, for example, an eye or face of an individual, by regulating multiple sources of illumination, e.g., for the purpose of machine vision.
- spatially controlled brightness e.g., substantially uniform or varying brightness
- This invention relates to illumination schemes for machine vision applications, particularly when surfaces and objects within the field-of-view of a camera vary in reflectivity.
- Conventional lighting schemes frequently produce shadows and/or glints (i.e., hot-spots) within scenes that contain curved surfaces or objects that rise above a single two- dimensional plane. These problems are exacerbated when lighting must be confined to a limited number of sources and/or when light sources must be located close to a scene that is being illuminated.
- Machine vision applications that involve machine vision are becoming increasingly common-place. In part, this has arisen as a result of technological advances in the electronics and software development industries, and decreases in the cost of cameras and information processing units.
- a small number of examples from the range of machine vision applications include: object identification, distance measurements, food inspection, quality control in assembly lines, reading bar codes, object counting, safety monitoring, and biometrics.
- Sectors and industries that utilize machine vision include military, medical, security, semiconductor fabrication, manufacturing, robotics, and toys.
- saturated pixels provide no information about changes in brightness levels for image processing algorithms to differentiate edges or boundaries. In some types of video cameras, saturated pixels can also "bleed over" to elevate the apparent brightness of nearby pixels.
- precise measurements of object size involve the detection of the object edges where edges arc identified as regions where there arc sharp gradients in color or luminance. If the camera's view of the edge of an object is distorted by shadows, then the reliability and accuracy of edge-detection algorithms are degraded.
- Machine vision applications that include object identification are particularly sensitive to lighting conditions. Dark corners, color changes due to illumination, luminance changes that result from different angles of surface illumination, shadows, and hot-spots can render an object unrecognizable due to lighting conditions.
- the controlled illumination of objects is particularly difficult when illumination sources are confined to be close to the objects being illuminated. This confinement can be due, for example, to a desire to make an apparatus compact and/or to reduce power consumption by confining illumination only to the field-of-view of a camera. Such is the case when illuminating the eye using an apparatus mounted to eyewear or a head-mounted device. Examples of these types of systems or apparatus may be found in U.S. Patent 7,515,054 B2 to William C. Torch, which discloses biosensor, communication, and controller applications facilitated by monitoring eye movements.
- the present invention is directed to apparatus and methods for controlling illumination of objects. More particularly, the present invention is directed to apparatus and methods for producing spatially controlled brightness, e.g., substantially uniform brightness, within video camera images of objects, for example, an eye or face of an individual, by regulating multiple sources of illumination, e.g., for the purpose of machine vision.
- spatially controlled brightness e.g., substantially uniform brightness
- the apparatus, systems, and methods herein may provide an improved illumination method and system for machine vision applications.
- the method generally includes two or more electromagnetic radiation sources that illuminate different regions of a scene at angles that are distinct from the viewing angle of a camera.
- These illumination sources may each include one or more illuminating devices, such as incandescent bulbs, arc lamps, or light-emitting-diodes (LEDs).
- the camera may be a component of a frame grabber and processing system that produces feedback signals to individually control illumination sources based on captured images.
- Target brightness levels may be constant or may vary as a function of space (for example, to produce a gradient in brightness across a scene) and/or time (for example, when a scene is alternately illuminated by sources with different wavelengths). Controlled and/or uniform illumination may allow machine vision applications that involve location measurements and object identification to be performed more simply, quickly and reliably.
- systems and methods are provided that produce camera images with substantially spatially uniform and/or controlled scene brightness, particularly when there are spatial variations in reflectance within a scene.
- systems and methods are provided that produce images with controlled or substantially uniform brightness from surfaces that arc curved and/or not coplanar with the image plane of a camera.
- the systems and methods may use one or more illumination sources that arc positioned well away for the linc-of-sight of the camera. Illumination from acute angles may help reveal fine structures such as surface cracks or indentations.
- systems and methods may reduce the effects of shadows generated by three-dimensional structures that rise above a surface.
- the effects of shadows may be reduced by using multiple sources of electromagnetic radiation that illuminate a scene from contrasting angles.
- systems and methods may reduce or avoid the effect of so-called "glints" or bright spots that arise as a result of point sources of illumination.
- the effects of glints may be avoided by steering light into a scene using sources at angles well away from the viewpoint of the camera.
- a system for monitoring an eye of a person that includes a device configured to be worn on a person's head; a camera mounted on the device and positioned for viewing a first eye of the person wearing the device; and a plurality of light sources on the device that preferentially illuminate respective focal regions of the person's face around the first eye and within the camera's ficld-of-view.
- the camera and/or light sources may be mounted remotely from the person, e.g., to a dashboard or other interior structure of a vehicle.
- a controller may be coupled to the camera and the light sources, the controller configured for sampling brightness in the respective focal regions of the light sources using the camera and modulating the light sources based on the sampled brightness to provide desired brightness levels within the respective focal regions.
- the controller may be configured for sampling brightness from multiple pixels of the camera that correspond to a first focal region within the camera's field-of-vicw that is illuminated by a first light source, the controller combining the sampled brightness to determine an average brightness provided by the first light source, the controller modulating the first light source to provide a desired brightness level within the first focal region.
- the controller may be configured for sampling brightness from a second or additional focal regions that arc illuminated by a second or additional respective light sources, and modulating the light sources to provide desired brightness levels within the corresponding focal regions.
- the controller may be configured for amplitude modulation of at least one of the current and the voltage to the light source to provide the desired brightness levels in the respective focal regions.
- the controller may be configured for pulse-width modulation of at least one of the current and the voltage to the light sources to provide the desired brightness levels.
- a processing unit may be coupled to the camera for receiving images of the first eye, e.g., to monitor and/or analyze the images of the first eye.
- the processing unit may include the controller or may be one or more separate processors.
- a feedback controlled system for producing spatially controlled illumination that includes a camera that measures scene brightness in two or more spatial regions of a field-of-vicw; light sources that preferentially illuminate corresponding regions witliin the camera's field-of-vicw; and a processor that computes average brightness within each region of the camera's ficld-of-view and modulates corresponding light sources to one or more target brightness levels to provide desired brightness levels within the field-of-view.
- the system may include an electronic display, and the camera and/or light sources may be mounted relative to the display such that the camera and light sources arc oriented towards a face of a person viewing the electronic display.
- the camera and/or light sources may be mounted to a dashboard or other structure of a vehicle such that the camera and/or light sources are oriented towards the face of an operator of the vehicle.
- the camera and/or light sources may be mounted to a structure adjacent an assembly line or conveyor belt for obtaining images of objects being directed along the assembly line or conveyor belt.
- the camera and/or light sources may be mounted to a structure adjacent a plurality of storage areas for obtaining images of objects located in the storage areas.
- the light sources are mounted adjacent the assembly line, conveyor belt, or storage areas to preferentially illuminate respective focal regions within the camera's ficld-of-view, e.g., to facilitate identification of objects based on images acquired by the camera.
- the light sources may be operated substantially continuously or intermittently.
- the light sources may be deactivated when the camera is inoperative.
- the light sources may be activated only during the periods when the camera is activated to acquire images and/or to sample brightness, although the light sources may be activated intermittently during these periods, e.g., if pulse width modulation is used to control brightness of the light sources.
- a method for controlling illumination of a first eye of a person.
- a camera may be positioned towards a person's face such that a first eye of the person lies within a ficld-of-vicw of the camera, and at least a portion of the person's face may be illuminated with a plurality of light sources that preferentially illuminate respective focal regions of the person's face around the first eye and within the camera's ficld-of-vicw.
- Brightness may be sampled in the respective focal regions of the light sources using the camera, and the light sources modulated based at least in part on the sampled brightness to provide desired brightness levels within the respective focal regions.
- a scene illumination system may produce spatially uniform or controlled brightness levels for machine vision applications.
- target brightness levels may be substantially the same throughout the camera's ficld-of-vicw, e.g., to generate substantially uniform brightness, or may vary as a function of location, e.g., to generate controlled brightness levels within a scene.
- target brightness levels may also be made to vary with time.
- the system includes a camera, multiple light sources that preferentially illuminate different regions within the camera's field-of-view, and a processing unit coupled to the camera and light sources. Focal regions of the light sources within the camera's field-of-view may be sampled to determine average regional brightness and compared to target brightness levels.
- the processing unit may control the light sources to increase or decrease illumination levels to converge toward the target brightness levels within the field-of-view. This modulation of the light sources may be repeated with each successive video image or with periodic images until target brightness levels are achieved.
- the iterative feedback control may be locked-in for some machine vision applications. For other applications, the iterative process may continue periodically or continuously to account for different scenes or changes in lighting conditions.
- a method for controlling illumination of a scene includes directing a camera towards a scene such that one or more objects in the scene are within a field-of-view of the camera; illuminating the scene with a plurality of light sources that preferentially illuminate respective focal regions within the camera's ficld-of-view; sampling brightness in the respective focal regions of the light sources using the camera; and modulating the light sources based at least in part on the sampled brightness to provide desired brightness levels within the respective focal regions.
- the scene may include at least a portion of a person's face, and the one or more objects may include at least one eye of the person.
- the camera and/or light sources may be mounted on a device on the person's head or at a location remotely from the person's head.
- the scene may include an electronic display, and the camera and/or light sources may be positioned relative to the display such that the camera and light sources arc oriented towards a face of a person viewing the electronic display.
- the light sources may be modulated to provide desired brightness levels to respective focal regions of the person's face within the camera's ficld-of-vicw.
- the camera and/or light sources may be mounted to a dashboard of a vehicle such that the camera is oriented towards the face of an operator of the vehicle.
- the scene may include at least a portion of the operator's face, and the light sources may be oriented towards the operator's face for preferentially illuminating respective focal regions of the operator's face within the camera's ficld-qf- vicw.
- the camera and/or light sources may be directed towards an assembly line or conveyor belt for obtaining images of objects being directed along the assembly line or conveyor belt.
- the camera and/or light sources may be directed towards a plurality of storage areas for obtaining images of objects located in the storage areas.
- the light sources may be directed towards the assembly line, conveyor belt, or storage areas to preferentially illuminate respective focal regions within the camera's ficld-of-vicw, e.g., to provide a substantially uniform brightness level within the field-of-view or variable brightness levels within the field-of-view.
- the light sources may be modulated to provide greater brightness levels for focal regions that are further from the camera than other focal regions.
- FIG. I is a perspective view of spatially uniform illumination of an eye and surrounding facial region using three light-emitting diodes under the control of a processing unit that acquires images from a camera.
- FIG. 2 is an example of illumination patterns of an eye and surrounding facial region generated by three separate light sources.
- FIG. 3 is an example of a camera image of an eye and surrounding facial region in which three clusters of pixels are sampled to determine average illumination intensities in three regions within a camera's ficld-of-view.
- FIG. 4 is an example of controlled illumination of grocery or warehouse shelves where a spatial gradient is desired in the illumination pattern. Illuminatiori is separated into four (4) horizontal regions with progressively increasing brightness toward the right side of the image.
- FIG. 5 is an example of the time sequence of feedback and signals used to control illumination using both amplitude modulation and pulse-width modulation techniques.
- FIG. 6 is a flow chart showing an exemplary algorithm that may be used to control illumination levels of multiple light sources to generate spatially uniform brightness.
- FIG. 7 is a perspective view of yet another embodiment of an apparatus for monitoring a person based upon movement of the person's eye.
- FIG. 1 shows an exemplary embodiment of a system 10 that may provide feedback-controlled illumination of an eye and/or nearby facial regions of a person.
- the system 10 includes a plurality of light sources 12 (three shown), one or more cameras 14 (one shown), and a processing unit 1 coupled to the light sources 12 and/or camera(s) 14.
- the components of the system 10 may be included on a single device or two or more separate devices.
- the camera 14 and light sources 12 may be provided on a frame 18 or other device configured to be worn on a person's head.
- the frame 18 includes a bridge piece 18a, a rim 18b extending above or around each eye and defining an opening 18c, and/or a pair of car supports 18d, e.g., similar to a pair of eyeglasses.
- lenses may or not be provided in the openings 18c, e.g., prescription, shaded, polarized, and/or protective lenses, and the like, as desired, although such lenses arc not necessary for operation of the system 10.
- components of the system 10 may be provided on other devices configured to be worn on a person's head, such as a helmet, a mask, a pair of goggles, a pull-down mask, and the like (not shown), such as those devices disclosed in U.S. Patent Nos. 6, 163,281 , 6,542,081 , or 7,488,294.
- the components may be provided on separate devices, e.g., stationary or moving structures, for monitoring a person, objects, and/or other scene, as described further elsewhere herein.
- the camera 14 and/or light sources 12 may be provided remotely from a person yet allow the person to be monitored.
- the camera 14 and/or light sources 12 may be mounted to a dashboard or elsewhere within the a cockpit or other interior region of a vehicle and oriented towards a driver, pilot, or other operator of the vehicle, a passenger, or other person within the vehicle.
- the positions of the camera 14 and/or light sources 12 may be substantially fixed or adjustable within the vehicle, e.g., such that the camera 14 and/or light sources 12 may be oriented towards the face, e.g., one or both eyes of the operator or other person.
- the camera 14 and/or light sources 12 may be mounted on or adjacent a display of a computer or other electronic device, for example, for monitoring one or both eyes of a user of the electronic device, e.g., to allow the user to control or operate the electronic device based at least in part on movement of one or both eyes.
- the processing unit 16 may also be carried by the frame 18 or may be separate and/or remote from the frame 18 (and/or other components of the system 10 if provided on other structures than the frame 18), as described elsewhere herein.
- the processing unit 16 may be provided in a casing separate from the frame 18, and may include one or more cables 17 (one shown for simplicity) extending from the frame 18.
- the cable(s) 17 may include individual cables or sets of wires coupled to the light sources 12, cameras 14, and/or other components on the frame 18 and to the processing unit 16.
- individual cables or sets of wires may be embedded in the frame 18, e.g., along the rim 18b, from respective light sources 12, cameras 14, 15, and the like, until captured within the cable 15, e.g., to reduce the overall profile of the frame 18, as desired.
- a camera 14 is mounted and/or positioned on the frame 18 such that the camera 14 includes a ficld-of-view that is directed towards a first eye 20 of a person wearing the frame 18 and/or otherwise being monitored with the system 10, as shown.
- the camera 14 may be offset from a respective opening 18c in the frame 18, e.g., to place the camera 14 away from the general viewing field of a person wearing the frame 18, e.g., as described in co-pending application Serial No. 12/55 1 ,547, filed January 13, 2010.
- the camera 14 may include a CCD or CMOS or other detector including an active area, e.g., including a rectangular or other array of pixels, for capturing images with the camera 14 and generating video signals representing the images.
- the active area of the camera 14 may have any desired shape, e.g., a square or rectangular shape, a circular or elliptical shape, and the like.
- the active area of the camera 14 may be substantially flat or may be curved, e.g., to lie within a curved plane that may be oriented towards the eye 22.
- Exemplary CMOS devices that may be used include Omnivision, Model No. OV7740, or Mocron Model No. MT9V032.
- the camera 14 may include one or more filters, lenses, and the like (not shown), if desired, e.g., to focus images on the active area, fi lter undesired intensities and/or wavelengths of light, and the like.
- a second camera may be provided that includes a field-of-view directed towards a second eye (not shown) of a person being monitored by the system 10.
- a pair of cameras 14 may be mounted on the frame 18, e.g., on a lower region of the rim 18b below each eye to minimize interference with the person's vision, thereby allowing both eyes of the person to be monitored.
- multiple cameras may be provided that are directed towards an individual eye of the person (or multiple cameras that arc directed towards each eye, not shown), e.g., providing separate or overlapping ficlds-of-view.
- one or more cameras 15 may be provided on the frame 18 that may be oriented away from the person wearing the frame 18, e.g., to acquire images of the person's surroundings, as disclosed in the patents identified elsewhere herein.
- the light sources 12 may be mounted on the frame 18 at several locations, e.g., around the opening 18c adjacent the camera 14. For example, as shown, three light sources 12a, 12b, 12c are shown, e.g., first and second light sources 12a, 12b on an upper region of the rim 18b and a third light source 12c on a lower region of the rim 18b. It will be appreciated that only two or more than three light sources (not shown) may be provided, if desired, and may be controlled using the systems and methods described herein. If the system 10 includes a second camera 14, as shown in FIG. 7, an additional set of light sources 1 2 may be provided on the frame 18 for illuminating a second eye and/or facial region (not shown) of a person wearing the frame 18.
- the system 10 includes multiple cameras directed towards an individual eye (not shown), the cameras may share a plurality of light sources or, alternatively, multiple sets of light sources may be provided for illuminating the respective fields-of-views of the cameras (also not shown).
- each light source 12 may include a light emitting diode configured for emitting a relatively narrow or wide bandwidth of the light, e.g., infrared light at one or more wavelengths between about 640-700 nanometers, broadband visible light, e.g., white light, and the like.
- the light sources 12 may include lenses, filters, diffusers, or other features (not shown), e.g., for facilitating lighting respective focal regions of the person's eye and/or face.
- the light sources 12 may be spaced apart from one another, e.g., in one or more arrays located around respective openings 1 c in the frame 18 to provide desired brightness levels, e.g., substantially uniform or variable le vel brightness of the person's eye and/or face and thereby of images of the person's eye and/or face captured using the camera 14.
- desired brightness levels e.g., substantially uniform or variable le vel brightness of the person's eye and/or face and thereby of images of the person's eye and/or face captured using the camera 14.
- the processing unit 16 may include one or more controllers or processors, e.g., one or more hardware components and/or software modules for operating various components of the system 10.
- the processing unit 16 may include a separate or integral controller (not shown) for controlling the light sources 12 and/or camera 14, for receiving and/or processing signals from the camera 14, and the like.
- one or more of the components of the processing unit 16 may be carried on the frame 18, e.g., on the ear supports 18d or rim 18b. similar to the embodiments described in the references identified elsewhere herein.
- the processing unit 16 may also include memory for storing image signals from the camcra(s) 14, 15, filters for editing and/or processing the image signals, and the like.
- the processing unit 16 may include one or more power sources (not shown), e.g., for operating the components of the system 10.
- the frame 18 and/or processing unit 16 may include one or more transmitters and/or receivers (not shown) for transmitting data, receiving instructions, and the like.
- the system 10 may include components that are remote from the frame 18 and/or processing unit 16, similar to embodiments disclosed in the references identified elsewhere herein.
- the system 10 may include one or more receivers, processors, and/or displays (not shown) at a remote location from the processing unit 16 and/or frame 18, e.g., in the same room, at a nearby monitoring station, or at a more distant location.
- an eye 20 and surrounding facial region 22 arc shown of a person being imaged using the system 10, e.g., for the purpose of automated pupil-tracking, in which substantially spatially uniform illumination may be useful.
- the facial region 22 including the eye 20 may be visualized by the camera 14 where images are communicated to the processing unit 16.
- the pupil 26 is located in the center of the iris 27.
- the iris 27, in turn, is located within the sclera 28 or white region of the eye.
- the location of the pupil 26 varies as a result of both voluntary and involuntary muscular control.
- the size and shape of the pupil 26, iris 27, and areas associated with the sclera 28 change within the field-of-vicw of the camera. In part, changes in size and shape may be due to the curvature of the eye 20.
- a number of factors can influence the intensity of light detected from different regions within the field-of-view of a camera: a) the distances between light sources and the field-of-view region, b) the intensity of each light source, c) the divergence angle of each light source, d) the reflectivity (at illumination wavelength(s)) of the illuminated surface, e) the curvature of the surface, and f) the efficiency of the camera in converting light into useful signals as well as the collection efficiency and spatial uniformity of the optics associated with the camera.
- three-dimensional structures can generate shadows if illuminated from a single or small number of light sources.
- shadows can be generated by structures that include eyelids, eye lashes, and skin folds.
- the brightness levels of different regions within the ficld-of-view of a camera may be used, in a feedback mode, to control the intensity of different illumination sources.
- the use of multiple light sources illuminating a scene from different angles may help to reduce the detrimental effects of shadows.
- Glints may also be reduced or avoided by using multiple illumination sources and/or by placing them at strategic locations, e.g., well away from the viewing angle of the camera.
- LEDs infrared light emitting diodes
- LED 12a preferentially illuminates the upper-left region of the ficld-of-view of the camera 14
- LED 12b preferentially illuminates the upper-right region of the ficld-of-view of the camera 14
- LED 12c preferentially illuminates the lower region of the field-of-view of the camera 14.
- FIG. 2 shows respective focal regions of illumination by different light sources.
- three light sources illuminate a pupil 26 and the facial region 22 around an eye 20.
- the dashes lines represent "focal regions," i.e., regions where specific light sources have a preferential or increased influence on illumination of a region compared to surrounding areas within the camera's field-of-view.
- LED 12a preferentially illuminates area 121 of FIG. 2 located in the upper-left region of the camera's field-of-view.
- LED 12b in FIG. 1 illuminates area 122 of FIG. 2 located in the upper-right region of the camera's field-of-view.
- LED 12c in FIG. 1 illuminates area 123 of FIG. 2 located in the lower region of the camera's field-of-view.
- the size of the illuminated areas or focal regions may be dependent on the divergence angle of the light source and or the distance between the light source and the illuminated surface.
- the general shape of the illuminated area may be dependent on the light profile of the light source as well as the angle between the location of the light source and a vector normal to the illuminated surface.
- the average brightness of each focal region may be controlled by adjusting the intensity of the associated light source(s).
- FIG. 3 is an example of pixel clusters that may be sampled to determine the average brightness of regions in the vicinity of a pupil 26.
- the average measured intensity of an eight by eight (8x8) element cluster of pixels 101 is used to assess the brightness of the upper-left region within a camera's field-of-view.
- Another eight by eight (8x8) element cluster of pixels 102 is used to compute the brightness of the upper-right region within a camera's field-of-view.
- a third eight by thirty two (8x32) element cluster of pixels 103 is used to assess the brightness of the lower region within the camera's field- of-view.
- brightness may be assessed from any number of regions with a field-of-view.
- the assessment of brightness may be determined from any number of pixels in clusters of any size, spacing or shape within each focal region. For example, the same pixels may be sampled during each sampling within each focal region, or different pixels, e.g., randomly selected within each focal region, may be sampled, if desired.
- Brightness may be sampled using actual video signals during monitoring the eye 20 or may be sampled outside the data stream using for monitoring the eye 20. For example, periodic frames from a series of video signals may be sampled in addition to being recorded, processed, or otherwise monitored, to estimate and/or modulate brightness levels using any of the systems and methods described herein.
- the processing unit 16 may sample predetermined pixels 101 , 102, 103 (shown in FIG. 3) in focal regions 121 , 122, 123 (shown in FIG. 2) periodically to estimate the average brightness level in each of the focal regions 12 1 , 1 22, 123.
- the processing unit 16 may then modulate the light sources 12a, 12b, 12c (shown in FIG. 1 ) to provide desired brightness levels within the respective focal regions 121 , 122, 123. For example, it may be desirable to have substantially uniform brightness when illuminating and imaging the eye 20 with the camera 14.
- the processing unit 16 may sample the sets of pixels 101 , 102, 103, determine the average brightness in each of the focal regions 121 , 122, 123, and increase or decrease the intensity of the light sources 12a, 12b, 12c, e.g., to maintain the average brightness substantially uniform and/or otherwise within desired ranges.
- one or more of the cameras may be used to sample brightness within the field-of- view of the camera(s).
- a first camera may be used exclusively for brightness sampling, e.g., as described above, while a second camera may be used to obtain images of the eye for other purposes, e.g., monitoring the person's awareness, physical and/or mental state, controlling one or more devices, and the like, as described elsewhere herein and in the references identified herein.
- multiple cameras may be used to sample brightness within focal regions of respective light sources and the sampled brightness from the multiple cameras may be averaged or otherwise compared to provide feedback control to the light sources.
- FIG. 4 shows such an example of controlling illumination in a situation where a (non-uniform) spatial gradient in brightness 242 may be desired.
- a camera 241 may be used to image objects on a constitus of shelves or conveyor system 240, e.g., such as those commonly found in grocery stores, warehouses, and manufacturing facilities.
- the camera 241 and light sources 231-234 may be mounted independently relative to one another, e.g., away from the shelves or conveyor system 240 and may be substantially stationary or moving.
- some objects and lettering on objects may appear to be smaller than others within images acquired using the camera 241 . This may be due to an actual reduced size of objects or because objects are further from the camera. Frequently, the performance of image processing algorithms may be improved by increasing the brightness (without generating saturation) of objects that appear smaller within the camera's field-of- view. Spatially controlled illumination may also be used to help image processing algorithms compensate for spatial variation in the resolution of lenses and other optical components by increasing the ratio of signal-to-noise. Most lenses, particularly small lenses, have a decreased spatial resolution and/or light-gathering capability nearer the outer edges of the lens compared to die central region of the lens.
- a horizontal gradient of brightness may be generated by dividing the ficld-of-vicw into 4 vertical regions.
- the left-most region 221 illuminated by LED 231 , may require the least brightness for reliable image processing.
- the next left-most region 222, illuminated by LED 232 may require the next least level of brightness for reliable image processing.
- Region 223, illuminated by LED 233 may require more brightness due to the presence of smaller objects.
- region 224, illuminated by LED 234, may require the most illumination because of the presence of small ob jects, the objects being furthest from the camera, and/or a decreased optical resolution near the edge of the images acquired using the camera 241.
- a processing unit may be coupled to the camera 241 to sample brightness levels within the vertical regions 221 -224, and/or may be coupled to the light sources 231 -234 to modulate the intensity of the light sources 231 -234 in response to the sampled brightness levels. For example, as explained above, the processing unit may modulate the light sources 231 -234 to provide increased brightness levels in each of the regions 221 -224 to facilitate monitoring images from the camera 241.
- FIG. 5 is an example of a time sequence of signals used to monitor and control illumination in different regions of a camera's field-of-view.
- Trace 141 may represent measured average brightness of an individual pixel cluster, while dashed line 140 may represent a target light intensity for a focal region of a light source.
- Dots 142 may represent times at which new camera images are collected for sampling brightness. If the measured average light intensity falls below the target intensity, different schemes may be used to increase the intensity of the light source associated with the corresponding region.
- die same scheme(s) may be used to decrease the brightness of the light source associated with the corresponding region.
- trace 143 illustrates a scheme in which the amplitude of cither the voltage or the current driving a light source may be used to control light intensity. This is generally referred to as "amplitude modulation.”
- trace 144 illustrates a scheme is which the duration or "dwell time" of a controlling voltage or current may be modified to control light intensity. This is generally referred to as “pulse- width modulation.”
- pulse- width modulation the duration or "dwell time" of a controlling voltage or current may be modified to control light intensity.
- illumination may be turned off at times when not needed such as when the camera is not converting light into useful signals or when the overall device is not in use, for example, to conserve energy and/or to reduce overall illumination intensities, e.g., for safety reasons.
- the light sources 231 -234 in FIG. 4 may be operated intermittently, e.g., deactivated when the camera 241 is inoperative.
- the camera 241 may be operated periodically to acquire images of its field-of-view as well as sample brightness, e.g., from the acquired images.
- the light sources 231 -234 may be activated only during the periods when the camera 241 is acquiring images with the brightness of the light sources controlled as described elsewhere herein, e.g., using amplitude modulation and/or pulse-width modulation during the periods of activation.
- the light sources 231 -234 may be activated only during the periods when the camera 241 is acquiring images with the brightness of the light sources controlled as described elsewhere herein, e.g., using amplitude modulation and/or pulse-width modulation during the periods of activation.
- amplitude modulation and/or pulse-width modulation during the periods of activation.
- the camera 241 may be operated periodically to acquire images of the field-of- view and separately to sample brightness.
- the light sources 231-234 may be activated intermittently only during the periods when the camera 241 is activated to acquire images and/or to sample brightness, e.g., using amplitude modulation and/or pulse- width modulation during the periods of activation.
- FIG. 6 is a flowchart of an exemplary algorithm that may be used to generate feedback control, e.g., to produce controlled or uniform illumination, such as using the system 10 of FIGS. 1 and 7 or the system of FIG. 4.
- the average brightness of each region may be measured successively within video images. For example, at step 3 10, a new image may be collected from a camera illuminated by a plurality of light sources.
- the average brightness of a first focal region corresponding to a first light source may be sampled by sampling a plurality of pixels from the new image within the first region. For example, the actual brightness levels of the plurality of pixels within the first region may be obtained and averaged to determine an average brightness for the first region.
- the average brightness may be compared with a target intensity for the first region. If the average brightness is greater than the target intensity (branch 330a), at step 332, the output to the first light source is decreased. If the average brightness is less than the target intensity (branch 330b), at step 334, the output to the first light source is increased. Thus, at step 336, the first light source is modulated to the desired intensity.
- the number "n” is increased and, at step 340, it is confirmed that another focal region exists that has not yet been sampled. If “n” is less than or equal to the total number of focal regions and light sources, the process is repeated, i.e., steps 320-336 for the second focal region and light source, etc. If all of the focal regions have been sampled from the new image and the light sources modulated, the process is repeated with a new image (starting again at step 3 10). Thus, with each sampling, the output to the entire array of light sources included in the system may be modulated to desired levels. The process may then be repeated upon collection of the next video image or only periodically, e.g., after every other or every tenth image acquired using the camera providing the sampled brightness.
- Target brightness levels may be the same for all regions throughout a camera's ficld- of-view to generate a scene with substantially uniform brightness, e.g., in the system 10 of FIGS. 1 and 7.
- target brightness levels may also be selected to vary in different spatial regions within a camera's field-of-view, such as the system depicted in FIG. 4, to partially compensate for the effects of reduced object size within images.
- target brightness within one or more regions may also be elevated to produce spatial gradient patterns, e.g., to enhance analysis, for example, in order to reveal detailed cracks in a surface or subtle changes in color.
- Target brightness levels may also be selected to vary as a function of time. If sets of illumination sources arc used that emit at different wavelengths, separate target brightness levels may be desired as each wavelength is selected to reveal different structures within camera images.
- brightness may be dominated by the electromagnetic sources that arc a part of the applied illumination system.
- light from the feedback-controlled illumination system may be superimposed on ambient light sources such as the sun or room lighting. In the latter case, it may be necessary to re-converge to desired brightness levels whenever ambient lighting levels change.
- Another example of an application where feedback-controlled dynamic lighting may be utilized includes security checkpoints where vehicles or individuals are identified as they approach a camera's ficld-of-view.
- cither visible or infrared illumination sources may be dynamically modulated to avoid shadows and'or illuminate objects for uniform scene brightness while avoiding hot-spots within images.
- Another example of an application where feedback-controlled dynamic lighting may be utilized is in sorting processes within assembly lines or conveyor belt systems such as those used to sort produce or other foodstuffs, such as apples or fish (not shown).
- lighting may be adjusted dynamically to better measure the dimensions and/or identify objects.
- the camera (or multiple cameras, as described elsewhere herein) and plurality of light sources may be mounted to stationary supports adjacent an assembly line or conveyor belt and directed towards objects on carried on the assembly line or conveyor belt (not shown).
- multiple stationary cameras may be provided and multiple sets of light sources may be provided, e.g.. that arc stationary relative to respective cameras or that arc mounted to an assembly line or conveyor belt such that light sources move with objects on the assembly line or conveyor belt.
- Another example of an application where feedback-controlled dynamic lighting may be utilized includes alignment systems such as those used to precisely align wheels and drive shafts. Another example of an application where feedback-controlled dynamic lighting may be utilized is in the field of face recognition.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Acoustics & Sound (AREA)
- Ophthalmology & Optometry (AREA)
- Optics & Photonics (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
- Circuit Arrangement For Electric Light Sources In General (AREA)
- Image Analysis (AREA)
- Exposure Control For Cameras (AREA)
Abstract
Description
Claims
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
RU2012141574/07A RU2012141574A (en) | 2010-03-01 | 2011-03-01 | SYSTEMS AND METHODS FOR A SPACE-CONTROLLED SCENE LIGHTING |
JP2012556180A JP2013521710A (en) | 2010-03-01 | 2011-03-01 | System and method for spatially adjusted scene lighting |
AU2011258826A AU2011258826A1 (en) | 2010-03-01 | 2011-03-01 | Systems and methods for spatially controlled scene illumination |
EP11787052.7A EP2543187B1 (en) | 2010-03-01 | 2011-03-01 | Systems and methods for spatially controlled scene illumination |
KR1020127025691A KR101829850B1 (en) | 2010-03-01 | 2011-03-01 | Systems and methods for spatially controlled scene illumination |
CN201180022056.1A CN102918834B (en) | 2010-03-01 | 2011-03-01 | The system and method illuminated for the scenery of spatial control |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/715,177 US8890946B2 (en) | 2010-03-01 | 2010-03-01 | Systems and methods for spatially controlled scene illumination |
US12/715,177 | 2010-03-01 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2011149577A2 true WO2011149577A2 (en) | 2011-12-01 |
WO2011149577A3 WO2011149577A3 (en) | 2012-02-02 |
Family
ID=44505076
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2011/026736 WO2011149577A2 (en) | 2010-03-01 | 2011-03-01 | Systems and methods for spatially controlled scene illumination |
Country Status (8)
Country | Link |
---|---|
US (2) | US8890946B2 (en) |
EP (1) | EP2543187B1 (en) |
JP (1) | JP2013521710A (en) |
KR (1) | KR101829850B1 (en) |
CN (1) | CN102918834B (en) |
AU (1) | AU2011258826A1 (en) |
RU (1) | RU2012141574A (en) |
WO (1) | WO2011149577A2 (en) |
Families Citing this family (150)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8730156B2 (en) | 2010-03-05 | 2014-05-20 | Sony Computer Entertainment America Llc | Maintaining multiple views on a shared stable virtual space |
US9250703B2 (en) | 2006-03-06 | 2016-02-02 | Sony Computer Entertainment Inc. | Interface with gaze detection and voice input |
US9158116B1 (en) | 2014-04-25 | 2015-10-13 | Osterhout Group, Inc. | Temple and ear horn assembly for headworn computer |
US9366867B2 (en) | 2014-07-08 | 2016-06-14 | Osterhout Group, Inc. | Optical systems for see-through displays |
US9715112B2 (en) | 2014-01-21 | 2017-07-25 | Osterhout Group, Inc. | Suppression of stray light in head worn computing |
US9400390B2 (en) | 2014-01-24 | 2016-07-26 | Osterhout Group, Inc. | Peripheral lighting for head worn computing |
KR101383235B1 (en) * | 2010-06-17 | 2014-04-17 | 한국전자통신연구원 | Apparatus for inputting coordinate using eye tracking and method thereof |
JP5814920B2 (en) * | 2010-07-22 | 2015-11-17 | 富士通株式会社 | Vein imaging device |
US20130154913A1 (en) * | 2010-12-16 | 2013-06-20 | Siemens Corporation | Systems and methods for a gaze and gesture interface |
US10120438B2 (en) * | 2011-05-25 | 2018-11-06 | Sony Interactive Entertainment Inc. | Eye gaze to alter device behavior |
RU2621633C2 (en) | 2011-10-28 | 2017-06-06 | Мэджик Лип, Инк. | System and method for augmented and virtual reality |
TWI471808B (en) * | 2012-07-20 | 2015-02-01 | Pixart Imaging Inc | Pupil detection device |
US9854159B2 (en) * | 2012-07-20 | 2017-12-26 | Pixart Imaging Inc. | Image system with eye protection |
TW201416908A (en) * | 2012-10-23 | 2014-05-01 | Pixart Imaging Inc | Pupil tracking device |
CN103793045B (en) * | 2012-10-31 | 2016-12-28 | 原相科技股份有限公司 | Pupil tracing device |
CN104903818B (en) | 2012-12-06 | 2018-12-14 | 谷歌有限责任公司 | Eyes track Worn type Apparatus and operation method |
US8761594B1 (en) | 2013-02-28 | 2014-06-24 | Apple Inc. | Spatially dynamic illumination for camera systems |
US9788906B2 (en) | 2013-03-15 | 2017-10-17 | Synaptive Medical (Barbados) Inc. | Context aware surgical systems for intraoperatively configuring imaging devices |
US10063782B2 (en) * | 2013-06-18 | 2018-08-28 | Motorola Solutions, Inc. | Method and apparatus for displaying an image from a camera |
CN113576398A (en) | 2013-09-03 | 2021-11-02 | 托比股份公司 | Portable eye tracking device |
US10310597B2 (en) | 2013-09-03 | 2019-06-04 | Tobii Ab | Portable eye tracking device |
US10686972B2 (en) | 2013-09-03 | 2020-06-16 | Tobii Ab | Gaze assisted field of view control |
TWI532413B (en) * | 2013-09-11 | 2016-05-01 | 晶睿通訊股份有限公司 | Light compensating system and method thereof |
US9736373B2 (en) * | 2013-10-25 | 2017-08-15 | Intel Corporation | Dynamic optimization of light source power |
US20150228119A1 (en) | 2014-02-11 | 2015-08-13 | Osterhout Group, Inc. | Spatial location presentation in head worn computing |
US10684687B2 (en) | 2014-12-03 | 2020-06-16 | Mentor Acquisition One, Llc | See-through computer display systems |
US9829707B2 (en) | 2014-08-12 | 2017-11-28 | Osterhout Group, Inc. | Measuring content brightness in head worn computing |
US10254856B2 (en) | 2014-01-17 | 2019-04-09 | Osterhout Group, Inc. | External user interface for head worn computing |
US9594246B2 (en) | 2014-01-21 | 2017-03-14 | Osterhout Group, Inc. | See-through computer display systems |
US9841599B2 (en) | 2014-06-05 | 2017-12-12 | Osterhout Group, Inc. | Optical configurations for head-worn see-through displays |
US10649220B2 (en) | 2014-06-09 | 2020-05-12 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
US11487110B2 (en) | 2014-01-21 | 2022-11-01 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US12105281B2 (en) | 2014-01-21 | 2024-10-01 | Mentor Acquisition One, Llc | See-through computer display systems |
US9753288B2 (en) | 2014-01-21 | 2017-09-05 | Osterhout Group, Inc. | See-through computer display systems |
US11669163B2 (en) | 2014-01-21 | 2023-06-06 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US12093453B2 (en) | 2014-01-21 | 2024-09-17 | Mentor Acquisition One, Llc | Eye glint imaging in see-through computer display systems |
US11892644B2 (en) | 2014-01-21 | 2024-02-06 | Mentor Acquisition One, Llc | See-through computer display systems |
US11737666B2 (en) | 2014-01-21 | 2023-08-29 | Mentor Acquisition One, Llc | Eye imaging in head worn computing |
US9846308B2 (en) | 2014-01-24 | 2017-12-19 | Osterhout Group, Inc. | Haptic systems for head-worn computers |
US20160018654A1 (en) | 2014-01-24 | 2016-01-21 | Osterhout Group, Inc. | See-through computer display systems |
US20170090557A1 (en) * | 2014-01-29 | 2017-03-30 | Google Inc. | Systems and Devices for Implementing a Side-Mounted Optical Sensor |
US9396571B2 (en) * | 2014-02-10 | 2016-07-19 | International Business Machines Corporation | Simplified lighting compositing |
US12112089B2 (en) | 2014-02-11 | 2024-10-08 | Mentor Acquisition One, Llc | Spatial location presentation in head worn computing |
US9651787B2 (en) | 2014-04-25 | 2017-05-16 | Osterhout Group, Inc. | Speaker assembly for headworn computer |
US10853589B2 (en) | 2014-04-25 | 2020-12-01 | Mentor Acquisition One, Llc | Language translation with head-worn computing |
US20160137312A1 (en) | 2014-05-06 | 2016-05-19 | Osterhout Group, Inc. | Unmanned aerial vehicle launch system |
US10564714B2 (en) | 2014-05-09 | 2020-02-18 | Google Llc | Systems and methods for biomechanically-based eye signals for interacting with real and virtual objects |
AU2015255652B2 (en) | 2014-05-09 | 2018-03-29 | Google Llc | Systems and methods for using eye signals with secure mobile communications |
US10663740B2 (en) | 2014-06-09 | 2020-05-26 | Mentor Acquisition One, Llc | Content presentation in head worn computing |
JP2016032257A (en) * | 2014-07-30 | 2016-03-07 | 株式会社デンソー | Driver monitoring device |
EP3198992B8 (en) * | 2014-09-25 | 2019-04-10 | Signify Holding B.V. | Control of lighting |
EP3226554B1 (en) * | 2014-11-27 | 2019-07-17 | Kyocera Corporation | Imaging device and vehicle |
US9674415B2 (en) | 2014-12-22 | 2017-06-06 | Google Inc. | Time-of-flight camera system with scanning illuminator |
US9635231B2 (en) | 2014-12-22 | 2017-04-25 | Google Inc. | Time-of-flight camera system and method to improve measurement quality of weak field-of-view signal regions |
US10878775B2 (en) | 2015-02-17 | 2020-12-29 | Mentor Acquisition One, Llc | See-through computer display systems |
US10909164B2 (en) | 2015-04-02 | 2021-02-02 | Essilor International | Method for updating an index of a person |
WO2016162554A1 (en) * | 2015-04-10 | 2016-10-13 | Essilor International (Compagnie Generale D'optique) | Head mounted display device |
NZ737402A (en) | 2015-05-19 | 2023-06-30 | Magic Leap Inc | Dual composite light field device |
US10139966B2 (en) | 2015-07-22 | 2018-11-27 | Osterhout Group, Inc. | External user interface for head worn computing |
US11003246B2 (en) | 2015-07-22 | 2021-05-11 | Mentor Acquisition One, Llc | External user interface for head worn computing |
JP2018533099A (en) | 2015-09-24 | 2018-11-08 | カリフォルニア インスティチュート オブ テクノロジー | Data visualization system and method using three-dimensional display |
US10602074B2 (en) * | 2015-11-10 | 2020-03-24 | Lumileds Holding B.V. | Adaptive light source |
CN106814518A (en) * | 2015-12-01 | 2017-06-09 | 深圳富泰宏精密工业有限公司 | Auto-focusing camera system and electronic installation |
US10480937B2 (en) * | 2016-01-28 | 2019-11-19 | Manatec Electronics Pvt. Ltd. | System for simultaneous measurement of wheel alignment angles and wheel runout of multi-axle vehicles |
US10850116B2 (en) | 2016-12-30 | 2020-12-01 | Mentor Acquisition One, Llc | Head-worn therapy device |
US10591728B2 (en) | 2016-03-02 | 2020-03-17 | Mentor Acquisition One, Llc | Optical systems for head-worn computers |
US10667981B2 (en) | 2016-02-29 | 2020-06-02 | Mentor Acquisition One, Llc | Reading assistance system for visually impaired |
US9880441B1 (en) | 2016-09-08 | 2018-01-30 | Osterhout Group, Inc. | Electrochromic systems for head-worn computer systems |
US9826299B1 (en) | 2016-08-22 | 2017-11-21 | Osterhout Group, Inc. | Speaker systems for head-worn computer systems |
US10684478B2 (en) | 2016-05-09 | 2020-06-16 | Mentor Acquisition One, Llc | User interface systems for head-worn computers |
US10466491B2 (en) | 2016-06-01 | 2019-11-05 | Mentor Acquisition One, Llc | Modular systems for head-worn computers |
US9910284B1 (en) | 2016-09-08 | 2018-03-06 | Osterhout Group, Inc. | Optical systems for head-worn computers |
US10084979B2 (en) * | 2016-07-29 | 2018-09-25 | International Business Machines Corporation | Camera apparatus and system, method and recording medium for indicating camera field of view |
KR102548199B1 (en) * | 2016-08-03 | 2023-06-28 | 삼성전자주식회사 | Electronic device and method for tracking gaze in the electronic device |
CN106128406B (en) * | 2016-09-08 | 2019-01-22 | 京东方科技集团股份有限公司 | Eye map sheet value adjusting method, data transmission method, circuit and display device |
US10925479B2 (en) * | 2016-10-13 | 2021-02-23 | Ronald Michael Kurtz | Networked system of mobile communication platforms for nonpharmacologic constriction of a pupil |
US10580234B2 (en) | 2017-01-20 | 2020-03-03 | Adesa, Inc. | Vehicle documentation system |
US10302764B2 (en) * | 2017-02-03 | 2019-05-28 | Microsoft Technology Licensing, Llc | Active illumination management through contextual information |
CN108507636B (en) * | 2017-02-27 | 2020-11-20 | 深圳市朗驰欣创科技股份有限公司 | Method for detecting water accumulation condition of channel, inspection device and water level detection equipment |
KR102451702B1 (en) | 2017-03-30 | 2022-10-05 | 매직 립, 인코포레이티드 | Non-blocking dual driver earphones |
US10977858B2 (en) | 2017-03-30 | 2021-04-13 | Magic Leap, Inc. | Centralized rendering |
IL269861B2 (en) | 2017-04-14 | 2023-11-01 | Magic Leap Inc | Multimodal eye tracking |
US10295827B1 (en) * | 2017-04-27 | 2019-05-21 | Facebook Technologies, Llc | Diffractive optics beam shaping for structured light generator |
US11079522B1 (en) | 2017-05-31 | 2021-08-03 | Magic Leap, Inc. | Fiducial design |
US10422995B2 (en) | 2017-07-24 | 2019-09-24 | Mentor Acquisition One, Llc | See-through computer display systems with stray light management |
US10578869B2 (en) | 2017-07-24 | 2020-03-03 | Mentor Acquisition One, Llc | See-through computer display systems with adjustable zoom cameras |
US11409105B2 (en) | 2017-07-24 | 2022-08-09 | Mentor Acquisition One, Llc | See-through computer display systems |
JPWO2019026925A1 (en) * | 2017-08-01 | 2020-07-30 | パナソニックIpマネジメント株式会社 | Personal authentication device |
US10969584B2 (en) | 2017-08-04 | 2021-04-06 | Mentor Acquisition One, Llc | Image expansion optic for head-worn computer |
WO2019079523A1 (en) | 2017-10-17 | 2019-04-25 | Magic Leap, Inc. | Mixed reality spatial audio |
US11556741B2 (en) | 2018-02-09 | 2023-01-17 | Pupil Labs Gmbh | Devices, systems and methods for predicting gaze-related parameters using a neural network |
EP3749172B1 (en) | 2018-02-09 | 2022-03-30 | Pupil Labs GmbH | Devices, systems and methods for predicting gaze-related parameters |
WO2019154509A1 (en) | 2018-02-09 | 2019-08-15 | Pupil Labs Gmbh | Devices, systems and methods for predicting gaze-related parameters |
CA3090281A1 (en) | 2018-02-15 | 2019-08-22 | Magic Leap, Inc. | Dual listener positions for mixed reality |
JP7313361B2 (en) | 2018-02-15 | 2023-07-24 | マジック リープ, インコーポレイテッド | mixed reality instrument |
CN116781827A (en) | 2018-02-15 | 2023-09-19 | 奇跃公司 | Mixed reality virtual reverberation |
EP3781976A1 (en) * | 2018-04-18 | 2021-02-24 | Gentex Corporation | Confined field of view illumination in surgical microscopy |
JP7438127B2 (en) * | 2018-04-19 | 2024-02-26 | シーイング マシーンズ リミテッド | Infrared light source protection system |
JP2021522552A (en) | 2018-04-24 | 2021-08-30 | メンター アクイジション ワン, エルエルシー | See-through computer display system with visual correction and increased content density |
WO2019221767A1 (en) | 2018-05-14 | 2019-11-21 | Virtualitics, Inc. | Systems and methods for high dimensional 3d data visualization |
US10509467B1 (en) * | 2018-06-01 | 2019-12-17 | Facebook Technologies, Llc | Determining fixation of a user's eyes from images of portions of the user's face enclosed by a head mounted display |
US10667072B2 (en) | 2018-06-12 | 2020-05-26 | Magic Leap, Inc. | Efficient rendering of virtual soundfields |
WO2019241760A1 (en) | 2018-06-14 | 2019-12-19 | Magic Leap, Inc. | Methods and systems for audio signal filtering |
WO2019241754A1 (en) | 2018-06-14 | 2019-12-19 | Magic Leap, Inc. | Reverberation gain normalization |
CN112567767B (en) | 2018-06-18 | 2023-01-03 | 奇跃公司 | Spatial audio for interactive audio environments |
CN112513983A (en) | 2018-06-21 | 2021-03-16 | 奇跃公司 | Wearable system speech processing |
EP4366337A3 (en) | 2018-07-24 | 2024-07-10 | Magic Leap, Inc. | Application sharing |
US10506681B1 (en) | 2018-08-09 | 2019-12-10 | Distech Controls Inc. | Environment controller and method for proportionally adjusting the light intensity of several lighting devices |
EP3857291A4 (en) | 2018-09-25 | 2021-11-24 | Magic Leap, Inc. | Systems and methods for augmented reality |
JP7194271B2 (en) | 2018-10-05 | 2022-12-21 | マジック リープ, インコーポレイテッド | Near-field audio rendering |
JP7545960B2 (en) | 2018-10-05 | 2024-09-05 | マジック リープ, インコーポレイテッド | Enhancements for Audio Spatialization |
EP3864575A4 (en) | 2018-10-09 | 2021-12-01 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality |
WO2020086760A2 (en) | 2018-10-24 | 2020-04-30 | Magic Leap, Inc. | Asynchronous asic |
CN111127537A (en) * | 2018-10-29 | 2020-05-08 | 托比股份公司 | Method and apparatus for detecting shadows in a head mounted device |
US11509870B2 (en) * | 2018-11-07 | 2022-11-22 | Sony Group Corporation | Image acquisition system and image acquisition method |
CN113272814A (en) * | 2018-11-26 | 2021-08-17 | Zkw集团有限责任公司 | Vehicle vision system with adaptive backup light |
US11221814B2 (en) | 2018-12-27 | 2022-01-11 | Magic Leap, Inc. | Systems and methods for virtual and augmented reality |
US11537202B2 (en) | 2019-01-16 | 2022-12-27 | Pupil Labs Gmbh | Methods for generating calibration data for head-wearable devices and eye tracking system |
CN113748462A (en) | 2019-03-01 | 2021-12-03 | 奇跃公司 | Determining input for a speech processing engine |
WO2020244752A1 (en) | 2019-06-05 | 2020-12-10 | Pupil Labs Gmbh | Devices, systems and methods for predicting gaze-related parameters |
WO2020247863A1 (en) | 2019-06-06 | 2020-12-10 | Magic Leap, Inc. | Photoreal character configurations for spatial computing |
KR20220020805A (en) * | 2019-06-11 | 2022-02-21 | 젠텍스 코포레이션 | wearable medical devices |
US11328740B2 (en) | 2019-08-07 | 2022-05-10 | Magic Leap, Inc. | Voice onset detection |
US11704874B2 (en) | 2019-08-07 | 2023-07-18 | Magic Leap, Inc. | Spatial instructions and guides in mixed reality |
CN110490017B (en) * | 2019-08-08 | 2022-02-22 | 广东斯玛特自动化科技有限公司 | Illumination method and device for code scanner |
CN114830182A (en) | 2019-10-18 | 2022-07-29 | 奇跃公司 | Gravity estimation and bundle adjustment of visual inertial odometer |
WO2021081442A1 (en) | 2019-10-25 | 2021-04-29 | Magic Leap, Inc. | Non-uniform stereo rendering |
US11304017B2 (en) | 2019-10-25 | 2022-04-12 | Magic Leap, Inc. | Reverberation fingerprint estimation |
US11959997B2 (en) | 2019-11-22 | 2024-04-16 | Magic Leap, Inc. | System and method for tracking a wearable device |
CN115698847A (en) | 2019-12-04 | 2023-02-03 | 奇跃公司 | Variable pitch color emissive display |
EP4070284A4 (en) | 2019-12-06 | 2023-05-24 | Magic Leap, Inc. | Environment acoustics persistence |
EP4073689A4 (en) | 2019-12-09 | 2022-12-14 | Magic Leap, Inc. | Systems and methods for operating a head-mounted display system based on user identity |
US11337023B2 (en) | 2019-12-20 | 2022-05-17 | Magic Leap, Inc. | Physics-based audio and haptic synthesis |
WO2021163224A1 (en) | 2020-02-10 | 2021-08-19 | Magic Leap, Inc. | Dynamic colocation of virtual content |
CN115698818B (en) | 2020-02-14 | 2024-01-23 | 奇跃公司 | Session manager |
EP4104000A4 (en) | 2020-02-14 | 2023-07-12 | Magic Leap, Inc. | Tool bridge |
US11763559B2 (en) | 2020-02-14 | 2023-09-19 | Magic Leap, Inc. | 3D object annotation |
US11778410B2 (en) | 2020-02-14 | 2023-10-03 | Magic Leap, Inc. | Delayed audio following |
EP4104456A4 (en) | 2020-02-14 | 2023-07-19 | Magic Leap, Inc. | Multi-application audio rendering |
US11205069B1 (en) * | 2020-02-20 | 2021-12-21 | Facebook Technologies, Llc | Hybrid cornea and pupil tracking |
JP2023515886A (en) | 2020-03-02 | 2023-04-14 | マジック リープ, インコーポレイテッド | Immersive audio platform |
US11917384B2 (en) | 2020-03-27 | 2024-02-27 | Magic Leap, Inc. | Method of waking a device using spoken voice commands |
JP2023527561A (en) | 2020-05-29 | 2023-06-29 | マジック リープ, インコーポレイテッド | Proper collision of surfaces |
EP4158615A4 (en) | 2020-05-29 | 2023-11-15 | Magic Leap, Inc. | Determining angular acceleration |
KR20220169207A (en) * | 2021-06-18 | 2022-12-27 | 삼성전자주식회사 | Electronic device and biometric liveness authentication method |
WO2023060015A1 (en) * | 2021-10-05 | 2023-04-13 | Rsi Worklete, Llc | Computing devices programmed to detect slippery surfaces within enclosures and methods/systems of used thereof |
US11592899B1 (en) | 2021-10-28 | 2023-02-28 | Tectus Corporation | Button activation within an eye-controlled user interface |
US11619994B1 (en) | 2022-01-14 | 2023-04-04 | Tectus Corporation | Control of an electronic contact lens using pitch-based eye gestures |
US12118138B2 (en) | 2022-02-14 | 2024-10-15 | Tectus Corporation | Vergence authentication |
US11874961B2 (en) | 2022-05-09 | 2024-01-16 | Tectus Corporation | Managing display of an icon in an eye tracking augmented reality device |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007092512A2 (en) | 2006-02-07 | 2007-08-16 | Attention Technologies, Inc. | Driver drowsiness and distraction monitor |
US20090018419A1 (en) | 2004-04-01 | 2009-01-15 | Torch William C | Biosensors, communicators, and controllers monitoring eye movement and methods for using them |
Family Cites Families (127)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3689135A (en) | 1971-01-26 | 1972-09-05 | Laurence R Young | Method for monitoring movement of a subject{40 s eyes |
US3863243A (en) | 1972-01-19 | 1975-01-28 | Max Skolnick | Sleep inhibiting alarm |
US3798599A (en) | 1972-02-24 | 1974-03-19 | H Kafafian | Single input controller for a communication system |
US3966310A (en) | 1974-02-15 | 1976-06-29 | Larson Merlin D | Pupillometer and method of use thereof |
US4102564A (en) | 1975-04-18 | 1978-07-25 | Michael Henry L | Portable device for the accurate measurement of eye movements both in light and obscurity |
US4359724A (en) | 1980-04-28 | 1982-11-16 | Ronald R. Zimmerman | Eyelid movement detector |
EP0125808A3 (en) | 1983-04-18 | 1986-01-29 | Lee S. Weinblatt | Eye movement monitoring technique |
JPS6332624A (en) | 1986-07-28 | 1988-02-12 | Canon Inc | Information processor |
US4953111A (en) | 1987-02-12 | 1990-08-28 | Omron Tateisi Electronics Co. | Doze detector |
US4850691A (en) | 1987-03-18 | 1989-07-25 | University Of Illinois | Method and apparatus for determining pupillary response parameters |
US4815839A (en) | 1987-08-03 | 1989-03-28 | Waldorf Ronald A | Infrared/video electronystagmographic apparatus |
JPH01312902A (en) | 1988-06-13 | 1989-12-18 | Konan Camera Kenkyusho:Kk | Examination device of motion of eyeball |
US4852988A (en) | 1988-09-12 | 1989-08-01 | Applied Science Laboratories | Visor and camera providing a parallax-free field-of-view image for a head-mounted eye movement measurement system |
JPH02164335A (en) | 1988-12-16 | 1990-06-25 | Konan Camera Kenkyusho:Kk | Eyeball motion analyzing device |
GB8916206D0 (en) | 1989-07-14 | 1989-11-08 | Marconi Gec Ltd | Helmet systems |
US4967186A (en) | 1989-08-18 | 1990-10-30 | Ariold Ludmirsky | Method and apparatus for fatigue detection |
US5778893A (en) | 1991-04-01 | 1998-07-14 | President And Fellows Of Harvard College | Method of diagnosing and monitoring a treatment for Alzheimer's disease |
US5189512A (en) | 1991-07-01 | 1993-02-23 | Camair Research, Inc. | Helmet integrated display system |
US5183512A (en) * | 1991-08-29 | 1993-02-02 | Aquotech, Inc. | Method of cleaning semiconductor wafer and computer disks by purified water treated with electric AC signal |
WO1995018565A1 (en) | 1991-09-26 | 1995-07-13 | Sam Technology, Inc. | Non-invasive neurocognitive testing method and system |
US5214456A (en) | 1991-10-09 | 1993-05-25 | Computed Anatomy Incorporated | Mapping of corneal topography with display of pupil perimeter |
US5583590A (en) | 1992-05-04 | 1996-12-10 | Wabash Scientific Corp. | Alert monitoring system |
US5341181A (en) | 1992-11-20 | 1994-08-23 | Godard Roger R | Systems and methods for capturing and presentng visual information |
US5345281A (en) | 1992-12-17 | 1994-09-06 | John Taboada | Eye tracking system and method |
US5461417A (en) * | 1993-02-16 | 1995-10-24 | Northeast Robotics, Inc. | Continuous diffuse illumination method and apparatus |
US5402109A (en) | 1993-04-29 | 1995-03-28 | Mannik; Kallis H. | Sleep prevention device for automobile drivers |
GB9323970D0 (en) | 1993-11-22 | 1994-01-12 | Toad Innovations Ltd | Safety device |
US5478239A (en) | 1993-12-21 | 1995-12-26 | Maximum Performance, Inc. | Dynamic visual acuity training method and apparatus |
US5481622A (en) | 1994-03-01 | 1996-01-02 | Rensselaer Polytechnic Institute | Eye tracking apparatus and method employing grayscale threshold values |
JP3293308B2 (en) | 1994-03-10 | 2002-06-17 | 三菱電機株式会社 | Person state detection device |
JPH086708A (en) | 1994-04-22 | 1996-01-12 | Canon Inc | Display device |
US5617872A (en) | 1994-07-25 | 1997-04-08 | Beth Israel Hospitcal Assoc. Inc. | Hypersensitive constriction velocity method for diagnosing Alzheimer's disease in a living human |
US5469143A (en) | 1995-01-10 | 1995-11-21 | Cooper; David E. | Sleep awakening device for drivers of motor vehicles |
US5583795A (en) | 1995-03-17 | 1996-12-10 | The United States Of America As Represented By The Secretary Of The Army | Apparatus for measuring eye gaze and fixation duration, and method therefor |
US5566067A (en) | 1995-03-23 | 1996-10-15 | The President And Fellows Of Harvard College | Eyelid vigilance detector system |
US5689241A (en) | 1995-04-24 | 1997-11-18 | Clarke, Sr.; James Russell | Sleep detection and driver alert apparatus |
US5570698A (en) | 1995-06-02 | 1996-11-05 | Siemens Corporate Research, Inc. | System for monitoring eyes for detecting sleep behavior |
US7738678B2 (en) * | 1995-06-07 | 2010-06-15 | Automotive Technologies International, Inc. | Light modulation techniques for imaging objects in or around a vehicle |
US5682144A (en) | 1995-11-20 | 1997-10-28 | Mannik; Kallis Hans | Eye actuated sleep prevention devices and other eye controlled devices |
US6003991A (en) | 1996-02-17 | 1999-12-21 | Erik Scott Viirre | Eye examination apparatus and method for remote examination of a patient by a health professional |
US6373961B1 (en) | 1996-03-26 | 2002-04-16 | Eye Control Technologies, Inc. | Eye controllable screen pointer |
US5726916A (en) | 1996-06-27 | 1998-03-10 | The United States Of America As Represented By The Secretary Of The Army | Method and apparatus for determining ocular gaze point of regard and fixation duration |
US5861936A (en) | 1996-07-26 | 1999-01-19 | Gillan Holdings Limited | Regulating focus in accordance with relationship of features of a person's eyes |
US6163281A (en) | 1996-08-19 | 2000-12-19 | Torch; William C. | System and method for communication using eye movement |
US6542081B2 (en) | 1996-08-19 | 2003-04-01 | William C. Torch | System and method for monitoring eye movement |
US6246344B1 (en) | 1996-08-19 | 2001-06-12 | William C. Torch | Method and apparatus for voluntary communication |
US5748113A (en) | 1996-08-19 | 1998-05-05 | Torch; William C. | Method and apparatus for communication |
JP3651204B2 (en) | 1996-12-18 | 2005-05-25 | トヨタ自動車株式会社 | Stereoscopic image display device, stereoscopic image display method, and recording medium |
US6070098A (en) | 1997-01-11 | 2000-05-30 | Circadian Technologies, Inc. | Method of and apparatus for evaluation and mitigation of microsleep events |
WO1998049028A1 (en) | 1997-04-25 | 1998-11-05 | Applied Science Group, Inc. | An alertness monitor |
US5867587A (en) | 1997-05-19 | 1999-02-02 | Northrop Grumman Corporation | Impaired operator detection and warning system employing eyeblink analysis |
AU8259498A (en) | 1997-06-19 | 1999-01-04 | Bioprobes, Inc. | System and method for screening for dementia |
US6007202A (en) | 1997-10-23 | 1999-12-28 | Lasersight Technologies, Inc. | Eye illumination system and method |
EP1027627B1 (en) | 1997-10-30 | 2009-02-11 | MYVU Corporation | Eyeglass interface system |
US6611618B1 (en) | 1997-11-13 | 2003-08-26 | Schepens Eye Research Institute, Inc. | Wide-band image enhancement |
US6055322A (en) | 1997-12-01 | 2000-04-25 | Sensor, Inc. | Method and apparatus for illuminating and imaging eyes through eyeglasses using multiple sources of illumination |
JP3361980B2 (en) | 1997-12-12 | 2003-01-07 | 株式会社東芝 | Eye gaze detecting apparatus and method |
US6088470A (en) | 1998-01-27 | 2000-07-11 | Sensar, Inc. | Method and apparatus for removal of bright or dark spots by the fusion of multiple images |
DE19803158C1 (en) | 1998-01-28 | 1999-05-06 | Daimler Chrysler Ag | Arrangement for determining the state of vigilance, esp. for machinery operator or vehicle driver |
US6152563A (en) | 1998-02-20 | 2000-11-28 | Hutchinson; Thomas E. | Eye gaze direction tracker |
US6003992A (en) * | 1998-06-02 | 1999-12-21 | Northeast Robotics Llc | Back lighting illumination system |
US6091378A (en) | 1998-06-17 | 2000-07-18 | Eye Control Technologies, Inc. | Video processing methods and apparatus for gaze point tracking |
US6867752B1 (en) | 1998-08-31 | 2005-03-15 | Semiconductor Energy Laboratory Co., Ltd. | Portable information processing system |
US6087941A (en) | 1998-09-01 | 2000-07-11 | Ferraz; Mark | Warning device for alerting a person falling asleep |
JP2000137792A (en) | 1998-10-30 | 2000-05-16 | Toyota Motor Corp | Eye part detecting device |
US6433760B1 (en) | 1999-01-14 | 2002-08-13 | University Of Central Florida | Head mounted display with eyetracking capability |
US7120880B1 (en) | 1999-02-25 | 2006-10-10 | International Business Machines Corporation | Method and system for real-time determination of a subject's interest level to media content |
WO2000052563A1 (en) | 1999-03-01 | 2000-09-08 | Bae Systems Electronics Limited | Head tracker system |
US6090051A (en) | 1999-03-03 | 2000-07-18 | Marshall; Sandra P. | Method and apparatus for eye tracking and monitoring pupil dilation to evaluate cognitive activity |
US20020024633A1 (en) | 1999-04-09 | 2002-02-28 | Daehoon Kim | Pupil evaluation system |
US6247813B1 (en) | 1999-04-09 | 2001-06-19 | Iritech, Inc. | Iris identification system and method of identifying a person through iris recognition |
US6820979B1 (en) | 1999-04-23 | 2004-11-23 | Neuroptics, Inc. | Pupilometer with pupil irregularity detection, pupil tracking, and pupil response detection capability, glaucoma screening capability, intracranial pressure detection capability, and ocular aberration measurement capability |
US6116736A (en) | 1999-04-23 | 2000-09-12 | Neuroptics, Inc. | Pupilometer with pupil irregularity detection capability |
US6346887B1 (en) | 1999-09-14 | 2002-02-12 | The United States Of America As Represented By The Secretary Of The Navy | Eye activity monitor |
GB0023172D0 (en) | 2000-09-20 | 2000-11-01 | Minter Kemp Martin J | Wakeup vizor |
JP2002143094A (en) | 2000-11-07 | 2002-05-21 | Nac Image Technology Inc | Visual axis detector |
US6864473B2 (en) | 2000-12-07 | 2005-03-08 | The United States Of America As Represented By The United States National Aeronautics And Space Administration | Dynamic optical filtration |
US6634553B2 (en) | 2001-03-21 | 2003-10-21 | Hewlett-Packard Development Company, L.P. | Method and apparatus for even illumination of a target for image capture |
JP3812355B2 (en) * | 2001-03-22 | 2006-08-23 | 日産自動車株式会社 | Driving attention detection device and drowsy driving detection device |
US6578962B1 (en) * | 2001-04-27 | 2003-06-17 | International Business Machines Corporation | Calibration-free eye gaze tracking |
US6959102B2 (en) | 2001-05-29 | 2005-10-25 | International Business Machines Corporation | Method for increasing the signal-to-noise in IR-based eye gaze trackers |
US6775060B2 (en) | 2001-07-26 | 2004-08-10 | The Schepens Eye Research Institute, Inc. | Bioptic telescope system embedded into a spectacle lens |
GB0119859D0 (en) | 2001-08-15 | 2001-10-10 | Qinetiq Ltd | Eye tracking system |
US6927694B1 (en) | 2001-08-20 | 2005-08-09 | Research Foundation Of The University Of Central Florida | Algorithm for monitoring head/eye motion for driver alertness with one camera |
US6997556B2 (en) | 2001-10-01 | 2006-02-14 | Ernst Pfleger | Method for detecting, evaluating, and analyzing look sequences |
AUPR872301A0 (en) | 2001-11-08 | 2001-11-29 | Sleep Diagnostics Pty Ltd | Alertness monitor |
US7206435B2 (en) * | 2002-03-26 | 2007-04-17 | Honda Giken Kogyo Kabushiki Kaisha | Real-time eye detection and tracking under various light conditions |
JP2004034938A (en) * | 2002-07-08 | 2004-02-05 | Yazaki Corp | Dozing-off drive prevention device and vehicle operation information collection device |
US20040061680A1 (en) | 2002-07-10 | 2004-04-01 | John Taboada | Method and apparatus for computer control |
NL1021496C2 (en) | 2002-09-19 | 2004-03-22 | Joannes Hermanus Heiligers | Monitoring device for alertness of person, especially vehicle driver, generates warning signal if eye is detected as being closed too long |
US6637883B1 (en) | 2003-01-23 | 2003-10-28 | Vishwas V. Tengshe | Gaze tracking system and method |
US7347551B2 (en) | 2003-02-13 | 2008-03-25 | Fergason Patent Properties, Llc | Optical system for monitoring eye movement |
US7391888B2 (en) | 2003-05-30 | 2008-06-24 | Microsoft Corporation | Head pose assessment methods and systems |
EP1482768B1 (en) | 2003-05-30 | 2009-01-07 | Continental Automotive GmbH | Method and driver for driving electroluminescent lamps |
US7731360B2 (en) | 2003-11-07 | 2010-06-08 | Neuro Kinetics | Portable video oculography system |
US20050100191A1 (en) * | 2003-11-11 | 2005-05-12 | Harbach Andrew P. | Imaging system and method for monitoring an eye |
US7963652B2 (en) * | 2003-11-14 | 2011-06-21 | Queen's University At Kingston | Method and apparatus for calibration-free eye tracking |
WO2005059630A2 (en) | 2003-12-17 | 2005-06-30 | The Schepens Eye Research Institute, Inc. | Peripheral field expansion device |
US7331671B2 (en) | 2004-03-29 | 2008-02-19 | Delphi Technologies, Inc. | Eye tracking method based on correlation and detected eye movement |
US7509043B2 (en) * | 2004-05-25 | 2009-03-24 | Nikon Corporation | Illuminating device for photographing and camera |
US7775679B2 (en) * | 2004-08-18 | 2010-08-17 | Advanced Illumination, Inc. | High intensity light source for a machine vision system and method of making same |
WO2006091893A2 (en) | 2005-02-23 | 2006-08-31 | Eyetracking, Inc. | Mental alertness level determination |
US7438418B2 (en) | 2005-02-23 | 2008-10-21 | Eyetracking, Inc. | Mental alertness and mental proficiency level determination |
CN101132729A (en) | 2005-03-04 | 2008-02-27 | 睡眠诊断学公司 | Measuring alertness |
US7430365B2 (en) * | 2005-03-31 | 2008-09-30 | Avago Technologies Ecbu (Singapore) Pte Ltd. | Safe eye detection |
JP2006307611A (en) | 2005-04-28 | 2006-11-09 | Shingo Kizai Kk | Opposed reflection diffusion type high output led internal illumination system road sign |
CA2613999C (en) | 2005-08-11 | 2013-01-15 | Sleep Diagnostics Pty Ltd | Alertness sensing spectacles |
US8358337B2 (en) * | 2005-08-23 | 2013-01-22 | National University Corporation Shizuoka University | Pupil detection device and pupil detection method |
EP1977374A4 (en) | 2005-11-30 | 2012-09-05 | Seeing Machines Pty Ltd | Visual tracking of eye glasses in visual head and eye tracking systems |
US8360578B2 (en) | 2006-01-26 | 2013-01-29 | Nokia Corporation | Eye tracker device |
JP2007195775A (en) | 2006-01-27 | 2007-08-09 | Keio Gijuku | Pupil detection method, pupil position detector, and transmission type eye-direction detector |
JP2007319174A (en) * | 2006-05-30 | 2007-12-13 | Matsushita Electric Ind Co Ltd | Photographic equipment and authentication apparatus using the same |
CN101485234B (en) | 2006-06-28 | 2012-08-08 | 皇家飞利浦电子股份有限公司 | Method of controlling a lighting system based on a target light distribution |
US7646422B2 (en) | 2006-10-04 | 2010-01-12 | Branislav Kisacanin | Illumination and imaging system with glare reduction and method therefor |
JP2008091840A (en) | 2006-10-05 | 2008-04-17 | Sony Corp | Solid state imaging device and electronic device |
US8225229B2 (en) | 2006-11-09 | 2012-07-17 | Sony Mobile Communications Ab | Adjusting display brightness and/or refresh rates based on eye tracking |
SE0602545L (en) | 2006-11-29 | 2008-05-30 | Tobii Technology Ab | Eye tracking illumination |
US7783077B2 (en) | 2006-12-01 | 2010-08-24 | The Boeing Company | Eye gaze tracker system and method |
US20080128498A1 (en) * | 2006-12-05 | 2008-06-05 | Fausak Andrew T | System, method, and apparatus for high value product management and tracking |
KR100850357B1 (en) | 2006-12-06 | 2008-08-04 | 한국전자통신연구원 | System and method for tracking gaze |
US20080166052A1 (en) | 2007-01-10 | 2008-07-10 | Toshinobu Hatano | Face condition determining device and imaging device |
US8253793B2 (en) * | 2007-04-20 | 2012-08-28 | Meinan Machinery Works, Inc. | Lumber inspection method, device and program |
US7633244B2 (en) | 2007-09-19 | 2009-12-15 | Zippy Technology Corp. | Method for dimming multiple lamps |
PL2242419T3 (en) | 2008-02-14 | 2016-05-31 | Nokia Technologies Oy | Device and method for determining gaze direction |
WO2010003410A1 (en) | 2008-07-08 | 2010-01-14 | It-University Of Copenhagen | Eye gaze tracking |
EP2236074B1 (en) | 2009-04-01 | 2021-05-26 | Tobii AB | Visual display with illuminators for gaze tracking |
EP2238889B1 (en) | 2009-04-01 | 2011-10-12 | Tobii Technology AB | Adaptive camera and illuminator eyetracker |
-
2010
- 2010-03-01 US US12/715,177 patent/US8890946B2/en active Active
-
2011
- 2011-03-01 EP EP11787052.7A patent/EP2543187B1/en active Active
- 2011-03-01 RU RU2012141574/07A patent/RU2012141574A/en not_active Application Discontinuation
- 2011-03-01 WO PCT/US2011/026736 patent/WO2011149577A2/en active Application Filing
- 2011-03-01 JP JP2012556180A patent/JP2013521710A/en not_active Withdrawn
- 2011-03-01 AU AU2011258826A patent/AU2011258826A1/en not_active Abandoned
- 2011-03-01 KR KR1020127025691A patent/KR101829850B1/en active IP Right Grant
- 2011-03-01 CN CN201180022056.1A patent/CN102918834B/en active Active
-
2014
- 2014-11-18 US US14/546,821 patent/US20150181100A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090018419A1 (en) | 2004-04-01 | 2009-01-15 | Torch William C | Biosensors, communicators, and controllers monitoring eye movement and methods for using them |
US7515054B2 (en) | 2004-04-01 | 2009-04-07 | Torch William C | Biosensors, communicators, and controllers monitoring eye movement and methods for using them |
WO2007092512A2 (en) | 2006-02-07 | 2007-08-16 | Attention Technologies, Inc. | Driver drowsiness and distraction monitor |
Non-Patent Citations (1)
Title |
---|
See also references of EP2543187A4 |
Also Published As
Publication number | Publication date |
---|---|
KR101829850B1 (en) | 2018-03-29 |
EP2543187B1 (en) | 2022-11-16 |
KR20130058665A (en) | 2013-06-04 |
WO2011149577A3 (en) | 2012-02-02 |
US20110211056A1 (en) | 2011-09-01 |
US8890946B2 (en) | 2014-11-18 |
JP2013521710A (en) | 2013-06-10 |
CN102918834A (en) | 2013-02-06 |
US20150181100A1 (en) | 2015-06-25 |
AU2011258826A1 (en) | 2012-10-11 |
EP2543187A4 (en) | 2017-12-20 |
RU2012141574A (en) | 2014-04-10 |
EP2543187A2 (en) | 2013-01-09 |
CN102918834B (en) | 2016-08-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8890946B2 (en) | Systems and methods for spatially controlled scene illumination | |
KR102669768B1 (en) | Event camera system for pupil detection and eye tracking | |
US9405365B2 (en) | Systems and methods for identifying gaze tracking scene reference locations | |
US9965672B2 (en) | Method of reducing visibility of pulsed illumination while acquiring high quality imagery | |
US20200390392A1 (en) | Wearable medical device | |
EP2823751B1 (en) | Eye gaze imaging | |
US20100254571A1 (en) | Face image pickup device and method | |
JP6327753B2 (en) | Pupil detection light source device, pupil detection device, and pupil detection method | |
WO2015108923A1 (en) | Method and apparatus for strobed illumination in eye based biometric systems | |
CN111132599B (en) | Image acquisition with reduced reflections | |
WO2023195872A1 (en) | Method and system for determining heartbeat characteristics |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180022056.1 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11787052 Country of ref document: EP Kind code of ref document: A2 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012556180 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 221738 Country of ref document: IL |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011258826 Country of ref document: AU |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2666/KOLNP/2012 Country of ref document: IN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011787052 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20127025691 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012141574 Country of ref document: RU |
|
ENP | Entry into the national phase |
Ref document number: 2011258826 Country of ref document: AU Date of ref document: 20110301 Kind code of ref document: A |