US7184075B2 - Imaged coded data source tracking product - Google Patents

Imaged coded data source tracking product Download PDF

Info

Publication number
US7184075B2
US7184075B2 US10/432,801 US43280103A US7184075B2 US 7184075 B2 US7184075 B2 US 7184075B2 US 43280103 A US43280103 A US 43280103A US 7184075 B2 US7184075 B2 US 7184075B2
Authority
US
United States
Prior art keywords
coded data
data source
imager
product
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US10/432,801
Other versions
US20040027455A1 (en
Inventor
Leonard Reiffel
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
REIFFEL TECHNOLOGIES LLC
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US10/432,801 priority Critical patent/US7184075B2/en
Publication of US20040027455A1 publication Critical patent/US20040027455A1/en
Application granted granted Critical
Publication of US7184075B2 publication Critical patent/US7184075B2/en
Assigned to REIFFEL TECHNOLOGIES, LLC reassignment REIFFEL TECHNOLOGIES, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: REIFFEL, LEONARD
Adjusted expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K7/00Methods or arrangements for sensing record carriers, e.g. for reading patterns
    • G06K7/10Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation
    • G06K7/10544Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum
    • G06K7/10821Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum further details of bar or optical code scanning devices
    • G06K7/10861Methods or arrangements for sensing record carriers, e.g. for reading patterns by electromagnetic radiation, e.g. optical sensing; by corpuscular radiation by scanning of the records by radiation in the optical part of the electromagnetic spectrum further details of bar or optical code scanning devices sensing of data fields affixed to objects or articles, e.g. coded labels
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S5/00Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
    • G01S5/16Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using electromagnetic waves other than radio waves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K19/00Record carriers for use with machines and with at least a part designed to carry digital markings
    • G06K19/06Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
    • G06K19/06009Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code with optically detectable marking
    • G06K19/06018Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code with optically detectable marking one-dimensional coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06KGRAPHICAL DATA READING; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K19/00Record carriers for use with machines and with at least a part designed to carry digital markings
    • G06K19/06Record carriers for use with machines and with at least a part designed to carry digital markings characterised by the kind of the digital marking, e.g. shape, nature, code
    • G06K2019/06215Aspects not covered by other subgroups
    • G06K2019/06225Aspects not covered by other subgroups using wavelength selection, e.g. colour code

Definitions

  • the tracking product images coded data sources to cause a track signal which represents the coded data source locations and the coded data source identities.
  • Coded data source locations can represent locations of objects.
  • Coded data source identities can represent identities of the objects.
  • Object identities can be represented generically. Examples are: text, a building, a door, a vehicle, a person, a container.
  • Object identities can be represented specifically. Examples are: specific instructions, a specific building, a specific address, a specific vehicle, a specific person, a specific container.
  • Imagers can be fixed in an environment. Imagers can be hand-held and can be worn. Data can be input to an imager to cause the imager to execute various applications.
  • the product can be used for security, safety, rehabilitation, finding, tracking, guiding, and various and numerous other cases. Specific examples are: giving a warning when an object is near, giving a warning when a child is approaching a dangerous object, finding an object—which can be a person—in a cluttered environment, tracking vehicles to optimize traffic flow, guiding a person—and a robot—through an unfamiliar environment.
  • FIG. 1 schematically depicts the product.
  • FIG. 2 depicts a personal form of the product used to distinguish containers.
  • the product comprises a plurality of coded data sources and comprises an imager.
  • the imager 11 has a field of view 31 extending along each of three orthogonal axes.
  • the imager is adapted to concurrently detect light 32 from several coded data sources from the plurality of coded data sources when the several coded data sources are spatially separated in the field of view—for example the several coded data sources 41 A, 41 B, 41 C.
  • Detect light here and throughout means not only detecting the presence of light but also means detecting the specific properties of the light which encode data so that the imager can cause a signal which represents the encoded data. Detecting light concurrently from several data sources distinguishes the imager from a bar code reader which can not concurrently detect light from several spatially separated bar codes.
  • a first coded data source 41 A from the plurality of coded data sources can be in the field of view at a first coded data source location relative to at least one reference point in the field of view.
  • the first coded data source represents a first coded data source identity. While all the coded data sources can use patterns of the same physical property—such as patterns of color bands—to represent identities, the first identity is represented by the first coded data source without any physical nor spatial nor temporal dependence on the representation of a second identity by a second coded data source from the plurality of coded data sources.
  • a first track signal (for example 12 —is caused when the imager detects light from the first coded data source.
  • the first track signal represents the first coded data source location relative to at least one reference point
  • the first track signal represents the first coded data source identity.
  • Reference points can be properties—such as boundaries—of the field of view. Reference points can be properties—such as pixel location—of the imager. Reference points can be coded data sources in the field of view—such as X Y in FIG. 3 . These reference point coded data sources can be fixed and can be moving in the field of view. The field of view can be fixed and can be moving.
  • the imager can be a personal imager manipulated by a person as depicted in FIG. 2 .
  • the personal imager can be part of a device such as a cell phone and a computer.
  • the imager can be a part of an environment as depicted in FIG. 3 , and can be fixed and can be moving in the environment.
  • the imager can be the dual mode imager of patent application PCTUS/01/13742 filed 30 Apr. 2001 and published as WO 01/84475 A1 Nov. 8, 2001.
  • the coded data sources depicted in FIG. 1 are a form which works well with the dual mode imager.
  • the bands 42 A, 42 B, 42 C and 44 A, 44 B, 44 C retro reflect infra red light.
  • the locate mode of the dual mode imager detects these bands.
  • the react mode of the dual mode imager detects light from the regions 43 A, 43 B, 43 C.
  • the 42 A, 42 B, 42 C bands are wider than the 44 A, 44 B, 44 C bands which defines directed axes 45 A, 45 B, 45 C.
  • Other imagers and other forms of coded data sources can be used.
  • the imager can have components— 11 F, 11 G for example—separated in space. This can enlarge the field of view. This can facilitate determining the locations of coded data sources relative to reference points.
  • the imager can have several elements which selectively detect portions of a range of values of a physical property of light from a coded data source, with the several portions detected by the several imager elements being combined to detect the full range of the physical property of light without forming an image of the coded data source.
  • Light from a coded data source can have various sources such as light reflected from ambient sources; a coded data source light source; light emitted after energizing by suitable radiation; light emitted with a characteristic decay time after energizing by suitable radiation; a light source adjacent to the imager illuminating the coded data source; and combinations of two and more of these.
  • Light is not limited to visible light. For example, infrared can be used, and millimeter and longer wavelengths can be used. Light can be radiating energy from any portion of the electromagnetic spectrum which can provide the functions required here. Other forms of radiating energy—such as acoustic energy—which can provide the functions required here are included in the meaning of “light” here.
  • the track signal 12 can be used to generate track data. Generation of track data can be caused via a processor 71 which is signal connected 12 to the imager.
  • a memory 81 can be signal connected 72 to the processor.
  • a data entry means 91 can be signal connected 73 to the processor.
  • Track data represent the track signal and thus represent the location and identity of each of the several coded data sources in the field of view.
  • Track data can also represent information stored in the memory about each of the several coded data sources in the field of view.
  • Track data can be stored in the memory for uses such as use in another application and use at a later time.
  • a display 61 can be signal connected 13 to the imager.
  • the display can output display data representing the track data.
  • the display can output display data in registration with an ordinary image of an environment.
  • the display can have any of various forms.
  • the display can be a personal display.
  • a personal display can be any of various devices such as a cell phone and a computer.
  • the display can be a public display.
  • a public display can be a large display which can be perceived by many persons.
  • a public display can be projected to a location close to a person so that his display can be perceived by the person and only also by other persons close to the person.
  • the display can be various combinations of two and more of these.
  • the display form can be tailored to the needs and wants of various persons using the product.
  • the display can be any data output means which can be perceived by a person such as visual, audible, and tactile means.
  • the display can be combinations of these means.
  • the display can be achieved by motions of a robot.
  • the data output means can be tailored to the needs and wants of various persons indicating objects.
  • the memory can have identity data which relates coded data source codes to object identities. Identity data can be entered to the memory by any data entry means such as via keys 91 D.
  • the memory can be—in part at least—a distributed memory accessed via a networked information system.
  • the memory can have reply data associated with a coded data source identity.
  • a first query can be entered via the data entry means to solicit first reply data about a first coded data source. This can be used to cause the display to output at least part of the first reply data
  • the tracking product can be used with other identifying and locating means such as infra red systems, radio frequency systems such as radio frequency identity stags, global positioning satellite systems, and barcode systems.
  • Identity data can—for example—be entered via any of these, other identifying means, and combinations thereof.
  • Any—and all—of the processor, memory, data entry means, and display can—in part at least—be integral with the imager. Any—and all—of the processor, memory, data entry means, and display can—in part at least—be constituents of a distributed system which is signal connected via a networked information system.
  • the tracking product can find a desired medication container among several medication containers.
  • the display 61 D can indicate a desired container by highlighting 64 D an image 62 D of the container, by illuminating the container, and by various other means.
  • the display can also indicate 63 D, 65 D a container which would be selected erroneously.
  • Display data can also identify the container and can provide information about the container such as correct usage of the medication.
  • a coded data source can be on the medication itself when the medication is in pill form. In cases—like choosing a pill—which are highly controlled and where the field of view can be very small, then coded data sources can be very small.
  • coded data sources can be associated with persons so that a person not having a coded data source identifying that person as being authorized to enter an a restricted area will be detected by the product.
  • a coded data source can be associated with a child.
  • the coded data source can be part of a garment worn by the child.
  • Other coded data sources can be associated with objects and locations in the environment the child is in.
  • the display can warn if the child is approaching an unsafe object and an unsafe part of the environment.
  • a coded data source can be associated with an object which will be moved automatically by machinery—such as a shell in a magazine.
  • machinery such as a shell in a magazine.
  • shells can be stored together and the tracking product can identify each, locate each, and track each.
  • Track data can keep an object from being handled erroneously. Track data can represent the disposition of each object from manufacture to use.
  • the coded data source can have cylindrically symmetric redundancy so that the coded data source can be imaged with the object in various orientations in the field of view.
  • the containers beginnings 46 D, 46 E—and the ends 47 D, 47 E—of code repeats are depicted.
  • a coded data source can be on a vehicle so that the product can track the locations of the vehicle.
  • the product can comprise a plurality of imagers.
  • the plurality of imagers can have a joint field of view which can cover at least places where the vehicle should not be. Places where the vehicle should not be can comprise places where dangerous materials can be loaded and places where dangerous materials should not be.
  • a coded data source can be associated with text
  • This coded data source could be a generic coded data source which indicates only that text is adjacent.
  • the product can be combined with optical character recognition so that the product can provide a display representing the text.
  • This display can use a sensory pathway—such as audible and tactile pathways—other than vision.
  • This generic coded data source can be associated with any text and characters which can be deciphered via optical character recognition.
  • a generic coded data source can indicate that an environment is safe.
  • a generic coded data source can indicate that an environment is not safe.
  • the product can display the distance between any coded data source and a person.
  • the product can display directions along a path between a person and any coded data source.
  • the product can be used with an information system which can generate a path from a starting point to a target point with the path optimized by distance traveled, time traveled, and other criteria.
  • Transmitted signals can be any of various point-to-point and broadcast forms of energy transmission such as wireless and via wires, cables, and fibers. Parts of transmitted signals can reside with one form of the transmitted signal, parts can reside with a second form of transmitted signal, and parts can reside with various combinations of transmitted signals.
  • the processing can utilize configured processing elements such as fixed circuits, can utilize configurable processing elements such as field programmable gate arrays and neural networks, can utilize instructions in a data-bearing medium, and can utilize combinations of these.
  • the processing be stand alone, can act via a local information system, can act via a networked information system, and can act via combinations of these.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Electromagnetism (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Remote Sensing (AREA)
  • General Health & Medical Sciences (AREA)
  • Toxicology (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Image Processing (AREA)
  • Transforming Light Signals Into Electric Signals (AREA)

Abstract

The tracking product images several coded data sources (41A–41C) concurrently to cause a track signal (12) which represents each of the several coded data source locations and represents each of the several coded data source identities.

Description

This application claims benefit of U.S. provisional application 60/256,086 filed 15 Dec. 2000, U.S. provisional application 60/311,803 filed 13 Aug. 2001, and U.S. provisional application 60/318,961 filed 11 Sep. 2001.
The tracking product images coded data sources to cause a track signal which represents the coded data source locations and the coded data source identities.
Coded data source locations can represent locations of objects. Coded data source identities can represent identities of the objects.
Object identities can be represented generically. Examples are: text, a building, a door, a vehicle, a person, a container.
Object identities can be represented specifically. Examples are: specific instructions, a specific building, a specific address, a specific vehicle, a specific person, a specific container.
Imagers can be fixed in an environment. Imagers can be hand-held and can be worn. Data can be input to an imager to cause the imager to execute various applications.
The product can be used for security, safety, rehabilitation, finding, tracking, guiding, and various and numerous other cases. Specific examples are: giving a warning when an object is near, giving a warning when a child is approaching a dangerous object, finding an object—which can be a person—in a cluttered environment, tracking vehicles to optimize traffic flow, guiding a person—and a robot—through an unfamiliar environment.
The invention provides progress over prior art shown for example in U.S. Pat. No. 5,912,700 by Honey 15 Jun. 1999, U.S. Pat. No. 4,998,441 by Stuart 12, Mar. 1991, U.S. Pat. No. 4,099,050 by Sauermann 4, Jul. 1978, and U.S. Pat. No. 4,053,233 by Bein 11, Oct. 1977.
FIG. 1 schematically depicts the product.
FIG. 2 depicts a personal form of the product used to distinguish containers.
The product comprises a plurality of coded data sources and comprises an imager. The imager 11 has a field of view 31 extending along each of three orthogonal axes. The imager is adapted to concurrently detect light 32 from several coded data sources from the plurality of coded data sources when the several coded data sources are spatially separated in the field of view—for example the several coded data sources 41A, 41B, 41C.
“Detect light” here and throughout means not only detecting the presence of light but also means detecting the specific properties of the light which encode data so that the imager can cause a signal which represents the encoded data. Detecting light concurrently from several data sources distinguishes the imager from a bar code reader which can not concurrently detect light from several spatially separated bar codes.
A first coded data source 41A from the plurality of coded data sources can be in the field of view at a first coded data source location relative to at least one reference point in the field of view. The first coded data source represents a first coded data source identity. While all the coded data sources can use patterns of the same physical property—such as patterns of color bands—to represent identities, the first identity is represented by the first coded data source without any physical nor spatial nor temporal dependence on the representation of a second identity by a second coded data source from the plurality of coded data sources.
A first track signal—for example 12—is caused when the imager detects light from the first coded data source. The first track signal represents the first coded data source location relative to at least one reference point The first track signal represents the first coded data source identity.
Reference points can be properties—such as boundaries—of the field of view. Reference points can be properties—such as pixel location—of the imager. Reference points can be coded data sources in the field of view—such as X Y in FIG. 3. These reference point coded data sources can be fixed and can be moving in the field of view. The field of view can be fixed and can be moving.
The imager can be a personal imager manipulated by a person as depicted in FIG. 2. The personal imager can be part of a device such as a cell phone and a computer. The imager can be a part of an environment as depicted in FIG. 3, and can be fixed and can be moving in the environment.
The imager can be the dual mode imager of patent application PCTUS/01/13742 filed 30 Apr. 2001 and published as WO 01/84475 A1 Nov. 8, 2001. The coded data sources depicted in FIG. 1 are a form which works well with the dual mode imager. The bands 42A, 42B, 42C and 44A, 44B, 44C retro reflect infra red light. The locate mode of the dual mode imager detects these bands. The react mode of the dual mode imager detects light from the regions 43A, 43B, 43C. The 42A, 42B, 42C bands are wider than the 44A, 44B, 44C bands which defines directed axes 45A, 45B, 45C. Other imagers and other forms of coded data sources can be used.
The imager—dual mode and otherwise—can have components—11F, 11G for example—separated in space. This can enlarge the field of view. This can facilitate determining the locations of coded data sources relative to reference points.
The imager can have several elements which selectively detect portions of a range of values of a physical property of light from a coded data source, with the several portions detected by the several imager elements being combined to detect the full range of the physical property of light without forming an image of the coded data source.
Light from a coded data source can have various sources such as light reflected from ambient sources; a coded data source light source; light emitted after energizing by suitable radiation; light emitted with a characteristic decay time after energizing by suitable radiation; a light source adjacent to the imager illuminating the coded data source; and combinations of two and more of these.
Light is not limited to visible light. For example, infrared can be used, and millimeter and longer wavelengths can be used. Light can be radiating energy from any portion of the electromagnetic spectrum which can provide the functions required here. Other forms of radiating energy—such as acoustic energy—which can provide the functions required here are included in the meaning of “light” here.
The track signal 12 can be used to generate track data. Generation of track data can be caused via a processor 71 which is signal connected 12 to the imager. A memory 81 can be signal connected 72 to the processor. A data entry means 91 can be signal connected 73 to the processor.
Track data represent the track signal and thus represent the location and identity of each of the several coded data sources in the field of view. Track data can also represent information stored in the memory about each of the several coded data sources in the field of view. Track data can be stored in the memory for uses such as use in another application and use at a later time.
A display 61 can be signal connected 13 to the imager. The display can output display data representing the track data. The display can output display data in registration with an ordinary image of an environment.
The display can have any of various forms. The display can be a personal display. A personal display can be any of various devices such as a cell phone and a computer. The display can be a public display. A public display can be a large display which can be perceived by many persons. A public display can be projected to a location close to a person so that his display can be perceived by the person and only also by other persons close to the person.
The display can be various combinations of two and more of these. The display form can be tailored to the needs and wants of various persons using the product.
The display can be any data output means which can be perceived by a person such as visual, audible, and tactile means. The display can be combinations of these means. The display can be achieved by motions of a robot. The data output means can be tailored to the needs and wants of various persons indicating objects.
The memory can have identity data which relates coded data source codes to object identities. Identity data can be entered to the memory by any data entry means such as via keys 91D. The memory can be—in part at least—a distributed memory accessed via a networked information system.
The memory can have reply data associated with a coded data source identity. A first query can be entered via the data entry means to solicit first reply data about a first coded data source. This can be used to cause the display to output at least part of the first reply data
The tracking product can be used with other identifying and locating means such as infra red systems, radio frequency systems such as radio frequency identity stags, global positioning satellite systems, and barcode systems. Identity data can—for example—be entered via any of these, other identifying means, and combinations thereof.
Any—and all—of the processor, memory, data entry means, and display can—in part at least—be integral with the imager. Any—and all—of the processor, memory, data entry means, and display can—in part at least—be constituents of a distributed system which is signal connected via a networked information system.
Coded data sources—41D, 42D, 43D, 44D, and 41E, 42E, 43E, 44E for example—can be associated with a container—51D, 51E—in the field of view 32D, 33D of a personal imager 11D as depicted in FIG. 2. The tracking product can find a desired medication container among several medication containers.
Here the display 61D can indicate a desired container by highlighting 64D an image 62D of the container, by illuminating the container, and by various other means. The display can also indicate 63D, 65D a container which would be selected erroneously. Display data can also identify the container and can provide information about the container such as correct usage of the medication.
A coded data source can be on the medication itself when the medication is in pill form. In cases—like choosing a pill—which are highly controlled and where the field of view can be very small, then coded data sources can be very small.
In the medication case—and in various and numerous other cases where there is danger that objects can be moved erroneously and can be handled without proper authorization—there can be an imager 11E associated with the environment of the objects. Then display data can warn if an object is being moved erroneously and can warn if an object is being handled without proper authorization.
In safety and security cases coded data sources can be associated with persons so that a person not having a coded data source identifying that person as being authorized to enter an a restricted area will be detected by the product.
A coded data source can be associated with a child. The coded data source can be part of a garment worn by the child. Other coded data sources can be associated with objects and locations in the environment the child is in. The display can warn if the child is approaching an unsafe object and an unsafe part of the environment.
A coded data source can be associated with an object which will be moved automatically by machinery—such as a shell in a magazine. Several kinds of shells can be stored together and the tracking product can identify each, locate each, and track each. Track data can keep an object from being handled erroneously. Track data can represent the disposition of each object from manufacture to use.
When an object has cylindrical symmetry—such as the containers in FIG. 2 the coded data source can have cylindrically symmetric redundancy so that the coded data source can be imaged with the object in various orientations in the field of view. On the containers beginnings 46D, 46E—and the ends 47D, 47E—of code repeats are depicted.
A coded data source can be on a vehicle so that the product can track the locations of the vehicle. The product can comprise a plurality of imagers. The plurality of imagers can have a joint field of view which can cover at least places where the vehicle should not be. Places where the vehicle should not be can comprise places where dangerous materials can be loaded and places where dangerous materials should not be.
A coded data source can be associated with text This coded data source could be a generic coded data source which indicates only that text is adjacent. The product can be combined with optical character recognition so that the product can provide a display representing the text. This display can use a sensory pathway—such as audible and tactile pathways—other than vision. This generic coded data source can be associated with any text and characters which can be deciphered via optical character recognition.
A generic coded data source can indicate that an environment is safe. A generic coded data source can indicate that an environment is not safe. The product can display the distance between any coded data source and a person. The product can display directions along a path between a person and any coded data source. The product can be used with an information system which can generate a path from a starting point to a target point with the path optimized by distance traveled, time traveled, and other criteria.
A “signal” from a first product part to a second product part and a first product part being “signal connected” with a second product part here, and throughout, mean that a first physical state of the first product part causes a second physical state of the second product part. This can occur by various direct causal means and can occur by any of various transmission means. Transmitted signals can be any of various point-to-point and broadcast forms of energy transmission such as wireless and via wires, cables, and fibers. Parts of transmitted signals can reside with one form of the transmitted signal, parts can reside with a second form of transmitted signal, and parts can reside with various combinations of transmitted signals.
The several causes here can act via any of various processing modes. The processing can utilize configured processing elements such as fixed circuits, can utilize configurable processing elements such as field programmable gate arrays and neural networks, can utilize instructions in a data-bearing medium, and can utilize combinations of these. The processing be stand alone, can act via a local information system, can act via a networked information system, and can act via combinations of these.

Claims (23)

1. A tracking product comprising:
a plurality of coded data sources;
an imager,
the imager having a field of view,
the field of view extending along each of three orthogonal axes,
the field of view having a reference point,
the imager being adapted to detect light concurrently from several coded data sources from the plurality of coded data sources when the several coded data sources are spatially separated in the field of view;
a first coded data source from the plurality of coded data sources,
the first coded data source being in the field of view at a first coded data source location relative to the reference point,
the first coded data source representing a first coded data source identity,
the first identity being represented by the first coded data source without any physical nor spatial nor temporal dependence on the representation of a second identity by a second coded data source from the plurality of coded data sources;
a first track signal,
the first track signal being caused when light from the first coded data source is detected by the imager,
the first track signal representing the first coded data source identity,
the first track signal representing the first coded data source location relative to the reference point.
2. The product of claim 1 wherein the light from the first coded data source is reflected by the first coded data source.
3. The product of claim 1 wherein the imager is adapted to being held by a person.
4. The product of claim 1 wherein the imager is associated with an environment independently of persons in the environment.
5. The product of claim 4 further comprising a display representing at least part of the first track signal.
6. The product of claim 5 wherein the display is adapted to being held by a person.
7. The product of claim 6 wherein the display is associated with an environment independently of persons in the environment.
8. The product of claim 5 wherein a first query solicits first reply data about the first coded data source and wherein the display is caused to output at least part of the first reply data.
9. A tracking product comprising:
a plurality of coded data sources;
an imager,
the imager having a field of view,
the field of view extending along each of three orthogonal axes,
the field of view having a reference point,
the imager being adapted to detect light concurrently from several coded data sources from the plurality of coded data sources when the several coded data sources are spatially separated in the field of view;
a first coded data source from the plurality of coded data sources,
the first coded data source being in the field of view at a first coded data source location relative to the reference point,
the first coded data source representing a first coded data source identity,
the first identity being represented by the first coded data source without any physical nor spatial nor temporal dependence on the representation of a second identity by a second coded data source from the plurality of coded data sources;
a first track signal,
the first track signal being caused when light reflected from the first coded data source is detected by the imager,
the first track signal representing the first coded data source identity,
the first track signal representing the first coded data source location relative to the reference point.
10. The product of claim 9 wherein the imager is adapted to being held by a person.
11. The product of claim 9 wherein the imager is associated with an environment independently of persons in the environment.
12. The product of claim 11 further comprising a display representing at least part of the first track signal.
13. The product of claim 12 wherein the display is adapted to being held by a person.
14. The product of claim 12 wherein the display is associated with an environment independently of persons in the environment.
15. The product of claim 12 wherein a first query solicits first reply data about the first coded data source and wherein the display is caused to output at least part of the first reply data.
16. A tracking product comprising:
a plurality of coded data sources;
an imager,
the imager adapted to being held by a person,
the imager having a field of view,
the field of view extending along each of three orthogonal axes,
the field of view having a reference point,
the imager being adapted to detect light concurrently from several coded data sources from the plurality of coded data sources when the several coded data sources are spatially separated in the field of view;
a first coded data source from the plurality of coded data sources,
the first coded data source being in the field of view at a first coded data source location relative to the reference point,
the first coded data source representing a first coded data source identity
the first identity being represented by the first coded data source without any physical nor spatial nor temporal dependence on the representation of a second identity by a second coded data source from the plurality of coded data sources;
a first track signal,
the first track signal being caused when light from the first coded data source is detected by the imager,
the first track signal representing the first coded data source identity,
the first track signal representing the first coded data source location relative to the reference point;
a display representing at least part of the first track signal.
17. The product of claim 16 wherein the display is adapted to being held by a person.
18. The product of claim 16 wherein the display is associated with an environment independently of persons in the environment.
19. The product of claim 16 wherein a first query solicits first reply data about the first coded data source and wherein the display is caused to output at least part of the first reply data.
20. A tracking product comprising:
a plurality of coded data sources;
an imager,
the imager adapted to being associated with an environment independently of people in the environment,
the imager having a field of view,
the field of view extending along each of three orthogonal axes,
the field of view having a reference point,
the imager being adapted to detect light concurrently from several coded data sources from the plurality of coded data sources when the several coded data sources are spatially separated in the field of view;
a first coded data source from the plurality of coded data sources,
the first coded data source being in the field of view at a first coded data source location relative to the reference point,
the first coded data source representing a first coded data source identity
the first identity being represented by the first coded data source without any physical nor spatial nor temporal dependence on the representation of a second identity by a second coded data source from the plurality of coded data sources;
a first track signal,
the first track signal being caused when light from the first coded data source is detected by the imager,
the first track signal representing the first coded data source identity,
the first track signal representing the first coded data source location relative to the reference point;
a display,
the display representing at least part of the first track signal.
21. The product of claim 20 wherein the display is adapted to being held by a person.
22. The product of claim 20 wherein the display is associated with an environment independently of persons in the environment.
23. The product of claim 20 wherein a first query solicits first reply data about the first coded data source and wherein the display is caused to output at least part of the first reply data.
US10/432,801 2000-12-15 2001-12-10 Imaged coded data source tracking product Expired - Fee Related US7184075B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US10/432,801 US7184075B2 (en) 2000-12-15 2001-12-10 Imaged coded data source tracking product

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US25608600P 2000-12-15 2000-12-15
US31180301P 2001-08-13 2001-08-13
US31896101P 2001-09-11 2001-09-11
US10/432,801 US7184075B2 (en) 2000-12-15 2001-12-10 Imaged coded data source tracking product
PCT/US2001/047620 WO2002049344A1 (en) 2000-12-15 2001-12-10 Imaged coded data source tracking product

Publications (2)

Publication Number Publication Date
US20040027455A1 US20040027455A1 (en) 2004-02-12
US7184075B2 true US7184075B2 (en) 2007-02-27

Family

ID=27400920

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/432,801 Expired - Fee Related US7184075B2 (en) 2000-12-15 2001-12-10 Imaged coded data source tracking product

Country Status (5)

Country Link
US (1) US7184075B2 (en)
EP (1) EP1350385A4 (en)
JP (2) JP4294319B2 (en)
AU (1) AU2002228950A1 (en)
WO (1) WO2002049344A1 (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110119073A1 (en) * 2009-11-18 2011-05-19 Al Cure Technologies LLC Method and Apparatus for Verification of Medication Administration Adherence
US20110153361A1 (en) * 2009-12-23 2011-06-23 Al Cure Technologies LLC Method and Apparatus for Management of Clinical Trials
US20110153360A1 (en) * 2009-12-23 2011-06-23 Al Cure Technologies LLC Method and Apparatus for Verification of Clinical Trial Adherence
US20110231202A1 (en) * 2010-03-22 2011-09-22 Ai Cure Technologies Llc Method and apparatus for collection of protocol adherence data
US8605165B2 (en) 2010-10-06 2013-12-10 Ai Cure Technologies Llc Apparatus and method for assisting monitoring of medication adherence
US9116553B2 (en) 2011-02-28 2015-08-25 AI Cure Technologies, Inc. Method and apparatus for confirmation of object positioning
US9256776B2 (en) 2009-11-18 2016-02-09 AI Cure Technologies, Inc. Method and apparatus for identification
US9293060B2 (en) 2010-05-06 2016-03-22 Ai Cure Technologies Llc Apparatus and method for recognition of patient activities when obtaining protocol adherence data
US9317916B1 (en) 2013-04-12 2016-04-19 Aic Innovations Group, Inc. Apparatus and method for recognition of medication administration indicator
US9399111B1 (en) 2013-03-15 2016-07-26 Aic Innovations Group, Inc. Method and apparatus for emotional behavior therapy
US9436851B1 (en) 2013-05-07 2016-09-06 Aic Innovations Group, Inc. Geometric encrypted coded image
US9665767B2 (en) 2011-02-28 2017-05-30 Aic Innovations Group, Inc. Method and apparatus for pattern tracking
US9679113B2 (en) 2014-06-11 2017-06-13 Aic Innovations Group, Inc. Medication adherence monitoring system and method
US9824297B1 (en) 2013-10-02 2017-11-21 Aic Innovations Group, Inc. Method and apparatus for medication identification
US9875666B2 (en) 2010-05-06 2018-01-23 Aic Innovations Group, Inc. Apparatus and method for recognition of patient activities
US9883786B2 (en) 2010-05-06 2018-02-06 Aic Innovations Group, Inc. Method and apparatus for recognition of inhaler actuation
US10116903B2 (en) 2010-05-06 2018-10-30 Aic Innovations Group, Inc. Apparatus and method for recognition of suspicious activities
US10558845B2 (en) 2011-08-21 2020-02-11 Aic Innovations Group, Inc. Apparatus and method for determination of medication location
US10762172B2 (en) 2010-10-05 2020-09-01 Ai Cure Technologies Llc Apparatus and method for object confirmation and tracking
US11170484B2 (en) 2017-09-19 2021-11-09 Aic Innovations Group, Inc. Recognition of suspicious activities in medication administration

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7137711B1 (en) 2000-03-21 2006-11-21 Leonard Reiffel Multi-user retro reflector data input
KR20030040407A (en) 2000-08-18 2003-05-22 레오나드 레이필 Annotating imaged data product
US7034803B1 (en) 2000-08-18 2006-04-25 Leonard Reiffel Cursor display privacy product
EP1350385A4 (en) 2000-12-15 2004-12-15 Leonard Reiffel Imaged coded data source tracking product
JP4081373B2 (en) 2000-12-15 2008-04-23 ライフェル レナード Coded data source converter by image
EP1354471A4 (en) * 2000-12-15 2006-02-08 Leonard Reiffel Multi-imager multi-source multi-use coded data source data input product
EP1390909A4 (en) * 2001-04-19 2004-12-29 Leonard Reiffel Combined imaging coded data source data acquisition
US20040135766A1 (en) * 2001-08-15 2004-07-15 Leonard Reiffel Imaged toggled data input product
US20060291797A1 (en) * 2003-05-27 2006-12-28 Leonard Reiffel Multi-imager multi-source multi-use coded data source data input product
US7325737B2 (en) 2005-11-30 2008-02-05 Symbol Technologies, Inc. Methods and apparatus for identifying candidate barcode fields
US7755061B2 (en) 2007-11-07 2010-07-13 Kla-Tencor Technologies Corporation Dynamic pattern generator with cup-shaped structure
US8241425B2 (en) * 2009-01-23 2012-08-14 Axcelis Technologies, Inc. Non-condensing thermos chuck
US8089051B2 (en) 2010-02-24 2012-01-03 Kla-Tencor Corporation Electron reflector with multiple reflective modes
US8373144B1 (en) 2010-08-31 2013-02-12 Kla-Tencor Corporation Quasi-annular reflective electron patterning device
EP2461268A1 (en) * 2010-12-06 2012-06-06 Gemalto SA Method for selecting an electronic device without contact among N

Citations (80)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4053233A (en) 1976-03-15 1977-10-11 Aerodyne Research, Inc. Retroreflectors
US4099050A (en) 1970-07-10 1978-07-04 The United States Of America As Represented By The Secretary Of The Air Force Codable optical transponder
US4228430A (en) 1976-12-17 1980-10-14 Hitachi, Ltd. CRT Display apparatus with changeable cursor indicia
EP0062473A2 (en) 1981-03-30 1982-10-13 Minnesota Mining And Manufacturing Company Method and apparatus for sensing and identifying information within a large interrogation zone
US4439672A (en) 1981-08-07 1984-03-27 Lord Electric Company, Inc. Control system for automated manipulator device
US4603231A (en) 1983-03-31 1986-07-29 Interand Corporation System for sensing spatial coordinates
US4637797A (en) 1985-01-11 1987-01-20 Access Learning Technology Corporation Software training system
US4650334A (en) 1985-10-18 1987-03-17 Caterpillar Inc. Optical straightness gauge and method
US4684349A (en) 1984-02-15 1987-08-04 Frank Ferguson Audio-visual teaching system and method
WO1987007106A1 (en) 1986-05-12 1987-11-19 Clinicom Incorporated Portable handheld terminal including optical bar code reader and electromagnetic transceiver means for interactive wireless communication with a base communications station
US4806741A (en) * 1986-07-16 1989-02-21 Telesis Controls Corporation Electronic code enhancement for code readers
US4945914A (en) 1987-11-10 1990-08-07 Allen George S Method and apparatus for providing related images over time of a portion of the anatomy using at least four fiducial implants
US4998441A (en) 1989-07-11 1991-03-12 Aura Systems, Inc. Force and torque measurement system
US5107350A (en) 1989-06-20 1992-04-21 Kabushiki Kaisha Toshiba Image reading apparatus for optically reading the contents of a conveyed document and obtaining image signal information
US5111410A (en) 1989-06-23 1992-05-05 Kabushiki Kaisha Oh-Yoh Keisoku Kenkyusho Motion analyzing/advising system
US5181015A (en) 1989-11-07 1993-01-19 Proxima Corporation Method and apparatus for calibrating an optical computer input system
US5214414A (en) 1991-04-12 1993-05-25 International Business Machines Corp. Cursor for lcd displays
WO1993018478A1 (en) 1992-03-12 1993-09-16 Norand Corporation Reader for decoding two-dimensional optical information
US5260556A (en) 1988-12-20 1993-11-09 Australian Meat & Live-Stock Research & Development Corp. Optically readable coded target
US5282045A (en) 1990-04-27 1994-01-25 Hitachi, Ltd. Depth-of-field control apparatus and image pickup apparatus having the same therein
FR2694827A1 (en) 1992-08-14 1994-02-18 Imr System Data managing system for objects with bar=codes for identification - transmits data from optical bar=code reader via various paths to remote audio-visual displays
US5415553A (en) 1992-11-13 1995-05-16 Szmidla; Andrew Device for identifying an object using an omnidirectional bar code
US5448261A (en) 1992-06-12 1995-09-05 Sanyo Electric Co., Ltd. Cursor control device
US5453015A (en) 1988-10-20 1995-09-26 Vogel; Peter S. Audience response system and method
US5483052A (en) * 1993-12-07 1996-01-09 Smith, Iii; Herbert J. System for reading, storing and using bar-encoded data from a coded business card or other printed material
US5507527A (en) 1993-12-30 1996-04-16 Tomioka; Makoto Two dimensional code for processing data
US5537211A (en) 1995-01-13 1996-07-16 Triliance Corporation Method and apparatus for selecting a wearable to match an object
US5563401A (en) 1995-01-03 1996-10-08 Lemelson; Jerome H. Bar codes and methods
WO1996032690A1 (en) 1995-04-10 1996-10-17 United Parcel Service Of America, Inc. Method for locating the position and orientation of a fiduciary mark
US5644126A (en) 1994-05-20 1997-07-01 Kabushikikaisha Wacom Manual implement for inputting incremental information by attitude control
US5710416A (en) 1995-10-05 1998-01-20 Ncr Corporation Price verifier
US5712658A (en) 1993-12-28 1998-01-27 Hitachi, Ltd. Information presentation apparatus and information display apparatus
US5729220A (en) 1989-11-22 1998-03-17 Russell; David C. Ergonomic customizable user/computer interface device
EP0840248A2 (en) 1993-10-13 1998-05-06 Dataquill Limited Data entry systems
US5756981A (en) 1992-02-27 1998-05-26 Symbol Technologies, Inc. Optical scanner for reading and decoding one- and-two-dimensional symbologies at variable depths of field including memory efficient high speed image processing means and high accuracy image analysis means
US5789732A (en) 1995-06-08 1998-08-04 Mcmahon; Steven A. Portable data module and system for consumer transactions
US5795161A (en) 1988-10-20 1998-08-18 Vogel; Peter S. Apparatus and method for calculating an absolute time at which an event occurred
US5822735A (en) 1992-09-17 1998-10-13 Ad Response Micromarketing Corporation Focused coupon system
US5821523A (en) 1992-03-12 1998-10-13 Bunte; Alan G. Combined code reader and digital camera using a common photodetector
US5825045A (en) 1992-02-13 1998-10-20 Norand Corporation Extended range highly selective low power consuming data tag and information display system
US5826578A (en) 1994-05-26 1998-10-27 Curchod; Donald B. Motion measurement apparatus
US5835237A (en) 1994-04-22 1998-11-10 Sony Corporation Video signal coding method and apparatus thereof, and video signal decoding apparatus
US5852211A (en) 1996-04-30 1998-12-22 Roche Vitamins Inc. Process for the conversion of the sodium salt of 2-keto-L-gulonic acid to the free acid
US5852823A (en) 1996-10-16 1998-12-22 Microsoft Image classification and retrieval system using a query-by-example paradigm
US5867265A (en) 1995-08-07 1999-02-02 Ncr Corporation Apparatus and method for spectroscopic product recognition and identification
JPH11143629A (en) 1997-11-07 1999-05-28 Seiko Epson Corp Remote coordinate input device/method
US5912700A (en) 1996-01-10 1999-06-15 Fox Sports Productions, Inc. System for enhancing the television presentation of an object at a sporting event
US5917486A (en) 1996-11-04 1999-06-29 Prolexia Reading Education Software Corporation System and method for client program control of a computer display cursor
US5917472A (en) 1996-05-29 1999-06-29 International Computers Limited Cursor control system with multiple pointing devices
WO1999036836A1 (en) 1998-01-20 1999-07-22 Temtec, Inc. Identification badge verification system
US5963145A (en) 1996-02-26 1999-10-05 Universal Electronics Inc. System for providing wireless pointer control
US5982352A (en) 1992-09-18 1999-11-09 Pryor; Timothy R. Method for providing human input to a computer
US5988505A (en) 1996-06-03 1999-11-23 Symbol Technologies, Inc. Omnidirectional reading of two-dimensional symbols
US6000612A (en) 1997-10-10 1999-12-14 Metanetics Corporation Portable data collection device having optical character recognition
WO1999066441A1 (en) 1998-06-19 1999-12-23 A.T.X. International, Inc. Method and apparatus for storing reference codes in a writing instrument and for retrieving information identified by the reference codes
US6048117A (en) 1998-05-08 2000-04-11 Xerox Corporation Network-based system for color calibration of printers
US6047893A (en) 1996-12-30 2000-04-11 Datalogic S.P.A. Method of locating an object-applied optical code
US6056199A (en) 1995-09-25 2000-05-02 Intermec Ip Corporation Method and apparatus for storing and reading data
US6082619A (en) 1998-12-16 2000-07-04 Matsushita Electric Industrial Co., Ltd. Method for locating and reading a two-dimensional barcode
EP1020810A2 (en) 1998-10-30 2000-07-19 Symbol Technologies, Inc. Improved integrated code reading systems including tunnel scanners
US6118848A (en) 1998-01-14 2000-09-12 Reiffel; Leonard System to stabilize an irradiated internal target
US6121953A (en) 1997-02-06 2000-09-19 Modern Cartoons, Ltd. Virtual reality system for sensing facial movements
US6155489A (en) 1998-11-10 2000-12-05 Ncr Corporation Item checkout device including a bar code data collector and a produce data collector
US6163946A (en) 1981-05-11 2000-12-26 Great Lakes Intellectual Property Vision target based assembly
US6167607B1 (en) 1981-05-11 2001-01-02 Great Lakes Intellectual Property Vision target based assembly
US6311214B1 (en) 1995-07-27 2001-10-30 Digimarc Corporation Linking of computers based on optical sensing of digital data
US6335685B1 (en) 2000-03-09 2002-01-01 International Business Machines Corporation Apparatus and method for locating containers and contents of containers using radio frequency tags
US20020036617A1 (en) 1998-08-21 2002-03-28 Timothy R. Pryor Novel man machine interfaces and applications
US20020183961A1 (en) 1995-11-06 2002-12-05 French Barry J. System and method for tracking and assessing movement skills in multidimensional space
US6542083B1 (en) 1999-11-23 2003-04-01 Xerox Corporation Electronic tag position detection using radio broadcast
US6545670B1 (en) 1999-05-11 2003-04-08 Timothy R. Pryor Methods and apparatus for man machine interfaces and related activity
US20030222145A1 (en) 2000-05-03 2003-12-04 Leonard Reiffel Dual mode imaging product
US20040027455A1 (en) 2000-12-15 2004-02-12 Leonard Reiffel Imaged coded data source tracking product
US20040041027A1 (en) 2000-12-15 2004-03-04 Leonard Reiffel Imaged coded data source transducer product
US6720949B1 (en) 1997-08-22 2004-04-13 Timothy R. Pryor Man machine interfaces and applications
US6750848B1 (en) 1998-11-09 2004-06-15 Timothy R. Pryor More useful man machine interfaces and applications
US20040125224A1 (en) 2000-08-18 2004-07-01 Leonard Reiffel Annotating imaged data product
US20040135766A1 (en) 2001-08-15 2004-07-15 Leonard Reiffel Imaged toggled data input product
US6766036B1 (en) 1999-07-08 2004-07-20 Timothy R. Pryor Camera based man machine interfaces
US20040195327A1 (en) 2001-04-19 2004-10-07 Leonard Reiffel Combined imaging coded data source data acquisition

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2881973B2 (en) * 1990-06-13 1999-04-12 ソニー株式会社 Information retrieval device
JP2884720B2 (en) * 1990-06-15 1999-04-19 ソニー株式会社 Portable search and display
JP3620914B2 (en) * 1995-01-25 2005-02-16 シンボル テクノロジーズ インコーポレイテッド Scanning device
JPH08263585A (en) * 1995-03-24 1996-10-11 Casio Comput Co Ltd Bar code reading device
JP2000148797A (en) * 1998-08-31 2000-05-30 Sony Corp Image processor and method for image processing and image processing program storage medium

Patent Citations (87)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4099050A (en) 1970-07-10 1978-07-04 The United States Of America As Represented By The Secretary Of The Air Force Codable optical transponder
US4053233A (en) 1976-03-15 1977-10-11 Aerodyne Research, Inc. Retroreflectors
US4228430A (en) 1976-12-17 1980-10-14 Hitachi, Ltd. CRT Display apparatus with changeable cursor indicia
EP0062473A2 (en) 1981-03-30 1982-10-13 Minnesota Mining And Manufacturing Company Method and apparatus for sensing and identifying information within a large interrogation zone
US6163946A (en) 1981-05-11 2000-12-26 Great Lakes Intellectual Property Vision target based assembly
US6167607B1 (en) 1981-05-11 2001-01-02 Great Lakes Intellectual Property Vision target based assembly
US6301763B1 (en) 1981-05-11 2001-10-16 Great Lakes Intellectual Property Ltd. Determining position or orientation of object in three dimensions
US6314631B1 (en) 1981-05-11 2001-11-13 Great Lakes Intellectual Property Vision target based assembly
US6317953B1 (en) 1981-05-11 2001-11-20 Lmi-Diffracto Vision target based assembly
US4439672A (en) 1981-08-07 1984-03-27 Lord Electric Company, Inc. Control system for automated manipulator device
US4603231A (en) 1983-03-31 1986-07-29 Interand Corporation System for sensing spatial coordinates
US4684349A (en) 1984-02-15 1987-08-04 Frank Ferguson Audio-visual teaching system and method
US4637797A (en) 1985-01-11 1987-01-20 Access Learning Technology Corporation Software training system
US4650334A (en) 1985-10-18 1987-03-17 Caterpillar Inc. Optical straightness gauge and method
WO1987007106A1 (en) 1986-05-12 1987-11-19 Clinicom Incorporated Portable handheld terminal including optical bar code reader and electromagnetic transceiver means for interactive wireless communication with a base communications station
US4806741A (en) * 1986-07-16 1989-02-21 Telesis Controls Corporation Electronic code enhancement for code readers
US4945914A (en) 1987-11-10 1990-08-07 Allen George S Method and apparatus for providing related images over time of a portion of the anatomy using at least four fiducial implants
US5453015A (en) 1988-10-20 1995-09-26 Vogel; Peter S. Audience response system and method
US5795161A (en) 1988-10-20 1998-08-18 Vogel; Peter S. Apparatus and method for calculating an absolute time at which an event occurred
US5260556A (en) 1988-12-20 1993-11-09 Australian Meat & Live-Stock Research & Development Corp. Optically readable coded target
US5107350A (en) 1989-06-20 1992-04-21 Kabushiki Kaisha Toshiba Image reading apparatus for optically reading the contents of a conveyed document and obtaining image signal information
US5111410A (en) 1989-06-23 1992-05-05 Kabushiki Kaisha Oh-Yoh Keisoku Kenkyusho Motion analyzing/advising system
US4998441A (en) 1989-07-11 1991-03-12 Aura Systems, Inc. Force and torque measurement system
US6330973B1 (en) 1989-10-30 2001-12-18 Symbol Technologies, Inc. Integrated code reading systems including tunnel scanners
US5181015A (en) 1989-11-07 1993-01-19 Proxima Corporation Method and apparatus for calibrating an optical computer input system
US5729220A (en) 1989-11-22 1998-03-17 Russell; David C. Ergonomic customizable user/computer interface device
US5282045A (en) 1990-04-27 1994-01-25 Hitachi, Ltd. Depth-of-field control apparatus and image pickup apparatus having the same therein
US5214414A (en) 1991-04-12 1993-05-25 International Business Machines Corp. Cursor for lcd displays
US5825045A (en) 1992-02-13 1998-10-20 Norand Corporation Extended range highly selective low power consuming data tag and information display system
US5756981A (en) 1992-02-27 1998-05-26 Symbol Technologies, Inc. Optical scanner for reading and decoding one- and-two-dimensional symbologies at variable depths of field including memory efficient high speed image processing means and high accuracy image analysis means
WO1993018478A1 (en) 1992-03-12 1993-09-16 Norand Corporation Reader for decoding two-dimensional optical information
US5821523A (en) 1992-03-12 1998-10-13 Bunte; Alan G. Combined code reader and digital camera using a common photodetector
US5448261A (en) 1992-06-12 1995-09-05 Sanyo Electric Co., Ltd. Cursor control device
FR2694827A1 (en) 1992-08-14 1994-02-18 Imr System Data managing system for objects with bar=codes for identification - transmits data from optical bar=code reader via various paths to remote audio-visual displays
US5822735A (en) 1992-09-17 1998-10-13 Ad Response Micromarketing Corporation Focused coupon system
US5982352A (en) 1992-09-18 1999-11-09 Pryor; Timothy R. Method for providing human input to a computer
US5415553A (en) 1992-11-13 1995-05-16 Szmidla; Andrew Device for identifying an object using an omnidirectional bar code
EP0840248A2 (en) 1993-10-13 1998-05-06 Dataquill Limited Data entry systems
US5483052A (en) * 1993-12-07 1996-01-09 Smith, Iii; Herbert J. System for reading, storing and using bar-encoded data from a coded business card or other printed material
US5712658A (en) 1993-12-28 1998-01-27 Hitachi, Ltd. Information presentation apparatus and information display apparatus
US5507527A (en) 1993-12-30 1996-04-16 Tomioka; Makoto Two dimensional code for processing data
US5835237A (en) 1994-04-22 1998-11-10 Sony Corporation Video signal coding method and apparatus thereof, and video signal decoding apparatus
US5644126A (en) 1994-05-20 1997-07-01 Kabushikikaisha Wacom Manual implement for inputting incremental information by attitude control
US5826578A (en) 1994-05-26 1998-10-27 Curchod; Donald B. Motion measurement apparatus
US5563401A (en) 1995-01-03 1996-10-08 Lemelson; Jerome H. Bar codes and methods
US5537211A (en) 1995-01-13 1996-07-16 Triliance Corporation Method and apparatus for selecting a wearable to match an object
WO1996032690A1 (en) 1995-04-10 1996-10-17 United Parcel Service Of America, Inc. Method for locating the position and orientation of a fiduciary mark
US5789732A (en) 1995-06-08 1998-08-04 Mcmahon; Steven A. Portable data module and system for consumer transactions
US6311214B1 (en) 1995-07-27 2001-10-30 Digimarc Corporation Linking of computers based on optical sensing of digital data
US5867265A (en) 1995-08-07 1999-02-02 Ncr Corporation Apparatus and method for spectroscopic product recognition and identification
US6056199A (en) 1995-09-25 2000-05-02 Intermec Ip Corporation Method and apparatus for storing and reading data
US5710416A (en) 1995-10-05 1998-01-20 Ncr Corporation Price verifier
US20020183961A1 (en) 1995-11-06 2002-12-05 French Barry J. System and method for tracking and assessing movement skills in multidimensional space
US5912700A (en) 1996-01-10 1999-06-15 Fox Sports Productions, Inc. System for enhancing the television presentation of an object at a sporting event
US5963145A (en) 1996-02-26 1999-10-05 Universal Electronics Inc. System for providing wireless pointer control
US5852211A (en) 1996-04-30 1998-12-22 Roche Vitamins Inc. Process for the conversion of the sodium salt of 2-keto-L-gulonic acid to the free acid
US5917472A (en) 1996-05-29 1999-06-29 International Computers Limited Cursor control system with multiple pointing devices
US5988505A (en) 1996-06-03 1999-11-23 Symbol Technologies, Inc. Omnidirectional reading of two-dimensional symbols
US5852823A (en) 1996-10-16 1998-12-22 Microsoft Image classification and retrieval system using a query-by-example paradigm
US5917486A (en) 1996-11-04 1999-06-29 Prolexia Reading Education Software Corporation System and method for client program control of a computer display cursor
US6047893A (en) 1996-12-30 2000-04-11 Datalogic S.P.A. Method of locating an object-applied optical code
US6121953A (en) 1997-02-06 2000-09-19 Modern Cartoons, Ltd. Virtual reality system for sensing facial movements
US6720949B1 (en) 1997-08-22 2004-04-13 Timothy R. Pryor Man machine interfaces and applications
US6000612A (en) 1997-10-10 1999-12-14 Metanetics Corporation Portable data collection device having optical character recognition
US6317118B1 (en) 1997-11-07 2001-11-13 Seiko Epson Corporation Remote coordinate input device and remote coordinate input method
JPH11143629A (en) 1997-11-07 1999-05-28 Seiko Epson Corp Remote coordinate input device/method
US6118848A (en) 1998-01-14 2000-09-12 Reiffel; Leonard System to stabilize an irradiated internal target
WO1999036836A1 (en) 1998-01-20 1999-07-22 Temtec, Inc. Identification badge verification system
US6048117A (en) 1998-05-08 2000-04-11 Xerox Corporation Network-based system for color calibration of printers
WO1999066441A1 (en) 1998-06-19 1999-12-23 A.T.X. International, Inc. Method and apparatus for storing reference codes in a writing instrument and for retrieving information identified by the reference codes
US20020036617A1 (en) 1998-08-21 2002-03-28 Timothy R. Pryor Novel man machine interfaces and applications
EP1020810A2 (en) 1998-10-30 2000-07-19 Symbol Technologies, Inc. Improved integrated code reading systems including tunnel scanners
US6750848B1 (en) 1998-11-09 2004-06-15 Timothy R. Pryor More useful man machine interfaces and applications
US6155489A (en) 1998-11-10 2000-12-05 Ncr Corporation Item checkout device including a bar code data collector and a produce data collector
US6082619A (en) 1998-12-16 2000-07-04 Matsushita Electric Industrial Co., Ltd. Method for locating and reading a two-dimensional barcode
US6545670B1 (en) 1999-05-11 2003-04-08 Timothy R. Pryor Methods and apparatus for man machine interfaces and related activity
US6766036B1 (en) 1999-07-08 2004-07-20 Timothy R. Pryor Camera based man machine interfaces
US6542083B1 (en) 1999-11-23 2003-04-01 Xerox Corporation Electronic tag position detection using radio broadcast
US6335685B1 (en) 2000-03-09 2002-01-01 International Business Machines Corporation Apparatus and method for locating containers and contents of containers using radio frequency tags
US20030222145A1 (en) 2000-05-03 2003-12-04 Leonard Reiffel Dual mode imaging product
US20040188525A1 (en) 2000-05-03 2004-09-30 Leonard Reiffel Dual mode data imaging product
US6708885B2 (en) 2000-05-03 2004-03-23 Leonard Reiffel Dual mode data imaging product
US20040125224A1 (en) 2000-08-18 2004-07-01 Leonard Reiffel Annotating imaged data product
US20040041027A1 (en) 2000-12-15 2004-03-04 Leonard Reiffel Imaged coded data source transducer product
US20040027455A1 (en) 2000-12-15 2004-02-12 Leonard Reiffel Imaged coded data source tracking product
US20040195327A1 (en) 2001-04-19 2004-10-07 Leonard Reiffel Combined imaging coded data source data acquisition
US20040135766A1 (en) 2001-08-15 2004-07-15 Leonard Reiffel Imaged toggled data input product

Cited By (70)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110119073A1 (en) * 2009-11-18 2011-05-19 Al Cure Technologies LLC Method and Apparatus for Verification of Medication Administration Adherence
US9652665B2 (en) 2009-11-18 2017-05-16 Aic Innovations Group, Inc. Identification and de-identification within a video sequence
US10297032B2 (en) 2009-11-18 2019-05-21 Ai Cure Technologies Llc Verification of medication administration adherence
US10929983B2 (en) 2009-11-18 2021-02-23 Ai Cure Technologies Llc Method and apparatus for verification of medication administration adherence
US11923083B2 (en) 2009-11-18 2024-03-05 Ai Cure Technologies Llc Method and apparatus for verification of medication administration adherence
US10380744B2 (en) 2009-11-18 2019-08-13 Ai Cure Technologies Llc Verification of medication administration adherence
US10388023B2 (en) 2009-11-18 2019-08-20 Ai Cure Technologies Llc Verification of medication administration adherence
US8781856B2 (en) 2009-11-18 2014-07-15 Ai Cure Technologies Llc Method and apparatus for verification of medication administration adherence
US11646115B2 (en) 2009-11-18 2023-05-09 Ai Cure Technologies Llc Method and apparatus for verification of medication administration adherence
US10402982B2 (en) 2009-11-18 2019-09-03 Ai Cure Technologies Llc Verification of medication administration adherence
US9256776B2 (en) 2009-11-18 2016-02-09 AI Cure Technologies, Inc. Method and apparatus for identification
US10297030B2 (en) 2009-11-18 2019-05-21 Ai Cure Technologies Llc Method and apparatus for verification of medication administration adherence
US10303855B2 (en) 2009-12-23 2019-05-28 Ai Cure Technologies Llc Method and apparatus for verification of medication adherence
US10496795B2 (en) 2009-12-23 2019-12-03 Ai Cure Technologies Llc Monitoring medication adherence
US10296721B2 (en) 2009-12-23 2019-05-21 Ai Cure Technology LLC Verification of medication administration adherence
US9454645B2 (en) 2009-12-23 2016-09-27 Ai Cure Technologies Llc Apparatus and method for managing medication adherence
US10566085B2 (en) 2009-12-23 2020-02-18 Ai Cure Technologies Llc Method and apparatus for verification of medication adherence
US10496796B2 (en) 2009-12-23 2019-12-03 Ai Cure Technologies Llc Monitoring medication adherence
US11222714B2 (en) 2009-12-23 2022-01-11 Ai Cure Technologies Llc Method and apparatus for verification of medication adherence
US20110153361A1 (en) * 2009-12-23 2011-06-23 Al Cure Technologies LLC Method and Apparatus for Management of Clinical Trials
US20110153360A1 (en) * 2009-12-23 2011-06-23 Al Cure Technologies LLC Method and Apparatus for Verification of Clinical Trial Adherence
US10303856B2 (en) 2009-12-23 2019-05-28 Ai Cure Technologies Llc Verification of medication administration adherence
US8731961B2 (en) 2009-12-23 2014-05-20 Ai Cure Technologies Method and apparatus for verification of clinical trial adherence
US8666781B2 (en) 2009-12-23 2014-03-04 Ai Cure Technologies, LLC Method and apparatus for management of clinical trials
US10395009B2 (en) 2010-03-22 2019-08-27 Ai Cure Technologies Llc Apparatus and method for collection of protocol adherence data
US20110231202A1 (en) * 2010-03-22 2011-09-22 Ai Cure Technologies Llc Method and apparatus for collection of protocol adherence data
US9183601B2 (en) 2010-03-22 2015-11-10 Ai Cure Technologies Llc Method and apparatus for collection of protocol adherence data
US11244283B2 (en) 2010-03-22 2022-02-08 Ai Cure Technologies Llc Apparatus and method for collection of protocol adherence data
US10650697B2 (en) 2010-05-06 2020-05-12 Aic Innovations Group, Inc. Apparatus and method for recognition of patient activities
US11328818B2 (en) 2010-05-06 2022-05-10 Ai Cure Technologies Llc Apparatus and method for recognition of patient activities when obtaining protocol adherence data
US10262109B2 (en) 2010-05-06 2019-04-16 Ai Cure Technologies Llc Apparatus and method for recognition of patient activities when obtaining protocol adherence data
US10646101B2 (en) 2010-05-06 2020-05-12 Aic Innovations Group, Inc. Apparatus and method for recognition of inhaler actuation
US10116903B2 (en) 2010-05-06 2018-10-30 Aic Innovations Group, Inc. Apparatus and method for recognition of suspicious activities
US11862033B2 (en) 2010-05-06 2024-01-02 Aic Innovations Group, Inc. Apparatus and method for recognition of patient activities
US11094408B2 (en) 2010-05-06 2021-08-17 Aic Innovations Group, Inc. Apparatus and method for recognition of inhaler actuation
US9883786B2 (en) 2010-05-06 2018-02-06 Aic Innovations Group, Inc. Method and apparatus for recognition of inhaler actuation
US11682488B2 (en) 2010-05-06 2023-06-20 Ai Cure Technologies Llc Apparatus and method for recognition of patient activities when obtaining protocol adherence data
US9875666B2 (en) 2010-05-06 2018-01-23 Aic Innovations Group, Inc. Apparatus and method for recognition of patient activities
US10872695B2 (en) 2010-05-06 2020-12-22 Ai Cure Technologies Llc Apparatus and method for recognition of patient activities when obtaining protocol adherence data
US9293060B2 (en) 2010-05-06 2016-03-22 Ai Cure Technologies Llc Apparatus and method for recognition of patient activities when obtaining protocol adherence data
US10762172B2 (en) 2010-10-05 2020-09-01 Ai Cure Technologies Llc Apparatus and method for object confirmation and tracking
US10506971B2 (en) 2010-10-06 2019-12-17 Ai Cure Technologies Llc Apparatus and method for monitoring medication adherence
US9844337B2 (en) 2010-10-06 2017-12-19 Ai Cure Technologies Llc Method and apparatus for monitoring medication adherence
US10149648B2 (en) 2010-10-06 2018-12-11 Ai Cure Technologies Llc Method and apparatus for monitoring medication adherence
US9486720B2 (en) 2010-10-06 2016-11-08 Ai Cure Technologies Llc Method and apparatus for monitoring medication adherence
US8605165B2 (en) 2010-10-06 2013-12-10 Ai Cure Technologies Llc Apparatus and method for assisting monitoring of medication adherence
US9116553B2 (en) 2011-02-28 2015-08-25 AI Cure Technologies, Inc. Method and apparatus for confirmation of object positioning
US9665767B2 (en) 2011-02-28 2017-05-30 Aic Innovations Group, Inc. Method and apparatus for pattern tracking
US9538147B2 (en) 2011-02-28 2017-01-03 Aic Innovations Group, Inc. Method and system for determining proper positioning of an object
US10511778B2 (en) 2011-02-28 2019-12-17 Aic Innovations Group, Inc. Method and apparatus for push interaction
US9892316B2 (en) 2011-02-28 2018-02-13 Aic Innovations Group, Inc. Method and apparatus for pattern tracking
US10257423B2 (en) 2011-02-28 2019-04-09 Aic Innovations Group, Inc. Method and system for determining proper positioning of an object
US11314964B2 (en) 2011-08-21 2022-04-26 Aic Innovations Group, Inc. Apparatus and method for determination of medication location
US10558845B2 (en) 2011-08-21 2020-02-11 Aic Innovations Group, Inc. Apparatus and method for determination of medication location
US10565431B2 (en) 2012-01-04 2020-02-18 Aic Innovations Group, Inc. Method and apparatus for identification
US11004554B2 (en) 2012-01-04 2021-05-11 Aic Innovations Group, Inc. Method and apparatus for identification
US10133914B2 (en) 2012-01-04 2018-11-20 Aic Innovations Group, Inc. Identification and de-identification within a video sequence
US9399111B1 (en) 2013-03-15 2016-07-26 Aic Innovations Group, Inc. Method and apparatus for emotional behavior therapy
US9317916B1 (en) 2013-04-12 2016-04-19 Aic Innovations Group, Inc. Apparatus and method for recognition of medication administration indicator
US10460438B1 (en) 2013-04-12 2019-10-29 Aic Innovations Group, Inc. Apparatus and method for recognition of medication administration indicator
US11200965B2 (en) 2013-04-12 2021-12-14 Aic Innovations Group, Inc. Apparatus and method for recognition of medication administration indicator
US9436851B1 (en) 2013-05-07 2016-09-06 Aic Innovations Group, Inc. Geometric encrypted coded image
US9824297B1 (en) 2013-10-02 2017-11-21 Aic Innovations Group, Inc. Method and apparatus for medication identification
US10373016B2 (en) 2013-10-02 2019-08-06 Aic Innovations Group, Inc. Method and apparatus for medication identification
US10475533B2 (en) 2014-06-11 2019-11-12 Aic Innovations Group, Inc. Medication adherence monitoring system and method
US9679113B2 (en) 2014-06-11 2017-06-13 Aic Innovations Group, Inc. Medication adherence monitoring system and method
US11417422B2 (en) 2014-06-11 2022-08-16 Aic Innovations Group, Inc. Medication adherence monitoring system and method
US9977870B2 (en) 2014-06-11 2018-05-22 Aic Innovations Group, Inc. Medication adherence monitoring system and method
US10916339B2 (en) 2014-06-11 2021-02-09 Aic Innovations Group, Inc. Medication adherence monitoring system and method
US11170484B2 (en) 2017-09-19 2021-11-09 Aic Innovations Group, Inc. Recognition of suspicious activities in medication administration

Also Published As

Publication number Publication date
WO2002049344A1 (en) 2002-06-20
US20040027455A1 (en) 2004-02-12
JP4294319B2 (en) 2009-07-08
AU2002228950A1 (en) 2002-06-24
EP1350385A1 (en) 2003-10-08
EP1350385A4 (en) 2004-12-15
JP2004527147A (en) 2004-09-02
JP2008135072A (en) 2008-06-12

Similar Documents

Publication Publication Date Title
US7184075B2 (en) Imaged coded data source tracking product
US9311586B2 (en) Apparatus and method for locating, tracking, controlling and recognizing tagged objects using active RFID technology
US8403225B2 (en) Vehicle license plate indicia scanning
US20160050309A1 (en) Restrictive cell phone use and anti-texting apparatus and methods
US8717165B2 (en) Apparatus and method for locating, tracking, controlling and recognizing tagged objects using RFID technology
KR101253337B1 (en) System for prrotection missing using location-aware
JP2008181542A (en) System and method for controlling remote devices
CN105379409A (en) EHF secure communication device
CN105094080B (en) System and method for dynamic body tracking and multi-tagging in an access control system
US10346604B2 (en) Light sequenced card validation
US10867490B2 (en) Object for theft detection
JP2018036920A (en) Obstacle outside visual field detection system
EP1325397B1 (en) Annotating imaged data product
Abd-Elminaam Smart life saver system for alzheimer patients, down syndromes, and child missing using IoT
JP2005107792A (en) Ic tag, article management system using it, entrance/exit management system using it, and tracking system using it
JP2008070923A (en) Authentication system, authentication method and terminal
KR20050089774A (en) Apparatus and method for searching a lost child via electric label
KR20160124543A (en) System and method for monitoring separated position of object by mobile device with antenna
Bhuvaneswary et al. IoT-based smart helmet for riders
US20050287986A1 (en) System and method for securing mobile equipment
Dejen DEVELOPMENT OF AN ANDROID FRAMEWORK FOR VISUAL NAVIGATION AND ORIENTATION
Aggarwal et al. All-in-One Companion for Visually Impaired
Pérez et al. Intelligent Jacket for Monitoring Mobility of People with Reduced Disabilities
US20180137739A1 (en) Item locator
KR20210070644A (en) Vehicle safety Sign Identification System Applied with Smart Code for Autonomous Vehicles

Legal Events

Date Code Title Description
STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: REIFFEL TECHNOLOGIES, LLC, ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:REIFFEL, LEONARD;REEL/FRAME:020064/0941

Effective date: 20070703

Owner name: REIFFEL TECHNOLOGIES, LLC,ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:REIFFEL, LEONARD;REEL/FRAME:020064/0941

Effective date: 20070703

REMI Maintenance fee reminder mailed
FPAY Fee payment

Year of fee payment: 4

SULP Surcharge for late payment
FPAY Fee payment

Year of fee payment: 8

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20190227