US20170038848A1 - User identification via collected sensor data from a wearable fitness monitor - Google Patents

User identification via collected sensor data from a wearable fitness monitor Download PDF

Info

Publication number
US20170038848A1
US20170038848A1 US15/231,620 US201615231620A US2017038848A1 US 20170038848 A1 US20170038848 A1 US 20170038848A1 US 201615231620 A US201615231620 A US 201615231620A US 2017038848 A1 US2017038848 A1 US 2017038848A1
Authority
US
United States
Prior art keywords
motion
user
fitness monitor
wearable fitness
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/231,620
Inventor
Shelten Gee Jao Yuen
James Park
Atiyeh Ghoreyshi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fitbit LLC
Original Assignee
Fitbit LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fitbit LLC filed Critical Fitbit LLC
Priority to US15/231,620 priority Critical patent/US20170038848A1/en
Assigned to FITBIT, INC. reassignment FITBIT, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GHOREYSHI, ATIYEH, YUEN, SHELTEN GEE JAO, PARK, JAMES
Publication of US20170038848A1 publication Critical patent/US20170038848A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0002Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/024Detecting, measuring or recording pulse rate or heart rate
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/024Detecting, measuring or recording pulse rate or heart rate
    • A61B5/02416Detecting, measuring or recording pulse rate or heart rate using photoplethysmograph signals, e.g. generated by infrared radiation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/02Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
    • A61B5/024Detecting, measuring or recording pulse rate or heart rate
    • A61B5/02438Detecting, measuring or recording pulse rate or heart rate with portable devices, e.g. worn by the patient
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1112Global tracking of patients, e.g. by using GPS
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1118Determining activity level
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/112Gait analysis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/11Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
    • A61B5/1123Discriminating type of movement, e.g. walking or running
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/117Identification of persons
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/44Detecting, measuring or recording for evaluating the integumentary system, e.g. skin, hair or nails
    • A61B5/441Skin evaluation, e.g. for skin disorder diagnosis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6802Sensor mounted on worn items
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6802Sensor mounted on worn items
    • A61B5/681Wristwatch-type devices
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/725Details of waveform analysis using specific filters therefor, e.g. Kalman or adaptive filters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7253Details of waveform analysis characterised by using transforms
    • A61B5/7257Details of waveform analysis characterised by using transforms using Fourier transforms
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7235Details of waveform analysis
    • A61B5/7264Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
    • A61B5/7267Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7271Specific aspects of physiological measurement analysis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/742Details of notification to user or communication with user or patient ; user input means using visual displays
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/7475User input or interface means, e.g. keyboard, pointing device, joystick
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B24/00Electric or electronic controls for exercising apparatus of preceding groups; Controlling or monitoring of exercises, sportive games, training or athletic performances
    • A63B24/0062Monitoring athletic performances, e.g. for determining the work of a user on an exercise apparatus, the completed jogging or cycling distance
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63BAPPARATUS FOR PHYSICAL TRAINING, GYMNASTICS, SWIMMING, CLIMBING, OR FENCING; BALL GAMES; TRAINING EQUIPMENT
    • A63B69/00Training appliances or apparatus for special sports
    • A63B69/36Training appliances or apparatus for special sports for golf
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C22/00Measuring distance traversed on the ground by vehicles, persons, animals or other moving solid bodies, e.g. using odometers, using pedometers
    • G01C22/006Pedometers
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01PMEASURING LINEAR OR ANGULAR SPEED, ACCELERATION, DECELERATION, OR SHOCK; INDICATING PRESENCE, ABSENCE, OR DIRECTION, OF MOVEMENT
    • G01P1/00Details of instruments
    • G01P1/02Housings
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01PMEASURING LINEAR OR ANGULAR SPEED, ACCELERATION, DECELERATION, OR SHOCK; INDICATING PRESENCE, ABSENCE, OR DIRECTION, OF MOVEMENT
    • G01P13/00Indicating or recording presence, absence, or direction, of movement
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S19/00Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S19/00Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
    • G01S19/01Satellite radio beacon positioning systems transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
    • G01S19/13Receivers
    • G01S19/14Receivers specially adapted for specific applications
    • G01S19/19Sporting applications
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • G06F1/16Constructional details or arrangements
    • G06F1/1613Constructional details or arrangements for portable computers
    • G06F1/163Wearable computers, e.g. on a belt
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/32User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
    • G06Q50/01Social networking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09BEDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
    • G09B5/00Electrically-operated educational appliances
    • G09B5/02Electrically-operated educational appliances with visual presentation of the material to be studied, e.g. using film strip
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2560/00Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
    • A61B2560/02Operational features
    • A61B2560/0223Operational features of calibration, e.g. protocols for calibrating sensors
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2562/00Details of sensors; Constructional details of sensor housings or probes; Accessories for sensors
    • A61B2562/02Details of sensors specially adapted for in-vivo measurements
    • A61B2562/0219Inertial sensors, e.g. accelerometers, gyroscopes, tilt switches
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0002Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
    • A61B5/0015Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
    • A61B5/0022Monitoring a patient using a global network, e.g. telephone networks, internet
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0002Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
    • A61B5/0015Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system
    • A61B5/0024Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by features of the telemetry system for multiple sensor units attached to the patient, e.g. using a body or personal area network
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/103Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
    • A61B5/1032Determining colour for diagnostic purposes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/117Identification of persons
    • A61B5/1171Identification of persons based on the shapes or appearances of their bodies or parts thereof
    • A61B5/1172Identification of persons based on the shapes or appearances of their bodies or parts thereof using fingerprinting
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/318Heart-related electrical modalities, e.g. electrocardiography [ECG]
    • A61B5/332Portable devices specially adapted therefor
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6844Monitoring or controlling distance between sensor and tissue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • G06F2218/10Feature extraction by analysing the shape of a waveform, e.g. extracting parameters relating to peaks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/15Biometric patterns based on physiological signals, e.g. heartbeat, blood flow
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients

Definitions

  • Methods and systems are provided for determining the identification of a user of a wearable fitness monitors or authenticating the user of the wearable fitness monitors.
  • One aspect of the disclosure relates to methods for using two motion signatures to determine an identity of a wearer of a wearable fitness monitor.
  • a method comprising: (a) obtaining a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a second motion signature obtained using data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor; (c) comparing the first and second motion signatures or a combination thereof to a reference motion feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • the one or more first motion sensors comprise an accelerometer, a gyroscope, a magnetometer, an altimeter, a GPS receiver, or any combination thereof.
  • the one or more second motion sensors comprise one or more of the first motion sensors.
  • the first motion signature is a time domain representation of a periodic motion of the movement experienced by the wearable fitness monitor.
  • the periodic motion is produced by a movement of the wearer, wherein the movement is a step rate, a metric derived from an amplitude of the time domain representation, a biking cadence, a rowing rate, a resistance-based repetition rate, a typing speed, a zero crossing rate, a peak-to-peak time, an arm swing rate, and any combination thereof.
  • the first motion signature comprises a frequency domain representation of a periodic motion of the movement experienced by the wearable fitness monitor.
  • the frequency domain representation of the periodic motion of the movement experienced by the wearable fitness monitor comprises a spectral component in the periodic motion.
  • the spectral component is a harmonic in the frequency domain representation of the periodic motion.
  • the first motion signature comprises a property of two or more harmonics in the frequency domain representation of the periodic motion.
  • the first motion signature comprises a motion periodicity and the second motion signature comprises a metric derived from an amplitude of the motion periodicity.
  • the first motion signature comprises a time domain representation of a periodic motion and the second motion signature comprises a frequency domain representation of the periodic motion.
  • the user's reference motion feature comprises a model of typical motion of the user.
  • the user's reference motion feature is generated using data obtained from the one or more first motion sensors and the one or more second motion sensors when the user wears the wearable fitness monitor.
  • the user's reference motion feature comprises a profile of a step by the user.
  • operation (c) of the method comprises comparing a combination of the first and second motion signatures to the user's reference motion feature, and wherein the user's reference motion feature comprises a line, a curve, or a look up table relating the first and second motion signatures for the user.
  • (c) comprises (i) determining a difference between the user's reference motion feature and the first and second motion signatures or the combination thereof, and (ii) determining whether the difference is greater than a threshold.
  • (c) comprises performing a linear discriminant analysis on the first and second motion signatures or the combination thereof with respect to the user's reference motion feature.
  • (c) comprises determining that at least one of the first and second motion signatures is an invalid motion for a human wearer.
  • the method further comprises determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user, wherein the one or more second motion sensors are located on a device that is separate from the wearable fitness monitor.
  • the separate device is a mobile phone.
  • determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user comprises determining whether the first and second motion signatures represent a characteristic of a periodic motion.
  • the first motion signature comprises a step count or a step rate and the second motion signature comprises a GPS or Bluetooth signature.
  • At least one of the first and second motion signatures comprises a cycle profile of a periodic motion performed by the user, and wherein the reference motion feature is a predetermined typical cycle for the user's periodic motion.
  • the cycle profile comprises a time varying amplitude of an output from the one or more first motion sensors.
  • the user's periodic motion is selected from the group consisting of running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, and riding an animal.
  • the method further includes repeating (a)-(d) at multiple times. In some implementations, wherein repeating (a)-(d) is performed automatically, without triggering by the wearer of the fitness monitor. In some implementations, the method further includes, responsive to determining that the identity of the wearer of the fitness monitor is not the user, preventing the wearable fitness monitor from allowing a transaction. In some implementations, the transaction comprises accessing a secure item or providing the user with an award for meeting an activity threshold.
  • the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is not the user, requiring the user to authenticate himself or herself.
  • requiring the user to authenticate comprises requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing.
  • the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is not the user, discrediting a fitness metric obtained for the user via the wearable fitness monitor.
  • the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is the user, crediting a fitness metric obtained for the user via the wearable fitness monitor.
  • the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is the user, allowing the wearable fitness monitor to facilitate a transaction.
  • the transaction comprises accessing a secure item or providing the user with an award for meeting an activity threshold.
  • At least one of the one or more first motion sensors and the one or more second motion sensors are the same sensors.
  • (c) comprises obtaining a function between the first motion signature and the second motion signature, and comparing the function to a reference function based on the reference motion feature.
  • (c) comprises obtaining an average motion signature by averaging the first motion signature and the second motion signature, and comparing the average motion signature to the reference motion feature.
  • (c) comprises: extracting features from the first motion signature and the second motion signature; forming a feature vector using the extracted features, and applying a classifier to the feature vector to determine whether the feature vector belongs to a class corresponding to the reference motion feature.
  • the obtaining the first motion signature obtained using the data from the one or more first motion sensors comprises: low-pass filtering the data from the one or more first motion sensors; and obtaining a cycle profile from the low-pass filtered data.
  • the obtaining the cycle profile from the low-pass filtered data comprises: obtaining local minima from the low-passed filtered data; dividing the low-passed filtered data into two or more segments using the local minima; and obtaining the cycle profile from the two or more segments.
  • the obtaining the cycle profile from the two or more segments comprises: (i) rejecting one or more outliers among the two or more segments that deviate from the mean of the two or more segments.
  • the obtaining the cycle profile from the two or more segments further comprises: repeating (i) one or more times among remaining segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: averaging remaining segments to obtain the cycle profile.
  • the obtaining the first motion signature obtained using the data from the one or more first motion sensors further comprises: extracting one or more features from the cycle profile or values derived from the cycle profile.
  • each feature is selected from the group consisting of: a slope, an inflection, a zero crossing, a derivative, a moment, a cumulant, and any combination thereof.
  • comparing the first motion signature to the reference motion feature for the user comprises: obtaining a classifier using motion data obtained from the user; and applying the classifier to the extracted one or more features, wherein the classifier takes the one or more features as inputs and provides a classification of the wearer being the user or not the user as an output.
  • the classifier comprises a linear discriminant analysis classifier. In some implementations, the classifier comprises a neural network classifier. In some implementations, the classifier is trained using at least one cycle profile derived from motion data obtained from the user. In some implementations, the method further involves updating the classifier using additional motion data obtained from the user.
  • a method involves: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a heartbeat waveform signature obtained using data from one or more heartbeat waveform sensors, wherein the heartbeat waveform signature characterizes a detected heartbeat waveform of a wearer of the wearable fitness monitor; (c) comparing the motion signature and the heartbeat waveform signature or a combination thereof to one or more reference features of a user; and (d) based on the comparison in (c), determining whether an identity of the wearer of the fitness monitor is the user.
  • a method of identifying a user involves: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a body characteristic obtained using data from one or more body characteristic sensors, wherein the body characteristic characterizes the body of a person wearing the wearable fitness monitor; (c) comparing the motion signature and the body characteristic or a combination thereof to at least one reference feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • the body characteristic is a detected response of the one or more body characteristic sensors to the wearer's skin.
  • at least one of the one or more body characteristic sensors comprises a light pulse emitter and a light pulse detector configured to determine a variable response of the detector to a variable intensity of light pulses from the emitter.
  • the body characteristic is body composition determined through bioelectrical impedance. In some implementations, at least one of the one or more body characteristic sensors is disposed on the wearable fitness monitor.
  • a further aspect of the disclosure relates to methods for using a motion signature to determine an identity of a wearer of a wearable fitness monitor.
  • a method includes: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) comparing the motion signature to a reference motion feature for a user; and (c) based on the comparison in (b), determining whether an identity of a wearer of the fitness monitor is the user.
  • the motion signature characterizes a cycle of periodic movement of the person wearing the wearable fitness monitor.
  • the reference motion feature is a reference cycle for periodic movement of a user.
  • the reference cycle is a predetermined typical cycle for the user's periodic motion.
  • the motion signature comprises a time-varying amplitude of an output from the one or more motion sensors.
  • the user's periodic motion is selected from the group consisting of running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, and riding an animal.
  • the reference motion feature is a characteristic of a periodic motion.
  • the reference motion feature is a metric derived from an amplitude.
  • the obtaining the motion signature obtained using the data from the one or more motion sensors comprises: low-pass filtering the data from the one or more motion sensors; and obtaining a cycle profile from the low-pass filtered data.
  • the obtaining the cycle profile from the low-pass filtered data comprises: obtaining local minima from the low-passed filtered data; dividing the low-passed filtered data into two or more segments using the local minima; and obtaining the cycle profile from the two or more segments.
  • the obtaining the cycle profile from the two or more segments comprises: (i) rejecting one or more outliers among the two or more segments that deviate from the mean of the two or more segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: repeating (i) one or more times among remaining segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: averaging remaining segments to obtain the cycle profile.
  • the obtaining the motion signature obtained using the data from the one or more motion sensors further comprises: extracting one or more features from the cycle profile or values derived from the cycle profile.
  • each feature is selected from the group consisting of: a slope, an inflection, a zero crossing, a derivative, a moment, a cumulant, and any combination thereof.
  • comparing the motion signature to the reference motion feature for the user comprises: obtaining a classifier using motion data obtained from the user; and applying the classifier to the extracted one or more features, wherein the classifier takes the one or more features as inputs and provides a classification of the wearer being the user or not the user as an output.
  • the classifier comprises a linear discriminant analysis classifier.
  • An additional aspect of the disclosure relates to methods for using a motion signature to determine that the motion of a wearable fitness monitor is generated by non-human, and preventing a transaction based on the determination.
  • a method includes: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a body movement of the person wearing the wearable fitness monitor; (b) determining whether the motion signature corresponds to an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human; and (c) based on the determination in (b), preventing the wearable fitness monitor from allowing a transaction.
  • the transaction comprises accessing a secure item or providing an award for meeting an activity threshold to a user associated with the wearable fitness monitor.
  • the method further involves, responsive to determining that the motion signature corresponds to the invalid motion feature, requiring a wearer of the fitness monitor to authenticate himself or herself
  • requiring the wearer to authenticate comprises requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing.
  • determining whether the motion signature corresponds to an invalid motion feature in (b) comprises: obtaining an additional signature using data from one or more additional sensors, and determining that the motion signature, the additional signature, or a combination thereof is inconsistent with at least one human activity.
  • the motion signature is step rate or step count and the additional signature is a heart rate or a heartbeat waveform.
  • determining whether the motion signature corresponds to the invalid motion feature in (b) comprises determining whether a periodicity for the motion signature is within a threshold periodicity for a given time period.
  • the invalid motion feature is a periodic motion having a cycle-to-cycle consistency greater than a threshold.
  • the invalid motion feature comprises one or more periodic motion contributions from one or more spatial dimensions that is less than a threshold.
  • a method involves: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a user, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining an additional signature obtained using the data from one or more additional sensors located on a device that is separate from the wearable fitness monitor, wherein the additional motion signature further characterizes the movement; (c) comparing the motion signature to the additional signature; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • the motion signature and the additional signature are obtained from the data collected at the same time. In some implementations, the comparing in (c) comprises determining whether the motion signature and the additional signature represent the same activity or activity level of the user. In some implementations, determining whether the motion signature and the additional signature represent the same activity or activity level of the user comprises determining whether the motion signature and the additional signature represent a characteristic of the user's gait.
  • the separate device is a mobile phone.
  • the motion signature comprises a step count or a step rate and the second motion signature comprises a GPS or Bluetooth signature.
  • a system includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, one or more second motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic.
  • the classification logic is configured to: (a) obtain a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtain a second motion signature obtained using data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor; (c) compare the first and second motion signatures or a combination thereof to a reference motion feature for a user; and (d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
  • a system includes (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, one or more heartbeat waveform sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic.
  • the classification logic is configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor, (b) obtain a heartbeat waveform signature obtained using data from the one or more heartbeat waveform sensors, wherein the heartbeat waveform signature characterizes a detected heartbeat waveform of a wearer of the wearable fitness monitor, (c) compare the motion signature and the heartbeat waveform signature or a combination thereof to one or more reference features of a user, and (d) based on the comparison in (c), determine whether an identity of the wearer of the fitness monitor is the user.
  • a system for identifying a user includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, one or more body characteristic sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic.
  • the classification logic is configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor, (b) obtain a body characteristic obtained using data from the one or more body characteristic sensors, wherein the body characteristic characterizes the body of a wearer of the wearable fitness monitor, (c) compare the motion signature and the body characteristic or a combination thereof to at least one reference feature for a user, and (d) based on the comparison in (c), determine whether an identity of the wearer of the wearable fitness monitor is the user.
  • a system includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor, (b) compare the motion signature to a reference motion feature for a user, and (c) based on the comparison in (b), determine whether an identity of a wearer of the fitness monitor is the user.
  • a system includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a body movement of a wearer of the wearable fitness monitor, (b) comparing the motion signature to an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human; and (c) based on the determination in (b), prevent the wearable fitness monitor from allowing a transaction.
  • a system includes: a wearable fitness monitor configured to be worn by a user and comprising: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; one or more additional sensors located on a device that is separate from the wearable fitness monitor; and classification logic.
  • the classification logic is configured to: (a) obtain a motion signature obtained using data from the one or more first motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtain an additional signature obtained using data from the one or more additional sensors, wherein the additional motion signature further characterizes the movement; (c) compare the motion signature to the additional signature; and (d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
  • FIG. 1 depicts a generalized schematic of an example wearable fitness monitor with which various operations described herein may be executed.
  • FIG. 2 shows one implementation of a wearable fitness monitor having an accelerometer, illustrating a coordinate system of the accelerometer.
  • FIG. 3 shows a representative 3-axis accelerometer signal from a wearable fitness monitor worn on the wrist of a user who is walking.
  • FIG. 4 depicts the step rate and signal energy for 4 unique users.
  • FIG. 5 depicts the step rate and signal energy for 3 unique users for a run.
  • FIG. 6 depicts power spectral density of motion data of two unique users who are otherwise indistinguishable in running by step rate and signal energy.
  • FIG. 7 shows examples of heart beat waveform data obtained from post-processing.
  • FIG. 8 shows a schematic illustration of a PQRST heart beat waveform.
  • FIG. 9 illustrates a number of time domain features of a PQRST waveform.
  • FIGS. 10A-10D depict representative features of the PPG waveform that may be used to identify the user.
  • FIG. 11 shows a flowchart of a method for determining whether an identity of an instant wearer of a fitness monitor matches that of a user.
  • FIG. 12 shows a process 1200 for training an LDA classifier and using the classifier to authenticate a wearer based on the classification result.
  • FIG. 13 shows an example of acceleration data as a function of time.
  • FIG. 14 shows motion data depicting multiple stride profiles from a same subject.
  • FIG. 15 shows eight mean stride cycle profiles for eight different subjects.
  • FIG. 16 shows an example of classification results for two-minute walking data for eight subjects.
  • This disclosure relates to methods, devices, and systems to recognize a user of a wearable fitness monitor using information obtained using the wearable fitness monitor.
  • the wearable fitness monitor can provide data to the correct user's digital account.
  • the wearable fitness monitor can estimate the identity of the user with some amount of certainty.
  • the wearable fitness monitor When data from the wearable fitness monitor is used to provide monetary incentives for user behavior, it may be useful for some embodiments that the wearable fitness monitor knows the identity of the user with some certainty. Additionally, trust of the user data itself (e.g., that a “step” is a true user step rather than a fake motion, hereinafter referred to as a “fake”) can be an important aspect to some embodiments.
  • the veracity of the data and the identity of the user can be an important feature to verify, which can ensure that the data used to compete is not faked.
  • this disclosure presents a method including the following operations: (a) obtaining a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a second motion signature obtained using the data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor; (c) comparing the first and second motion signatures or a combination thereof to a reference motion feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • the one or more first motion sensors include an accelerometer, a gyroscope, a magnetometer, an altimeter, a GPS receiver, or any combination thereof.
  • a “combination” of motion signatures in (c) may be used in cases where the signatures are not analyzed separately, but together as in the case of a point or a curve in a multi-dimensional signature space.
  • each of the first and second signatures is separately compared against the reference motion signature.
  • the “movement” experienced by the wearable fitness monitor is a voluntary body movement such as the wearer's intentional movement of her head, neck, eyelid, mouth, shoulder, arm, wrist, finger, torso, hips, knee, ankle, and/or toe.
  • the movement is characterized by a wearer's gait or stride when walking or running.
  • the movement may be associated with a particular activity type such as running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc.
  • a particular activity type such as running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc.
  • the movement does not include involuntary motions such as heartbeats which may be determined using, e.g., sensor technology include photoplethysmography (PPG), electrocardiography (ECG), etc
  • the comparison of the signature(s) and the reference motion signature may indicate that the wearer is not the user. Such cases may arise where the wearer is a different individual than the user and cases where the “wearer” is a robot or other automaton.
  • the “reference motion feature” used in operation (c) may be a user reference motion feature (a template built using historical motion data of the user).
  • determining whether an identity of the wearer of the fitness monitor is the user may be a matter of probability or a prediction of whether the user is likely wearing the monitor.
  • the determination may be made by mathematical or other logical techniques that determine a distance or difference between the wearer's current motion signature(s) and the user's reference motion feature in movement signature space.
  • the determination may rely on a classification technique which provides a likelihood, rather than certainty, that the wearer is the user. For instance, classifiers using one or more of the following techniques may be used to determine the likelihood that the wearer is the user: linear discriminant analysis, neural network, clustering techniques, support vector machine, logistic regression, naive Bayes, random forest, decision tree, etc.
  • the sensor producing the data for the first and second motion signatures may be the same or different or they may overlap, with some first motion sensors being the same as some second motion sensors.
  • the one or more second motion sensors include at least one motion sensor from the one or more first motion sensors.
  • the first motion signature is a time domain representation of a periodic motion of the movement experienced by the wearable fitness monitor, which periodic motion may be a person's gait, which may be, for example, a step rate, a metric derived from an amplitude of the wearer's periodic motion, a biking cadence, a rowing rate, a resistance-based repetition rate (for, e.g., weightlifting), a typing speed, a zero-crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof.
  • periodic motion may be a person's gait, which may be, for example, a step rate, a metric derived from an amplitude of the wearer's periodic motion, a biking cadence, a rowing rate, a resistance-based repetition rate (for, e.g., weightlifting), a typing speed, a zero-crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof.
  • the first motion signature is a frequency domain representation of a person's periodic motion (e.g., gait), which may include, for example, contributions of a spectral component (e.g., the fundamental frequency or a harmonic thereof) in the first motion signature.
  • the spectral component includes a combination of the harmonics or the fundamental frequency and one or more harmonics.
  • a combination of the contributions of harmonics may be a property of two or more harmonics (or the fundamental frequency) such as a ratio of the powers of the individual harmonics.
  • the first motion signature includes a motion periodicity parameter (e.g., step rate) and the second motion signature includes a metric derived from an amplitude of the wearer's periodic motion.
  • the first motion signature includes a time domain representation of a periodic motion (e.g., a wearer's gait) and the second motion signature includes a frequency domain representation of the periodic motion.
  • the user's reference motion feature typically has characteristics that facilitate comparison with the motion signatures.
  • the user's reference motion feature may include a predetermined typical motion signature for the user.
  • Such feature may be obtained in various ways such as by using data obtained from the one or more first motion sensors and the one or more second motion sensors when the user wears the wearable fitness monitor. In such cases, it should be established that the user is actually wearing a fitness monitor when capturing data for generating her reference motion feature.
  • the user's reference motion feature comprises a profile of a step by the user.
  • comparing the motion signature(s) to the user's reference motion feature in (c) includes comparing a combination of the first and second motion signatures to the user's reference motion feature.
  • the user's reference motion feature may be a relationship between the first and second motion signatures for the user.
  • the relationship may be a line, a curve, a look up table, etc. relating the first and second motion signatures for the user.
  • comparing the motion signature(s) to the reference motion feature includes (i) determining a distance (or difference) between the user's reference motion feature and the first and second motion signatures or the combination thereof, and (ii) determining whether the distance (or difference) is greater than a threshold.
  • the distance is a minimum separation between two points, lines, surfaces, etc. in multidimensional analysis. It may be viewed as a type of “difference.”
  • the comparison in (c) includes performing a linear discriminant analysis (LDA) on the first and second motion signatures or the combination thereof with respect to the user's reference motion feature.
  • LDA linear discriminant analysis
  • the comparison in (c) includes determining that at least one of the first and second motion signatures is an invalid motion for a human user (e.g., it is unnatural for a human user).
  • a machine provides movements from which the fitness monitor generates data having unnaturally high consistency over time or from cycle to cycle.
  • data produced from machine movements may be unnaturally limited to contributions from one or a few axes (e.g., one axis of a three-axis accelerometer or gyroscope).
  • the comparison between motion signatures and a reference motion feature may be performed by various classification techniques such as LDA, neural network, clustering, logistic regression, support vector machine, naive Bayes, etc.
  • the method includes an operation of determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user. In some implementations, determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user includes determining whether the first and second motion signatures represent a characteristic of a periodic motion. For example, the motion signatures may be compared to determine whether they represent the same gait of a wearer. As an example, the first motion signature includes a step count or a step rate and the second motion signature includes a GPS or Bluetooth signature.
  • the one or more second motion sensors are located on a device that is separate from the wearable fitness monitor, such as a smart phone or a second monitoring device worn or carried by the wearer of the fitness monitor.
  • the separate device is a mobile phone or other portable device with one or more sensors.
  • the motion signatures may be or include representations of multiple cycles of the wearer's movement or even a single cycle.
  • at least one of the first and second motion signatures may include a cycle profile of a periodic motion performed by the user.
  • the reference motion feature may be a predetermined typical cycle for the user's periodic motion.
  • the cycle profile is a time varying amplitude of an output from the one or more first motion sensors.
  • the user's periodic motion is running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc.
  • the process of obtaining the wearer's motion signatures and comparing them with the user's reference motion feature is performed repeatedly, sometime continuously.
  • the above method may involve repeating operations (a)-(d) at multiple times.
  • the repeating of operations (a)-(d) may be performed automatically, without triggering by the wearer of the fitness monitor.
  • Determining whether or not the wearer of the fitness monitor is the user can be used in various contexts, often in the same or similar way as biometric information is conventionally used.
  • various actions may be taken to block a transaction involving the wearer, require the wearer to take additional steps to authenticate or otherwise identify herself, etc.
  • the method responsive to determining that the identity of the wearer of the fitness monitor is not the user, the method prevents the wearable fitness monitor from allowing a transaction.
  • the transaction may be accessing a secure item or providing the user with an award for meeting an activity threshold, which may be determined from quantifiable biometric information.
  • the method responsive to determining that the identity of the wearer of the fitness monitor is not the user, the method requires the user to authenticate himself or herself.
  • requiring the user to authenticate may include requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing.
  • the method responsive to determining that the identity of the wearer of the fitness monitor is not the user, the method discredits a fitness metric obtained for the user via the wearable fitness monitor.
  • various actions may be taken to credit the user or allow the user to engage in a transaction.
  • the method credits a fitness metric obtained for the user via the wearable fitness monitor.
  • the method allows the wearable fitness monitor to facilitate a transaction. Examples of such transactions include accessing a secure item or providing the user with an award for meeting an activity threshold.
  • At least one of the one or more first motion sensors and the one or more second motion sensors are the same sensors.
  • the data from the first and second motions sensors includes at least a first datum from the one or more first motion sensors and a second datum from the one more second motion sensors.
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a heartbeat waveform signature obtained using data from one or more heartbeat waveform sensors, wherein the heartbeat waveform signature characterizes a detected heartbeat waveform of a wearer of the wearable fitness monitor; (c) comparing the motion signature and the heartbeat waveform signature or a combination thereof to one or more reference features of a user; and (d) based on the comparison in (c), determining whether an identity of the wearer of the fitness monitor is the user.
  • a wearer is a non-human wearer and, in certain cases, a “heartbeat waveform” is detected where the non-human lacks a heartbeat.
  • the method may be employed to identify the user, deauthenticate the user, etc. as described above.
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a body characteristic obtained using data from one or more body characteristic sensors, wherein the body characteristic characterizes the body of a person wearing the wearable fitness monitor; (c) comparing the motion signature and the body characteristic or a combination thereof to at least one reference feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • the method may be employed to identify the user, deauthenticate the user, etc. as described above.
  • the body characteristic is one or more characteristics of the wearer's skin (e.g., the wearer's skin color). Skin color or another skin characteristic can be determined (or approximated) by various techniques.
  • at least one of the one or more body characteristic sensors includes a light pulse emitter and a light pulse detector configured to determine a variable response of the detector to a intensity of light pulses from the emitter. The variability of the response of the detector is can be influenced by the user's skin color. Thus, this response may be used as a signature that is compared against a reference feature of a user.
  • the body characteristic is body composition determined through bioelectrical impedance.
  • at least one of the one or more body characteristic sensors is disposed on the wearable fitness monitor.
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, where the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) comparing the motion signature to a reference motion feature for a user; and (c) based on the comparison in (b), determining whether an identity of a wearer of the fitness monitor is the user.
  • data from the one or more motion sensors are preprocessed before the motion signature is obtained.
  • the reference motion feature may be updated continuously or periodically after the identity of the wearer has been determined.
  • the comparison in operation (b) is implemented using an LDA classifier.
  • the motion signature characterizes a cycle of periodic movement of the person wearing the wearable fitness monitor.
  • a cycle contains information about a wearer's step or other unit of periodic motion. Examples of the user's periodic motion include running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc.
  • a motion signature for a cycle includes a time-varying amplitude of an output from the one or more motion sensors.
  • the motion signature for a cycle may be obtained for data generated during a single instance of the cycle or from multiple instances, with the instances being averaged or otherwise combined to provide the “cycle” used in this method.
  • the reference motion feature may be a reference cycle for periodic movement of a user. In some cases, such reference cycle is a predetermined typical cycle for the user's periodic motion. The method may be employed to identify the user, deauthenticate the user, etc. as described above.
  • the reference motion feature is a characteristic of a periodic motion.
  • the reference motion feature is a metric derived from an amplitude (e.g., an amplitude a reference cycle).
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, where the motion signature characterizes a body movement of the person wearing the wearable fitness monitor; (b) determining whether the motion signature corresponds to an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human; and (c) based on the determination in (b), preventing the wearable fitness monitor from allowing a transaction (or deauthenticating a user associated with the fitness monitor).
  • the transaction includes accessing a secure item or providing an award for meeting an activity threshold to a user associated with the wearable fitness monitor.
  • the invalid motion feature is simply an unnatural (for humans or non-machines) level of consistency or repetition in the motion signature.
  • an invalid motion feature has limited dimensional range; for example, the motion feature unnaturally (for a human or non-machine) emphasizes one or two spacial dimensions.
  • the user identification logic can identify this level of consistency or limited dimensional range by comparing the relevant motion signature component to a defined threshold.
  • unnatural motion signatures might be generated by automatons or other machines. Additionally, an unnatural motion signature may be generated by a non-human animal wearing the fitness monitor.
  • an approach as described here allows a system or entity to deauthenticate a user, discredit a user, prevent access to a secure item, and/or prevent a transaction involving the user. For example, in certain embodiments, responsive to determining that the motion signature corresponds to an invalid motion feature, the method requires a wearer of the fitness monitor to authenticate himself or herself.
  • requiring the wearer to authenticate includes requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing.
  • operation (c) includes determining that the motion signature matches an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human.
  • determining whether the motion signature corresponds to an invalid motion feature in (b) includes obtaining an additional signature using data from one or more additional sensors, and determining that the motion signature and/or the additional signature are/is inconsistent with a human activity.
  • the motion signature is step rate or step count and the additional signature is a heart rate or a heartbeat waveform.
  • determining whether the motion signature corresponds to the invalid motion feature in (b) includes determining whether a periodicity for the motion signature is within a threshold periodicity for a given time period.
  • determining whether the motion signature corresponds to the invalid motion feature in (b) includes determining a relation (e.g., a function) between step rate and heart rate for human activities, and determining that a combination of the motion signature (e.g., measured step rate) and the additional signature (e.g., measured heart rate) is inconsistent with the relation determined for one or more human activities.
  • a measured step rate is compared to a reference step rate range associated with human activities to determine whether the measure step rate is outside of the step rate range of human activities.
  • measured heart rate is compared to a heart rate range of human activities to determine whether the measured heart rate is outside of the heart rate range of human activities.
  • Another aspect of the disclosure concerns a method including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining an additional signature obtained using data from one or more additional sensors located on a device that is separate from the wearable fitness monitor, wherein the additional motion signature further characterizes the movement experienced by the wearable fitness monitor, or characterizes movement experienced by the device that is separate from the wearable fitness monitor; (c) comparing the motion signature to the additional signature; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • the motion signature and the additional signature are obtained from the data collected at the same time.
  • the comparing in (c) includes determining whether the motion signature and the additional signature represent the same activity or activity level of the user. In some implementations, determining whether the motion signature and the additional signature represent the same activity or activity level of the user comprises determining whether the motion signature and the additional signature represent a characteristic of the user's gait.
  • the separate device is a mobile phone, a second fitness monitor, a headset, or other portable device.
  • the motion signature includes a step count or a step rate and the second motion signature includes a GPS or Bluetooth signature.
  • One aspect of the disclosure relates to systems that use sensor data to verify the identity of a wearer of a wearable fitness monitor.
  • the sensor data include data from one or more motion sensors of the wearable fitness monitor.
  • Various implementations of the systems are configured to perform any of the methods described above.
  • the system includes a wearable fitness monitor configured to be worn by a person.
  • the wearable fitness monitor includes one or more first motion sensors, one or more second motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor.
  • the system also includes classification logic configured to perform operations to implement one or more methods described above.
  • the classification logic is configured to: (a) obtain a first motion signature obtained using data from the one or more first motion sensors of the wearable fitness monitor, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor, (b) obtain a second motion signature obtained using the data from the one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor, (c) compare the first and second motion signatures or a combination thereof to a reference motion feature for a user, and (d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
  • the wearable fitness monitor is configured to be worn by a person and includes: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor.
  • the wearable fitness monitor also includes one or more heartbeat waveform sensors.
  • the wearable fitness monitor also includes one or more body characteristic sensors.
  • the wearable fitness monitor includes: one or more first motion sensors, one or more body characteristic sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor.
  • the system includes (a) a wearable fitness monitor configured to be worn by a person.
  • the wearable fitness monitor includes: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor.
  • the system also includes (b) classification logic, and (c) one or more sensors located on a device that is separate from the wearable fitness monitor.
  • the methods and systems described herein identify users of fitness monitors that produce data from a motion sensor that responds to the user's movements (e.g., motion of a limb, head, torso, wrist, etc.).
  • Such voluntary movements are typically produced for purposes other than user identification or authentication, typically for the purpose of activity tracking and monitoring. They may be generated from the user's normal walking or fitness activities.
  • Motion signatures may include any data relating motion periodicity (otherwise referred to as periodic motion) obtained from the motion sensor are analyzed by, e.g., classification logic to determine whether the detected motions are those of the user.
  • data relating to motion periodicity include step rate, a metric derived from the amplitude of the motion signal, a biking cadence, a rowing rate, a resistance-based repetition rate (e.g., weightlifting repetition rates), a typing speed, a zero crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof.
  • a metric derived from the amplitude of the motion signal may be referred to herein as “motion signal energy.”
  • Motion energy may be calculated through various methods, such as classical energy formulas as well as simplified calculations, such as finding the absolute difference of motion signals.
  • One or more motion signatures may be used to identify a user. When multiple motion signatures are used, they may be used separately or in combination to identify a user.
  • One example of a combination is a line or curve relating two or more signature types for a user. For example, a user's motion signal energy may vary as a function of step rate in a reproducible manner.
  • Another example of a combination is where the values of multiple motion signatures are hashed or otherwise used to index a lookup table.
  • sensors other than motion sensors in the wearable fitness monitor are used to help identify the user wearing the fitness monitor. Such other sensors include heartbeat waveform sensors (PPG sensors and ECG sensors), bioimpedance sensors, and EEGs, the like. In some cases, such sensors measure involuntary body motions such a heart beats and respiration.
  • An example of combination of a motion signature and another signature is a relationship between heartrate and step rate for a user.
  • a positive or negative identification of a user may be applied in various contexts such as insurance (e.g., an insured's fitness level factors into her premium), automatically switching users sharing a wearable fitness device, accessing secure devices such as automobile ignition systems, door locks, media systems, etc., and fitness competitions where the method credits or discredits the user's fitness results.
  • insurance e.g., an insured's fitness level factors into her premium
  • automatically switching users sharing a wearable fitness device e.g., accessing secure devices such as automobile ignition systems, door locks, media systems, etc.
  • fitness competitions e.g., a fitness competitions where the method credits or discredits the user's fitness results.
  • Identifying a user as described herein may be performed at various times and frequencies. In some cases, the identifying is performed at a time when she needs to enter into a transaction such as accessing a secure device. For example, the system may evaluate a current motion signature from a wearable fitness monitor to confirm that a user can execute a transaction (such as accessing a secure device, or receiving an insurance benefit) at the time when the user wants to execute the transaction. Such can occur when the activity tracker detects an interaction with a payment system, as may occur via an NFC protocol, a Bluetooth connection, or a triggering event initiated by an application executing on the activity tracker.
  • a user who wears a fitness monitor regularly over a period of days, weeks, months, years, etc. may have her identity checked periodically, with or without notifying the user.
  • the classification logic may check the identity of the wearer of the fitness monitor automatically, without prompting from the user. Changes in status may occur with or without notifying the user, e.g., user identification logic may determine that a fitness monitor that was identified as being worn by the user is no longer being worn by the user. Or the classification logic may determine that it can no longer conclude with a requisite level of confidence that the monitor is being worn by the user.
  • Such checks and adjustments may be made in the “background,” that is to say that they are performed without the user's input and/or without notifying the user.
  • evaluating motion signatures from a wearable fitness monitor results in deauthenticating a previously authenticated wearable fitness monitor and blocks execution of a transaction.
  • the process or classification logic starts by authenticating a user by using a technique other than a continuous or quasi-continuous fitness tracking measurement based on gait, heartbeat waveform, etc. Examples of the initial authentication techniques include fingerprint capture, ECG measurement, personal identification number (PIN) entry, and/or bioimpedance measurement. Such techniques often require only a short duration (e.g., a minute or less) to authenticate a user.
  • the wearable fitness monitor motion signatures are continually or periodically evaluated by the classification logic to determine whether to maintain authentication or deauthenticate the user/device.
  • Such evaluations may be conducted in the background by the classification logic during normal operation of the wearable fitness monitor.
  • the classification logic determines that the user is authenticated or deauthenticated, it may or may not notify user. In some embodiments, the user is not notified until she attempts to execute a transaction.
  • the user identification logic initially authenticates the user using the short duration technique (e.g., fingerprint, ECG, bioimpedance, etc.), and the logic then acquires the motion signatures produced by the voluntary and/or involuntary actions of the user wearing the fitness monitor, and uses these motion signatures to train a profile or other relationship between the motion signatures and a specific user.
  • the short duration technique e.g., fingerprint, ECG, bioimpedance, etc.
  • the device or logic may prompt the user to re-authenticate by verifying a code provided by a trusted device (e.g., a mobile phone in which authentication is achieved via fingerprint or PIN code) or any of the motion, heartbeat, or other fitness monitor-based authentication methods described in this disclosure.
  • a trusted device e.g., a mobile phone in which authentication is achieved via fingerprint or PIN code
  • Prompting may occur by a vibration or haptic interaction from the wearable fitness monitor. Prompting may occur at the next moment in which the wearable fitness monitor's user interface is engaged (e.g., upon pressing a button or providing a motion gesture such as moving a wrist wearable toward the face).
  • Wearable fitness monitors suitable for use with the methods and systems described herein collect data for the user of the device such as activity, sleep, and physiological measures. Examples include steps, distance traveled, calories burned, pace, floors climbed, elevation, number of active minutes, the start and stop of a sleep period, the duration of sleep, the number of awakenings during sleep, sleep disturbances from external stimuli, sleep stages, apnea hypopnea index, heart rate, resting heart rate, maximum heart rate, heart rate variability, time spent in at a specific exertion level (e.g., “cardio” zone), blood pressure, arterial stiffness, cardiovascular fitness, blood glucose, stress and/or relaxation levels, power output on a bicycle, number of swimming laps in a pool, number of swimming strokes, type of swimming strokes, lap splits, running mile splits, path walked or run overland (e.g., via GNSS tracking), location, time spent at work, home or the gym, number and/or length of sedentary periods in a day, start and stop of exercise
  • Wearable fitness monitors are devices that are worn on or carried by a person. They come in many form factors: wrist band, watch, clip, shoe pod, shoe, pendant, earbuds, clothing (shirt, socks, pants, undergarments), belt, cuff links, glasses, ring, earring (nose ring, studs, etc.), helmet, hat, hair clip, and socks.
  • wrist band watch, clip, shoe pod, shoe, pendant, earbuds
  • clothing shirt, socks, pants, undergarments
  • belt belt
  • cuff links glasses
  • helmet hat, hair clip, and socks.
  • hat hair clip
  • a set of protective, attachable and/or wearable cases that enable a user to wear a single wearable fitness monitor in multiple fashions or body locations may be provided.
  • a wearable fitness monitor may be designed such that it may be inserted into, and removed from, a plurality of compatible cases.
  • the wearable fitness monitors may be permanently or semi-permanently mounted into (or joined to) straps, clips, clasps, bands, or other attachments for wear.
  • the various individual elements of the various example cases and/or biometric tracking devices shown herein may also be combined with elements from other example cases and/or biometric tracking devices shown herein, e.g., a necklace or pendant case for a removable wearable fitness monitor may also be provided for a permanently-mounted wearable fitness monitor.
  • a wearable fitness monitor or biometric tracking device combined with a case or some other means allowing it to be worn or easily carried by a person may be referred to herein as a “biometric monitoring system” or “biometric tracking system.”
  • FIG. 1 depicts a generalized schematic of an example wearable fitness monitor or other device with which the various operations described herein may be executed.
  • the wearable fitness monitor 102 may include a processing unit 106 having one or more processors, a memory 108 , an operator interface 104 , one or more biometric sensors 110 , and input/output 112 .
  • the processing unit 106 , the memory 108 , the operator interface 104 , the one or more biometric sensors 110 , and the input/output 112 may be communicatively connected via communications path(s) 114 (it is to be understood that some of these components may also be connected with one another indirectly).
  • the wearable fitness monitor may collect one or more types of biometric data, e.g., data pertaining to physical characteristics of the human body (such as heartbeat, perspiration levels, etc.) and/or data relating to the physical interaction of that body with the environment (such as accelerometer readings, gyroscope readings, etc.), from the one or more biometric sensors 110 and/or external devices (such as an external heart rate monitor, e.g., a chest-strap heart rate monitor) and may then store such information for later use, e.g., for communication to another device via the I/O 112 , e.g., a smartphone or to a server over a wide-area network such as the Internet.
  • biometric data e.g., data pertaining to physical characteristics of the human body (such as heartbeat, perspiration levels, etc.) and/or data relating to the physical interaction of that body with the environment (such as accelerometer readings, gyroscope readings, etc.)
  • the processing unit 106 may also perform an analysis on the stored data and may initiate various actions depending on the analysis. For example, the processing unit 106 may determine that the data stored in the memory 108 indicates that a goal threshold has been reached and may then display content on a display of the portable biometric tracking device celebrating the achievement of the goal.
  • the display may be part of the operator interface 104 (as may be a button or other control, not pictured, that may be used to control a functional aspect of the wearable fitness monitor).
  • a wearable fitness monitor may incorporate one or more types of user interfaces including but not limited to visual, auditory, touch/vibration, or combinations thereof.
  • the wearable fitness monitor may, for example, display the state of one or more of the data types available and/or being tracked by the wearable fitness monitor through, for example, a graphical display or through the intensity and/or color of one or more LEDs.
  • the user interface may also be used to display data from other devices or internet sources.
  • the device may also provide haptic feedback through, for instance, the vibration of a motor or a change in texture or shape of the device.
  • the biometric sensors themselves may be used as part of the user interface, e.g., accelerometer sensors may be used to detect when a person taps the housing of the biometric monitoring unit with a finger or other object and may then interpret such data as a user input for the purposes of controlling the wearable fitness monitor.
  • accelerometer sensors may be used to detect when a person taps the housing of the biometric monitoring unit with a finger or other object and may then interpret such data as a user input for the purposes of controlling the wearable fitness monitor.
  • double-tapping the housing of the wearable fitness monitor may be recognized by the wearable fitness monitor as a user input that will cause the display of the wearable fitness monitor to turn on from an off state or that will cause the wearable fitness monitor to transition between different monitoring states, e.g., from a state where the wearable fitness monitor may interpret data according to rules established for an “active” person to a state where the wearable fitness monitor may interpret data according to rules established for a “sleeping” person.
  • the wearable fitness monitor 102 may calculate and store a user's step count while the user is wearing the wearable fitness monitor 102 and then subsequently transmit data representative of step count to the user's account on a web service like www.fitbit.com, to a mobile phone paired with the portable biometric monitoring unit, and/or to a standalone computer where the data may be stored, processed, and visualized by the user.
  • the device may measure, calculate, or use a plurality of other physiological metrics in addition to, or in place of, the user's step count.
  • caloric energy expenditure floors climbed or descended
  • heart rate heart rate variability
  • heart rate recovery location and/or heading (e.g., through GPS technology including a GPS receiver)
  • location and/or heading e.g., through GPS technology including a GPS receiver
  • elevation ambulatory speed and/or distance traveled, swimming lap count, bicycle distance and/or speed
  • blood pressure blood glucose, skin conduction, skin and/or body temperature
  • electromyography data electroencephalographic data
  • weight body fat, and respiration rate.
  • Some of this data may be provided to the wearable fitness monitor from an external source, e.g., the user may input their height, weight, and stride in a user profile on a fitness-tracking website and such information may then be communicated to the biometric tracking device and used to evaluate, in tandem with data measured by the biometric sensors 110 , the distance traveled or calories burned of the user.
  • the device may also measure or calculate metrics related to the environment around the user such as barometric pressure, weather conditions, light exposure, noise exposure, and magnetic field.
  • the communications interface may include wireless communication functionality so that when the wearable fitness monitor comes within range of a wireless base station or access point, the stored data automatically uploads to an Internet-viewable source such as a website, e.g., www.fitbit.com.
  • the wireless communications functionality may be provided using one or more communications technologies known in the art, e.g., Bluetooth, RFID, Near-Field Communications (NFC), Zigbee, Ant, optical data transmission, etc. Some of these communications technologies such as Bluetooth and NFC may be characterized as low power and/or short range in comparison to some other wireless communications technologies such as cellular and Wifi.
  • the wearable fitness monitor also contains wired communication capability, e.g., USB.
  • FIG. 1 illustrates a generalized implementation of a wearable fitness monitor 102 that may be used to implement a portable wearable fitness monitor or other device in which the various operations described herein may be executed. It is to be understood that in some implementations, the functionality represented in FIG. 1 may be provided in a distributed manner between, for example, an external sensor device and communication device, e.g., a chest-strap heart rate sensor that may communicate with a wearable fitness monitor.
  • an external sensor device and communication device e.g., a chest-strap heart rate sensor that may communicate with a wearable fitness monitor.
  • the memory 108 may also store configuration data or other information used during the execution of various programs or instruction sets or used to configure the wearable fitness monitor.
  • the processing unit may be implemented by a general or special purpose processor (or set of processing cores) and thus may execute sequences of programmed instructions to effectuate the various operations associated with sensor device syncing, as well as interaction with a user, system operator or other system components.
  • the processing unit may be an application-specific integrated circuit (ASIC) or programmable hardware, such as a FPGA.
  • ASIC application-specific integrated circuit
  • various other functional blocks may be provided as part of the wearable fitness monitor 102 according to other functions it may be required to perform, e.g., environmental sensing functionality, etc.
  • Other functional blocks may provide wireless telephony operations with respect to a smartphone and/or wireless network access to a mobile computing device, e.g., a smartphone, tablet computer, laptop computer, etc.
  • the functional blocks of the wearable fitness monitor 102 are depicted as being coupled by the communication path 114 which may include any number of shared or dedicated buses or signaling links. More generally, however, the functional blocks shown may be interconnected using a variety of different architectures and may be implemented using a variety of different underlying technologies and architectures.
  • the memory 108 may include non-volatile storage media such as fixed or removable magnetic, optical, or semiconductor-based media to store executable code and related data and/or volatile storage media such as static or dynamic RAM to store more transient information and other variable data.
  • non-volatile storage media such as fixed or removable magnetic, optical, or semiconductor-based media to store executable code and related data
  • volatile storage media such as static or dynamic RAM to store more transient information and other variable data.
  • the various methods and techniques disclosed herein may be implemented through execution of one or more sequences of instructions, e.g., software programs, by the processing unit 106 (e.g., a generalized or specialized processor) or by a custom-built hardware ASIC (application-specific integrated circuit) or a programmable hardware device such as an FPGA (field-programmable gate array), or any combination thereof within or external to the processing unit 106 .
  • the processing unit 106 e.g., a generalized or specialized processor
  • ASIC application-specific integrated circuit
  • FPGA field-programmable gate array
  • the wearable fitness monitor may include computer-executable instructions for controlling one or more processors of the wearable fitness monitor to obtain biometric data from one or more biometric sensors.
  • the instructions may also control the one or more processors to receive a request, e.g., an input from a button or touch interface on the wearable fitness monitor, a particular pattern of biometric sensor data (e.g., a double-tap reading), etc., to display an aspect of the obtained biometric data on a display of the wearable fitness monitor.
  • the aspect may be a numerical quantity, a graphic, or simply an indicator (a goal progress indicator, for example).
  • the display may be an illuminable display so as to be visible when displaying data but otherwise invisible to a casual observer.
  • the instructions may also cause the one or more processors to cause the display to turn on from an off state in order to display the aspect of the biometric data.
  • Motion sensors provide an output a signal responsive to motion experienced.
  • Examples of motion sensors include accelerometers, gyroscopes, compasses, switches (for example, mechanical), GPS modules, piezoelectric film and/or pedometers to determine, calculate and/or detect one or more steps of the user; notably, the exemplary motion sensor may be incorporated into portable monitoring devices such as wearable fitness monitors.
  • the portable monitoring device may estimate, calculate and/or determine, calorie consumption, burn and/or expenditure using data which is representative of the intensity of user motion—for example, as provided or determined by one or more single axis or multi-axis accelerometers.
  • the signals from the one or more accelerometers may be filtered using time domain or frequency domain filtering techniques to produce a parameter indicative of the intensity of user motion, often referred to as a “count”.
  • a count may be computed as the sum of the rectified filtered accelerometer output taken over a suitable time epoch, for example, 10 seconds, with or without additional processing such as thresholding and/or saturation.
  • the portable monitoring device may calculate, determine and/or estimate calorie consumption, burn and/or expenditure as a function of the current count value or a sequence of count values. Further descriptions of various motion sensors are provided in U.S. Patent Application Publication No. 2015/0134268 (Attorney Reference: FTBTP014D15C2US), titled PORTABLE MONITORING DEVICES AND METHODS OF OPERATING SAME, filed on Jan. 22, 2015, which is incorporated by reference in its entirety.
  • An accelerometer is often used as a motion sensor.
  • this disclosure adopts a coordinate system as outlined in FIG. 2 .
  • coordinate systems are a matter of convenience and may be arbitrarily defined—for example, a tri-axial accelerometer may be flipped upside down to reverse the orientations of two of the three axes of the tri-axial accelerometer or may be subjected to two 90° rotations about mutually-perpendicular axes to cause all three axes to be aligned differently from the conventions used herein.
  • the techniques discussed herein may be practiced using tri-axial accelerometers (and their corresponding measurement outputs) that are aligned with coordinate systems different from the convention used in this disclosure (as outlined in FIG. 2 ). It is to be further understood that the data from tri-axial accelerometers that are aligned with other coordinate systems may still be used to perform the techniques discussed herein if the data from such tri-axial accelerometers is transformed in order to align with the coordinate system convention adopted herein or if the techniques outlined herein are adapted, e.g., transformed, to account for the shift in coordinate systems (for example, if an axis is reversed from the convention used herein, a condition stating that an acceleration along that axis be less than ⁇ 0.125 g may have an equivalent condition in the new coordinate system that the acceleration along that axis be more than 0.125 g).
  • the accelerations obtained from the accelerometer(s) may be first subjected to one or more pre-processing steps prior to being used in the present techniques.
  • the accelerations may be used in raw form (counts or accelerations converted from counts, for example) or may first be smoothed or otherwise processed (e.g., by using a moving average filter) to reduce noise and produce a more stable signal.
  • the techniques of concepts presented herein are intended to provide more reliable, more responsive recognition of motion signatures while simultaneously having a low impact on battery life.
  • various operations that are discussed may be performed slightly differently in actual practice. For example, as part of one technique, the magnitude of acceleration measured by a tri-axial accelerometer may be evaluated to see if it exceeds a threshold acceleration.
  • Another type of motion sensor is an angular motion measurement system.
  • a detailed description of angular motion measurement systems is provided in U.S. Provisional Patent Application No. 62/054,341 (Attorney Reference: FTBTP002X1KP), titled HIGH-DYNAMIC RANGE ANGULAR MOTION SENSING SYSTEM, filed on Sep. 23, 2014, which is incorporated by reference in its entirety.
  • Such systems may obtain angular motion measurement data using a hybrid system incorporating two different, non-gyroscopic angular motion measurement sensors.
  • Such a system includes, at a minimum, a multi-accelerometer angular rate sensor (MAARS), an accelerometer/magnetometer angular rate sensor (AMARS), and logic for determining which of the two angular rate sensors (ARS's) were to actively used to collect data at any given instant in time.
  • MAARS multi-accelerometer angular rate sensor
  • AMARS accelerometer/magnetometer angular rate sensor
  • ARS's angular rate sensors
  • wearable fitness monitors are used to control or facilitate control of electronic and/or digital devices and systems such as household appliances, automobiles, door locks, and the like. Such devices may be secure in the sense that they cannot be controlled or otherwise accessed without a form of authentication (e.g., a password) or other user identification.
  • the fitness monitor serves to identify a user and allow access a secure electronic device or system. This form of identification may have other applications such as enabling access to secure physical areas or property, and customizing user experiences/interfaces for a service.
  • Further examples include controlling a television, unlocking and/or opening the door of a residence, office, car, or other locked space, providing a digital ID and password (pin or otherwise) to access a computer, banking account, online shopping site, or other computer account, changing the music played in a room for a specific user, selecting the goods to display to a user on a shopping site, etc.
  • a user who increases daily activity by 10% over a baseline daily activity value for 6 months may obtain a reduced insurance premium for the following 6 months.
  • the wearer of the fitness monitor must be identified as the user before such rewards are distributed to the user. This identification can occur at different points in time as the user is wearing the device and making progress towards the goal. The points in time may be selected according to a set frequency, period, schedule (with different intervals between identification points to appear random) or triggered based on a condition that considers progress toward a goal (e.g., if the wearable fitness monitor detects activity of a given step rate, time period of activity, or an activity being performed).
  • the data from wearable fitness monitor may also be used to compete against friends (e.g., in a social network), coworkers, or in a game. For example, a user who runs a specific path the fastest may be granted a title (e.g., “king of the hill”). A user who has the most steps in a week (or any other time period) may be placed at the top of a leaderboard. A user in a digital race game may be ranked against peers on the basis of running distance in a month. In each of these and other cases, the wearer of the fitness monitor must be identified as the user before the fitness monitor's measured metric is ascribed to the competing user.
  • the competition may be individually based where the metrics of an individual are ranked or compared against other individuals.
  • the competitions may be group based, where groups are formed from multiple individuals and the metrics of the members of the group are aggregated (e.g., summed, averaged, etc.) and the aggregated metric of a group is compared against the aggregated metric of the other groups.
  • Embodiments may allow groups to be formed based on an attribute such as a department, company, geography (state, city), a school, a dorm, a social group, a family connection, friend connection, favorite team, or any other suitable attribute.
  • embodiments may allow groups to be formed based on an invitation scheme.
  • a wearable fitness monitor may be shared amongst multiple users. For example, a gym, household, school, work place, or any other community may provide a wearable fitness monitor to members of that community. The use of the wearable fitness monitor may be time-sliced among the member. Rather than requiring each user initiating a session with the wearable fitness monitor, the classification logic may authenticate the wearer of the classification logic and, once authenticate, correlate tracked activity data from the wearer's use of the wearable fitness monitor with the digital account of the wearer.
  • the wearer of a fitness monitor is identified as the user by obtaining one or more motion signatures from the fitness monitor and determining whether such signature(s) are produced by the user's movement. Individual humans have movement characteristics which individually or collectively be used as a biometric identifier for the user. Such motion signatures are described further herein. These signatures and their comparison to reference features linked to the user may allow the wearable fitness monitor to authenticate or otherwise identify the wearer of the wearable fitness monitor.
  • a motion signature may be obtained from a motion sensor such as one of the sensor types mentioned above. It may be viewed as a form of biometric information that can be collected or presented at any time to authenticate or otherwise identify the wearer of the wearable fitness monitor.
  • the wearer of a fitness monitor is identified as the user when the motion signature (or the motion signature in combination with other wearer information) matches, to a degree required by an appropriate user classification procedure, the information in one or more reference features.
  • motion signatures may be obtained via data from motion sensors and may be used by the user identification logic. Some motion signatures may represent motion in the time domain (e.g., amplitude, power, intensity, phase, field strength, etc., each as a function of time). Others represent motion in the frequency domain (e.g., amplitude, power, intensity, phase, field strength, etc., each as a function of frequency). Other types of motion signatures are provided only in association with certain activity types such as running, walking, swimming, bicycling, rowing, weightlifting, etc. For example, curl repetition count is a motion signature that is associated with weightlifting but not running or bicycling.
  • Some other types of motion signature employ a single cycle which may be characterized by its “profile,” which may have either time or frequency as an independent variable.
  • the cycle motion signature is compared against a reference cycle feature, with the comparison matching features of the cycle such as total magnitude and/or duration (e.g., peak-to-peak), maximum or minimum magnitude, maximum positive or negative slope, the number and relative locations of inflection points, envelope, etc.
  • the comparison can be performed by many different techniques including pattern recognition algorithms or classification algorithms known to those of skill in the art.
  • motion signatures include step count, step rate, cadence, a metric derived from an amplitude of a periodic motion, a biking cadence, a rowing rate, a resistance-based repetition rate, a typing speed, a zero crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof.
  • motion signatures are obtained from data taken from motion sensors.
  • the data may be processed little or substantially to obtain motion signatures.
  • the motion sensor data used to obtain the motion signatures may be the raw “absolute” signal, or may be obtained after filtering (by, e.g., bandpass filtering, low pass filtering), scaling, and the like.
  • the 2-norm of the 3-axis accelerometer motion signal may be used in lieu of or in combination with the (x, y, z) signal features to provide information used in a motion signature.
  • Motion signatures representing repetitive movement may be obtained using, e.g., data processed to obtain peak counts, zero crossings, spectral information from, e.g., a FFT, and the like.
  • a Fourier series decomposition may be performed to extract the contributions of multiple periodic motions to sensor signal. These contributions may be harmonics associated with, e.g., steps and arm motion.
  • Each harmonic may be a motion signature and/or the ratio of the harmonics' powers may be a motion signature. For example, it has been found that a user's step impact has a big effect on the power observed in higher harmonics. It is to be appreciated that although some embodiments have been described in the context of harmonics, other embodiments can operate on any contribution of a spectral component.
  • the wearing configuration of a fitness monitor affects the resulting motion signature.
  • a single user motion may produce one motion signature when obtained using a fitness monitor clipped to the user's hip and a second motion signature, different from the first motion signature, when obtained using a fitness monitor worn on the user's wrist.
  • the motion signature analysis may therefore account for the type and worn location of the fitness monitor.
  • FIG. 3 shows a representative 3-axis accelerometer signal from a wearable fitness monitor worn on the wrist of a user who is walking.
  • motion signatures such as the step rate (e.g., steps/min) and signal energy may typify a user. For example, for a user who is walking, the accelerometer signal energy increases with step rate.
  • a user may be identified, distinguished from another user, or determined to be a fake by comparing the pair of motion signatures (step/min, signal energy) from the motion signal provided by the wearable fitness monitor to data characterizing the user's walk (e.g., features from signals previously supplied by the user or an entity responsible for enrolling the user and characterizing the user's walk).
  • trusted information characterizing movement typical of the user is referred to as a reference motion feature for the user.
  • FIG. 4 depicts the step rate and signal energy for 4 unique users. The curve may be approximated as a line and a fitness monitor or the wearer of the fitness monitor may be classified by the nearest line to the data from the wearable fitness monitor (i.e., to one or more motion signatures).
  • the data may be considered fake (e.g., in the sense of a linear discriminant).
  • a line is used in this example, but any model may be employed such as an arbitrary polynomial, lookup table, etc.
  • the classifier employed may be a neural network, support vector machine, random forest, decision tree, or other machine learning or heuristic algorithm.
  • the step rate and signal energy during a run may identify the user, distinguish the user from another, or determined to be a fake. In embodiments where there is no enrollment for a user, default curves or lookup tables for these quantities may be employed.
  • FIG. 5 depicts the step rate and signal energy for 3 unique users for a run.
  • one user is distinguished from another by the spectral characteristics of the motion data.
  • the spectral characteristic is a harmonic
  • the ratios of the second to first harmonic, third to first harmonic, and third to second harmonic observed in the 2-norm of a 3-axis accelerometer may be used.
  • Higher harmonic content corresponds to more impact in walking.
  • the approach can be used for running.
  • FIG. 6 depicts two unique users who are otherwise indistinguishable in running by step rate and signal energy, but are clearly distinguished by motion harmonics. Harmonic features may also be used in combination with the aforementioned walking/running features.
  • the accelerometer motion signal is split into “cycles” (e.g., periods between two steps), aligned, time warped, and used to construct a composite or typical step profile for a user by which features such as the (x, y, z) axis peak-to-peak heights, envelopes, and peak-to-peak duration may be used to build a model of the user's typical motion.
  • a “cycle-type” motion signature and associated reference motion feature may be used to identify a user in the manner described elsewhere herein.
  • a machine learning algorithm such as an LDA classifier, an artificial neural network, decision tree, support vector machine, or the like may be used to classify the user.
  • the wearable fitness monitor may have a gyroscope and similar or identical approaches may be employed in lieu of or in combination with an accelerometer.
  • signal data and signal processing operations performed they may be performed on the raw “absolute” signal, or after filtering (e.g., bandpass filter, low pass filter), scaling, and the like.
  • the wearable fitness monitor may store one or more invalid motion features that each individually or collectively characterize motion likely to be faked (e.g., performed by a non-human). To detect fakes, the user's identity may be rejected based on an invalid motion feature that characterizes a detected step rate being too consistent (e.g., not varying by more than 5 steps/min from minute to minute over a 10 minute time window) or the motion being too periodic (e.g., each “cycle” of motion corresponding to a step being nearly identical to the previous), the signal energy being contained to nearly one axis of motion (e.g., through principal component analysis with over a threshold value (e.g., 50%) of the 2-norm of a 3-axis accelerometer signal being comprised of one motion axis), the duration of motion being too long (e.g., over 1 hour with no breaks), the motion observed on an accelerometer lacking high harmonic structure (e.g., being too smooth or sinusoidal, no clear presence of integer harmonics from the fundamental step frequency), or too
  • a feature may thus be viewed broadly as data or logic that defines an expected signature of a user, either in the positive (e.g., reference motion signatures) or in the negative (e.g., invalid motion features).
  • Such features may be expressed in a data driven manner (such as through a line, curve, graph, data point), functionally (e.g., logic that defines an expected threshold around a given motion signature), or some combination thereof.
  • a functional expression of a feature is an invalid feature that specifies that motion signatures that do not exceed a minimal variance are to result in deauthenticating the wearer, as may be the case where a mechanical device may be causing the motion data detected by the wearable fitness monitor.
  • the classification logic may analyze the motion signature to determine whether the cycles represented by the motion signature vary beyond a threshold.
  • the wearable fitness monitor comprises a location sensor (e.g., GPS, WiFi interface) or is in communication with a device that has a location sensor (e.g., a smartphone)
  • location may be used to determine that the wearable is not with the intended user. For example, if the activity data provided by the wearable fitness monitor does not spend significant periods of time at the user's registered home and/or workplace (e.g., at least 8 hours at home), then the activity data may be classified as fake in relation to a cash incentive program or competitive challenge.
  • the wearable fitness monitor may be linked to the user's smartphone (e.g., via an account on an app running on the phone) but is not in proximity of the phone or does not “sync” data with the user's phone for a period of time (e.g., 1 week), then the activity data provided by the wearable fitness monitor may be rejected as fake for the purposes of a cash incentive program or competitive challenge.
  • the wearable fitness monitor comprises a wireless communication system such as Bluetooth, Bluetooth Low Energy, Near Field Communication, Wifi, Body Area Network (e.g., communication routed through the body), and the like.
  • a wireless communication system such as Bluetooth, Bluetooth Low Energy, Near Field Communication, Wifi, Body Area Network (e.g., communication routed through the body), and the like.
  • the device or a system incorporating the device e.g., a mobile phone and the wearable fitness monitor or a cloud-based server and the wearable fitness monitor
  • the names of enrolled or commonly observed Wifi networks in the user's typical areas of movement may be used to determine if the wearable fitness monitor is in a foreign environment, which may trigger a de-authentication.
  • the device may prompt the user to re-authenticate by verifying a code provided by a trusted device (e.g., a mobile phone in which authentication is achieved via fingerprint or PIN code) or any of the authentication methods described in this disclosure. Prompting may occur by a vibration or haptic interaction from the wearable fitness monitor.
  • a trusted device e.g., a mobile phone in which authentication is achieved via fingerprint or PIN code
  • Prompting may occur at the next moment in which the wearable fitness monitor's user interface is engaged (e.g., upon pressing a button or providing a motion gesture such as moving a wrist wearable toward the face).
  • the system infer the geographical location (e.g., address, latitude and longitude) of the user based on the communication data—it is sufficient to maintain a list of networks and/or devices that are in communication with the user.
  • the wearable fitness monitor includes a Body Area Network communication system.
  • the device may transmit data through the user's body to another device that is in close proximity or in contact with the user.
  • the user may touch a door knob that likewise includes a Body Area Network communication system and the door knob may unlock in response to the user's touch.
  • the user may touch or come in close proximity (e.g., less than 1 cm) to an automobile door handle, secured door in an office, etc., and the lock may disengage and/or open in response.
  • the same touch and/or proximity may engage “syncing” of the wearable fitness monitor in the sense of transmitting activity, sleep, and other biometric data to the device that is in proximity and, thereafter, a cloud-based service (e.g., www.fitbit.com).
  • a cloud-based service e.g., www.fitbit.com.
  • a user may have a Body Area Network enabled lamp at home and touching the lamp may set the color and/or intensity of the light based on the user's preferences (including time of day) and also initiate the transmission of the wearable fitness monitor's activity data to the lamp (if so enabled to receive this data) or other communication hub (e.g., a computer in the user's residence).
  • a user may authenticate to a wearable fitness monitor through features of a heart rate sensor, such as an electrocardiogram (ECG).
  • ECG electrocardiogram
  • the device may contain a two electrode system where one electrode (lead 1) is in contact with the wrist and an electrode on the outer surface of the bracelet (lead 2) that is accessible by the opposite hand.
  • one electrode lead 1
  • an electrode on the outer surface of the bracelet lead 2
  • an ECG signal can be collected for multiple heart beat cycles or for some duration of time (for example 10 heart beats or 10 seconds).
  • the ECG waveform collected is then post-processed into a signature waveform. This may involve techniques such as overlapping the periodic components of the waveform (e.g., individual PQRST sections), alignment, normalization, outlier removal, and linear/nonlinear filtering, outputting a composite or typical PQRST waveform signature.
  • FIG. 7 shows examples of data obtained from post-processing.
  • Panel (a) of FIG. 7 shows the raw ECG waveform acquired from the wearable fitness monitor.
  • Panel (b) of FIG. 7 shows the ECG waveform after filtering.
  • Panel (c) of FIG. 7 shows overlapping multiple repeating PQRST waveforms.
  • Panel (d) of FIG. 7 shows the final ECG signature waveform used for feature extraction.
  • Signatures are then extracted from the final ECG signature waveform.
  • These signatures may include pathological characteristics of the PQRST wave (shown in FIG. 8 below) use in cardiac monitoring such as the time and magnitude features of the PR interval, PR segment, QRS complex, ST segment, and QT interval.
  • Signatures may also include other time domain (shown in FIG. 9 below) such as the slopes, areas, curvatures, and polynomial fits that may not directly have any direct physical or medical significance, but present uniqueness useful for authentication.
  • Frequency domain characteristics can also be used as features such as Fourier and cosine transforms of the PQRST.
  • Authentication is determined through comparing the features of a reference ECG vs. that of a the ECG signal detected from the current wearer.
  • Techniques for authentications could involve a combination of techniques such as neural network, support vector machine, logistic regression, naive Bayes, random forest, decision tree, or other machine learning or heuristic algorithms. After authentication is successful, the wearer can remain authenticated until the bracelet is determined to be off wrist, such as due to device removal or loss of contact.
  • the device may provide continuous ECG measurements of the user. Similar to the embodiments described above, the ECG may be split into “beats”, aligned, and used to construct a composite PQRST waveform over a moving window, which is then used to extract features to compare against a template for the user.
  • authentication may be continuous and the user may have a trust score that degrades if the features do not match for a period of time longer than the moving window. If the trust score goes below a designated threshold for a designated period of time, the user may be de-authenticated from the device.
  • the presence of a live user may be determined both by a continuous (or nearly continuous) heart rate signal and that the user is wearing the shirt continuously.
  • Using clothing to house a wearable fitness monitor is described in US Patent Application Publication No. 2010/0292599, which is incorporated herein by reference in its entirety.
  • the wearable fitness monitor may include another type of heart rate sensor, such as a photoplethysmograph (PPG) sensor, for instance, to monitor the heart rate of the user when it is worn against the wrist.
  • PPG photoplethysmograph
  • the shape of the PPG waveform (e.g., in the sense of a template), which is correlated to the age of the user, may also be used as a feature after appropriate temporal and spatial normalization.
  • a PPG sensor may be used to characterize the level of skin pigmentation for the wearer of the device based on a response of the PPG sensor. This may be used as an additional feature in a user authentication system. For example, if the intensity level of the PPG's light emitter is high but the return signal to the PPG's light detector is low, then the user has higher skin pigmentation. The ratio of the return signal to the emitter output may be used as a feature to characterize a user. Similarly, testing with different light intensities and/or wavelengths may provide a transfer function (or table lookup) for the user that may be used to identify the user.
  • the wearable PPG sensor may have multiple wavelength LEDs that can be set to different intensities and/or multiple wavelength (e.g., spectral response) photodiodes to do this characterization.
  • the PPG sensor may be used to determine if the wearable fitness monitor is being worn. That is, if the PPG response is not representative of human skin (e.g., the return signal is low relative to the emitted output because there is nothing against the sensor), then the device can determine that it is not being worn, at least not on the wrist. This may be determined in combination with a motion sensor (e.g., that the device is stationary on a surface). Similarly, the absence of a heart rate signal in the PPG data, or lack of heart rate variability (e.g., the duration between heart beats is too consistent), may be used to determine that activity data from the device is faked. Moderate to vigorous activity (e.g., walking or running) in the presence of low heart rate (e.g., below 60 bpm for walking and 90 bpm for running) in the PPG data may likewise indicate a fake.
  • Moderate to vigorous activity e.g., walking or running
  • low heart rate e.g., below 60 bpm for walking and 90
  • the wearable fitness monitor contains a motion sensor and an ECG.
  • the user authenticates to the device with the ECG and the authentication may be lost if the motion signature is not representative of the registered user.
  • the wearable fitness monitor comprises a contact sensor (e.g., in the clasp, a PPG, or capacitive sensor against the wrist of the user) that detects when the sensor is removed from the user's wrist and then authentication is lost.
  • the wearable fitness monitor contains a motion sensor and a PPG.
  • Activity data from the device may be rejected as fake if the motion signature or the PPG-derived data do not correspond to the registered user or to human activity.
  • the wearable fitness monitor contains an ECG and PPG.
  • the user authenticates not only by matching the ECG morphological signatures and PPG morphological signatures with those previously enrolled by the user (e.g., the reference features), but also in comparing the heart rate and heart rate variability of the two signals to each other.
  • the wearer of the device may be determined to not be the authorized user of the device based on the heart rate exertion of the user observed during an exercise. For example, the user may walk at a moderate pace and if the heart rate divided by the pace (as observed, say, by GPS) or step cadence is significantly higher or lower than is characteristic for the authorized user, the current user of the device is determined to not be the authorized user.
  • the wearable fitness monitor automatically tracks exercises such as elliptical, bicycling, and the like
  • equivalent metrics of heart rate per unit activity e.g., elliptical strokes, bicycling pace
  • optical monitors are used in the wearable monitor, implementing different modes of operation by emitting pulses of light and detecting light after it interacts with the user's skin or other tissue, to thereby capture data that may be used to obtain the user's heartbeat waveform, worn state, user characteristics, etc.
  • the optical monitor is used as a heartbeat waveform monitor, and while much of the following description refers to such monitors as heartbeat waveform monitors, such monitors need not be configured or designed to measure heartbeat waveforms. It is sufficient that the monitor emit and detect pulses and interpret the pulsing information to accomplish the described results.
  • the current disclosure provides methods for operating a wearable fitness monitoring device having a heart rate monitor (HRM) in a low power state when the device determines that the device is not worn by a user, or is “off-wrist” when implemented in a wrist-worn device.
  • HRM heart rate monitor
  • This feature of the HRM is also referred to as an “automatic off” function.
  • the automatic off function is implemented by operating the HRM in an “unworn” (or “off-wrist”) detection mode, and the automatic off function automatically turns off the heart rate monitoring operations of the HRM to conserve energy if the device determines that it is not being worn by the user.
  • Other benefits of the automatic off function include providing more accurate heart rate estimation.
  • a heart rate detection algorithm may reset.
  • the algorithm stops running when off-wrist is detected, and restarts when on-wrist is detected.
  • the heart rate monitor restarts, it resets.
  • the current disclosure provides methods for operating a wearable fitness monitoring device having a heart rate monitor in a normal power state when the device is worn by the user, or “on-wrist” when implemented in a wrist-worn device.
  • This feature of the HRM is also referred to as an “automatic on” function.
  • the automatic on function is implemented by operating the HRM in a “worn” (or “on-wrist”) detection mode. The automatic on function automatically takes the HRM out of a low power state and turns on the heart rate monitoring operations of the HRM if the device detects motion and determines that it is worn by the user.
  • the unworn (or off-wrist) and worn (or on-wrist) detection may be implemented by light (e.g., LED) probing, which emits light pulses and detects signals after the light pulses interact with the user's skin and tissues.
  • the unworn and worn probing may share some hardware, firmware, software, and/or parameters for light emission, light detection, and analyses of detected signals.
  • the two probing modes employ different hardware, firmware, software, and/or parameters for light emission, light detection, and analyses may be used for unworn and worn detection.
  • the wearable fitness monitoring device goes in and out of the low power state regulated by a probe light (e.g., LED) and a motion detector, implementing automatic off and on functions.
  • a probe light e.g., LED
  • a motion detector e.g., a Bosch Sensortec BMA150 accelerometer
  • the heart rate monitor saves power by turning off, or scaling back operation of, its LED light source and its photodetector.
  • other light sources and light detectors e.g., photodiodes, photomultiplier tubes, CCD, or CMOS may be used to implement the automatic off and on functions.
  • Some embodiments provide a method of operating a heart rate monitor of a wearable fitness monitoring device having a plurality of sensors.
  • the method includes: (a) operating the heart rate monitor in a first mode while also operating in a second mode configured to detect near proximity of the wearable fitness monitoring device to a user's skin, where the first mode is configured to determine one or more characteristics of a user's heartbeat waveform when the wearable fitness monitoring device is in near proximity to the user; (b) from information collected in the second mode, determining that the heart rate monitor is not proximate to the user's skin; and (c) in response to determining that the heart rate monitor is not proximate to the user's skin, ending operating the heart rate monitor in the first mode.
  • the one or more characteristics of the user's heartbeat waveform include the user's heart rate.
  • the wearable fitness monitor includes a motion sensor, and the method further involving: prior to (c), determining from information output by the motion detecting sensor that the wearable fitness monitoring device has had been still for at least a defined period; and in response to detecting that the wearable fitness monitoring device has had been still for at least the defined period, performing (c).
  • the device prior to (a) while the first mode is not operating, the device (i) detects motion of the wearable fitness monitoring device using a motion detecting sensor and/or detecting proximity of the heart rate monitor to the user′ skin by operating the heart rate monitor in a third mode; and (ii) initiates operation of the first mode of the heart rate monitor when the wearable fitness monitoring device is determined to be in near proximity to the user.
  • the wearable fitness monitor has a bioimpedance sensor (possibly sharing the same electrodes as the ECG) and the bioimpedance of the user is further used with the ECG and/or PPG to determine the current user of the wearable fitness monitor.
  • the wearable fitness monitor has a fingerprint sensor (e.g., capacitive, ultrasound) that images the pattern of skin ridges on the finger(s) of a user to authenticate the user to a device (e.g., when the device is put on, when the device is used as a proxy for a credit card).
  • the device may include an ECG, PPG, and/or bioimpedance sensor to further enhance the authentication system with user-specific biometric data.
  • the device may maintain authentication through motion signatures of the user and PPG-based signatures of the user.
  • Removal of the device may de-authenticate the wearer of the device.
  • the authorized user of the wearable fitness monitor may also have a smartphone that tracks user activity such as walking. If the smartphone walking activity does not match the time and approximate count of the wearable fitness monitor walking activity data for more than a threshold percentage (e.g., 50% of time over some period), then the data of the wearable fitness monitor is considered fake. If the wearable fitness monitor is used as a proxy for a credit card or other secure service, the authorization of the user is deactivated. Reauthentication may be established by displaying a code or image on the wearable fitness monitor and entering the same on the user's mobile phone.
  • a threshold percentage e.g. 50% of time over some period
  • the mobile phone is used as a proxy for a credit card or other secure service, the authorization of it may be deactivated until reauthentication by matching a code or image to the wearable is performed.
  • the wearable fitness monitor comprises a location sensor (e.g., GPS), it may activate the sensor and broadcast its location when it next synchronizes to a cloud-based service.
  • the wearable fitness monitor tracks the sleep of the user based on sensor data generated by the wearable fitness monitor.
  • the wearable may use motion data and/or heart rate data to infer when a user is asleep.
  • Other examples include the use of skin temperature, galvanic skin response, SpO2, blood pressure, and time of day in combination with or in lieu of the preceding data to infer when the user is asleep.
  • the authentication may be “blacked out” so that the wearable fitness monitor cannot be used to access secure digital accounts, open secure areas, etc. Upon waking, the user may retain authentication.
  • Skin color may be used to define one or more features of a user.
  • skin color may affect heart rate measurements. This section discloses techniques for measuring skin color as a user feature and using skin measurements to improve measurements of other features such as heart rate.
  • Ambient light and skin color may make it difficult to extract a user's heart rate from a PPG signal.
  • the effect of ambient light may be reduced by subtracting a value of the received detected light signal when the PPG light source is off from the value of the received detected light signal when the PPG light source is on (assuming that both signals are obtained in close temporal proximity to each other).
  • the effect of skin color may be reduced by changing the intensity of the PPG light source, the wavelength of the light emitted from the light source, and/or by using the ratio or difference of received signal corresponding to two different wavelengths.
  • Skin color may be determined by using user input (e.g.
  • the effect of skin color (and tightness with which the user is wearing the device) on the raw PPG signal may also be measured by sending in a signal of known amplitude to the light source(s) and then measuring the received signal from the photodetector(s). Such a signal may be sent for a prolonged period of time (so as to capture data through multiple expected heart beats) and then averaged to produce a steady-state data set that is not heart-rate dependent. This amplitude may then be compared to a set of values stored in a table to determine algorithm calibration, transmitter amplitude and the receiver gain.
  • the disclosure provides methods and devices to accurately measure heartbeat waveform for different user characteristics, such as skin colors, motion, sweat, position, and physiologic state (e.g., skin thickness, body fat, etc.) of the users. Because darker skin has lower reflectance of light, the relations between photodetector reading and light pulse intensity, e.g., DAC, tends to have a lower slope than for paler skin.
  • the signals for skin characterization may operate intermittently at higher frequency than the light pulses of the first mode for heart rate monitoring.
  • Some embodiments provide a method for adjusting at least one setting for operating a heart rate monitor in a wearable fitness monitoring device.
  • the method involves: (a) pulsing a light source in the heart monitor in a skin characterization mode by emitting a succession of light pulses, at least some having variable intensity with respect to one another; (b) detecting a variation in intensity of light from the light pulses emitted in the skin characterization mode after the light has interacted with the user's skin; (c) determining a response characteristic of the user's skin from the variation in intensity of light detected in (b); and (d) using the response characteristic of the user's skin to adjust a gain and/or light emission intensity of the heart rate monitor operating in a first mode for detecting one or more characteristics of the user's heartbeat waveform.
  • the response characteristic is dependent on an opacity value of the user's skin.
  • operating in the first mode and operating in the skin characterization mode are performed concurrently.
  • operating in the first mode and operating in the skin characterization mode concurrently involves periodically determining a response characteristic of the user's skin while continuously operating in the first mode.
  • operating in the first mode involves pulsing the light source in the heart rate monitor at a first frequency and detecting light from the light source, after the light has interacted with the user's skin, at the first frequency.
  • operating in the skin characterization mode involves pulsing the light source in the heart rate monitor at a second frequency and detecting light from the light source at the second frequency.
  • the wearable fitness monitor may store a reference feature relating to the response of a sensor to the skin of a user to later verify the user based on a current response of the sensor to the skin of the wearer of the wearable fitness monitor using, for example, the aforementioned techniques.
  • Various classification and identification techniques may be applied to compare motion and/or other signatures to user reference features. Generally, such techniques determine whether or not it is likely that the motion signature obtained from a fitness monitor was created by a user in question wearing the device. In this way, a wearer of the fitness monitor can be authenticated or otherwise identified.
  • the logic used to compare a signature to a reference feature may be a classifier or other routine implemented on the fitness monitor and/or a secondary device as described elsewhere herein.
  • the classifier employed may be an LDA classifier, neural network, support vector machine, random forest, decision tree, or other machine learning or heuristic algorithm.
  • the motion signal may be split into “cycles” (e.g., periods between two steps), aligned, time warped, and used to construct a composite or typical step profile for a user by which features such as the (x, y, z) axis peak-to-peak heights, envelopes, and peak-to-peak duration may be used to build a model of the user's typical motion.
  • a machine learning algorithm such as a neural network, decision tree, support vector machine, and the like to classify the user.
  • the classification logic compares signatures to reference features and applies a level of confidence (provided directly or indirectly by the comparison algorithm) for authenticating or otherwise identifying the user.
  • the confidence level for identifying the user may be set as appropriate for the application (insurance versus fitness credits in casual competition).
  • the level may also be set for the type of classification algorithm used to compare the wearer's signature(s) to the user's reference feature.
  • FIG. 11 shows a flowchart of a method for determining whether an identity of an instant wearer of a fitness monitor matches that of a user. If an instant wearer's identity matches that of the user, the instant wearer is authenticated as the user. For example, the user may be an owner or an authorized user of the wearable fitness monitor. Because the motion features of the user are provided as a reference against which a wearer's data are compared, the user is also referred to as the reference user herein.
  • the operations of process 1100 are performed by a single wearable fitness monitor. In other implementations, some of the operations can be performed by the wearable fitness monitor, while others operations can be performed by an external device associated with the wearable fitness monitor such as smart phone, a personal computer, a tablet, or a webserver that is associated or communicatively linked to the wearable fitness monitor.
  • an external device associated with the wearable fitness monitor such as smart phone, a personal computer, a tablet, or a webserver that is associated or communicatively linked to the wearable fitness monitor.
  • Process 1100 involves obtaining a reference motion feature of the reference user (or user as used elsewhere herein) using one or more motion sensors on the wearable fitness monitor worn by the user. See block 1102 .
  • the motion sensors may be selected from accelerometers, gyroscopes, GPS sensors, and other motion sensors described herein elsewhere.
  • the reference motion feature comprises a motion cycle profile as described above and hereinafter at block 1206 of FIG. 12 .
  • process 1100 involves training a classifier using the reference motion feature.
  • the classifier is a binary classifier.
  • the classifier is a linear discriminant (LDA) classifier.
  • LDA linear discriminant
  • an LDA classifier is provided as an example below, in various implementations, other classifiers described herein or known in the field may be used instead of or in combination with an LDA. For instance, clustering methods, neural networks, support vector machines, linear and nonlinear models, decision trees, etc., may be used as classifiers to classify the wearer.
  • test data are classified as two classes.
  • the classifier may determine three or more classes. In such implementations, the classifier may be implemented to authenticate three or more users.
  • a C-class LDA classifier may be used, where C is not fewer than three.
  • a binary LDA provides a method to classify data in a multidimensional space into two classes: a target class and a non-target.
  • Data points from each of the two classes are provided to train the classifier.
  • the LDA projects the data points into a new space that best separates the two classes of the data points. More specifically, the projected data points have an optimal combination of mean difference and class variance, with the largest difference in means of the two classes and the smallest variance within each class.
  • the LDA determines a hyperplane that separate the two classes of data. The projections of data points from the same class are very close to each other and at the same time the projected means of the two classes are as far apart from each other as possible.
  • the classifier is applied to a test vector.
  • the test vector belongs to the target class if the test vector is located on the same side of the hyperplane as the target class, and the location of the hyper plane is defined by a threshold value.
  • the process 1100 further involves obtaining a motion signature from motion data of an instant wearer, using the one or more motion sensors of the wearable fitness monitor. See block 1106 .
  • the process 1100 shown in FIG. 11 uses the same motion sensors and the same wearable fitness monitor to obtain both the reference motion feature and the motion signature.
  • the reference motion feature may be provided by sensors or wearable fitness monitors different from those producing the motion signature.
  • the reference motion feature may be imported from another wearable fitness monitor or computer, and then stored on the instant wearable fitness monitor or an instant computer performing one or more operations of process 1100 . Then the imported reference motion feature may be compared with the motion signature on the instant wearable fitness monitor or the instant computing device.
  • Operations 1102 and 1104 represent a different phase of the process from operations 1106 - 1110 , with 1102 and 1104 performed initially and the result used repeatedly in 1106 - 10 .
  • Training the classifier is qualitatively different from using the classifier.
  • 1102 and 1104 can be performed more than once to update the classifier as described in the example below. But typically a single trained classifier can be used repeatedly to determine identity.
  • the process involves obtaining two or more motion signatures from motion data of an instant wearer.
  • the two or more motion signatures comprise two or more motion cycle profiles.
  • the two more motion signatures comprise two different motion features, such as step rate and motion signal power.
  • Process 1100 involves comparing the motion signature to the reference motion signature by applying the classifier (e.g., a LDA classifier) to the motion signature. See block 1108 .
  • a LDA classifier e.g., an LDA classifier
  • other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA.
  • a feature vector is extracted from the reference motion feature, which feature vector is then provided as a data point belonging to the target class to train the LDA classifier.
  • the LDA classifier may also be trained by additional data points belonging to the target class and additional data points belonging to the nontarget class. When more data points are provided to train the LDA classifier, the confidence of classification may be improved.
  • the motion signature is analyzed to extract a feature vector, which can then be tested using the LDA classifier to determine whether the motion signature from the instant wearer matches the reference motion feature from the reference user.
  • the LDA classifier takes a feature vector as an input and provides a target or a non-target classification as an output.
  • the process can determine that the identity of the wearer of the fitness monitor if the reference user. See block 1110 .
  • two or more motion signatures when they are obtained, they can be combined into one value or function, which can then be compared to the reference motion feature.
  • two motion cycle profiles may be averaged, and then the average profile may be compared to a reference cycle profile.
  • values of two motion signatures can form a function (e.g., power as a function of step rate), which can then be compared to a reference function.
  • Other probabilistic approaches instead of or in addition to Na ⁇ ve Bayes (e.g., a mixture model of multiple probability functions) may also be used to combine multiple motion signatures.
  • FIG. 12 shows an implementation of a process 1200 for training a classifier and using the classifier to authenticate a wearer based on the classification result.
  • an LDA classifier is provided as an example, other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA.
  • Operation 1102 of obtaining a reference motion feature of the user may be implemented according to operations 1202 , 1204 , and 1206 of FIG. 12 .
  • Operation 1104 of training a linear discriminant analysis classifier may be implemented as the operations 1208 and 1112 of FIG. 12 .
  • operation 1106 of obtaining a motion signature may be implemented as operations 1202 , 1204 , and 1206 .
  • the operation 1108 of comparing the motion signature to the reference motion feature may be implemented as operation 1208 and the 1214 in FIG. 12 .
  • Operation 1110 of determining whether the motion signature matches the reference motion feature may be implemented as operation 1216 of FIG. 12 .
  • Process 1200 starts operating one or more motion sensors of a wearable fitness monitor to generate motion data. See block 1202 .
  • the one or more motion sensors are selected from accelerometers, gyroscopes, magnetometers, GPS sensors, etc.
  • the motion data include, for instance, about 2 minutes of data sampled at 25 Hz.
  • motion data include about 5 minutes of data.
  • reference data may be obtained when the wearable fitness monitor is worn by a reference user in a training phase. Then in a testing phase, data provided by an instant wearer are compared to the reference data. When the instant wearer's data match the reference data, the wearer's identity is determined to match the identity of the user, thereby providing authentication to the wearer.
  • the owner of the wearable fitness monitor is a reference user in this context.
  • Training data are collected from the owner of the wearable fitness monitor.
  • the wearable fitness monitor may determine whether an instant wearer of the wearable fitness monitor has the same identity as the owner, thereby authenticating the wearer as the reference user.
  • the wearable fitness monitor may experience motion caused by walking or running at various speeds.
  • data from different speeds or speed ranges are used to train and generate different classifiers.
  • an LDA classifier is provided as an example, other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA.
  • the different classifiers will be applied for data associated with different speeds.
  • data generated by motion at different speeds or speed ranges may be used to train and generate a single LDA classifier.
  • data are normalized on the time dimension to obtain a single cycle profile of the motion data.
  • movement speed may be provided as a feature of a feature vector, and the feature vector being provided to train the LDA classifier or test the LDA classifier.
  • motion data includes data of one, two or three motion sensors.
  • each motion sensor includes three axes.
  • motion data from one axis of a motion sensor are used. Such implementations can provide effective classification and efficient analysis when signal is sufficiently strong.
  • Process 1200 involves preprocessing motion data generated by the motion sensors using various techniques. See block 1204 .
  • preprocessing techniques described herein are optional.
  • raw motion sensor data are low-pass filtered to smooth the data.
  • data may be smoothed by a rolling time window.
  • local minima are obtained from the smoothed data.
  • FIG. 13 shows an example of acceleration data as a function of time. As shown in the example, the filtering and smoothing provide are more regular and cyclic data. The smoothed data can then be segmented into stride profiles or cycle profiles. See block 1206 of FIG. 12 .
  • FIG. 14 shows motion data depicting multiple stride profiles from a same subject.
  • data of multiple stride profiles are normalized on the time dimension.
  • a profile of a mean stride is calculated from multiple strides.
  • outlier profiles that deviate from the mean stride over a criterion are removed.
  • a new mean stride is obtained from stride profiles having the outliers removed.
  • outlier removal and averaging are performed for additional iterations to further improved representativeness of the obtained mean stride profile.
  • one cycle profile of a mean stride is obtained at operation 1206 .
  • two or more cycle profiles may be obtained from the motion data to provide training data to train the classifier.
  • cycle profiles from multiple users are obtained, which are then used to train the classifier.
  • At least one cycle profile belongs to the target class of the reference user, and at least one cycle profile belongs to the nontarget class other than the reference user.
  • FIG. 15 shows eight mean stride cycle profiles for eight different subjects.
  • Process 1200 further involves extracting at least one feature vector from the at least one cycle profile. See block 1208 .
  • signal values based on amplitude may be used to extract features such as slopes, minima, maxima, accelerations, relative distance between features, etc.
  • features may be based on moments, cumulants, time frequency domain functions, etc.
  • additional features may be added to the feature vector to train or test the classifier.
  • the additional features include motion data not reflected by the cycle profile. For instance, speed of motion may be used as an additional feature.
  • other biometric data may be used to provide additional features of the feature vector, which can then be used to train the classifier or be tested using the classifier. For instance, heart rate, blood pressure, respiration rate, skin color, and other metrics described herein may be included as additional features.
  • Process 1200 further involves determining whether to train the classifier or to apply the classifier to authenticate a wearer. See block 1210 .
  • An LDA classifier is an example of a suitable classifier. Other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA.
  • the process determines to train the LDA classifier the motion data is configured to be obtained from the reference user.
  • the reference user's identity is the reference identity, against which an instance wearer's identity is compared to. If the instant wearer's identity matches that of the reference user, the instant wearer is authenticated as the reference user.
  • feature vectors are used to train the classifier. See block 1212 .
  • the at least one cycle profile obtained in 1206 , along non-target data, is used to train the LDA classifier.
  • the at least one cycle profile is equivalent to the reference motion feature in blocks 1104 , 1108 , and 1110 of FIG. 11 .
  • at least one feature vector of a target cycle profile and at least one feature vector from a non-target cycle profile are obtained to train the LDA classifier.
  • the target cycle profile is obtained from the reference user.
  • the non-target profile can be obtained from a person other than the reference user.
  • the non-target profile may be obtained from the reference user performing a motion that is different from a target motion.
  • numerous feature vectors are obtained from the reference user to train the LDA classifier.
  • the multiple feature vectors obtained from the reference user provide data points of the target class.
  • one or more feature vectors obtained from motion data from individuals other than the reference user are also provided as data points of the nontarget class to train the LDA classifier.
  • process 1200 determines whether to continue to train the LDA classifier with additional data. See block 1222 . If the decision is positive, process 1200 returns to operation 1202 to generate more motion data using the one or more motion sensors of the wearable fitness monitor. The training LDA classifier operations described above are repeated.
  • operation 1210 decides to apply the LDA classifier to determine whether the instant wearer is the reference user.
  • the at least one cycle profile obtained in block 1206 is equivalent to the motion signature in blocks 1106 , 1108 , and 1110 in FIG. 11 .
  • the process uses the LDA classifier by applying it to the at least one feature vector. See block 1214 .
  • this operation compares the feature vector obtained from the motion data of the instant wearer to that obtained from the reference user.
  • the LDA classifier takes the feature vector as an input and provides an output of a classification of whether or not the data belongs to the target class, i.e., the reference user.
  • the wearable fitness monitor provides authentication to the wearer, determining that the wearer has the same identity as the reference user. See block 1218 . Otherwise, the process does not provide authentication to the wearer, or requires the wearer to provide an alternative method to authenticate the wearer's identity. For instance, the wearer may be required to provide a fingerprint, password, retina scan, a heart rate measurement, or other biometric data for authenticating the wearer. In some implementations, the wearer may be required to provide more motion data to repeat the authentication operations described above.
  • process 1200 proceeds to determine whether to continue to train or use the LDA classifier. If the decision is positive, the process loops back to operation 1202 to generate more motion data to train or use the LDA classifier.
  • decision 1222 determines to continue to train the LDA classifier.
  • Such an implementation may provide continuous learning of the classifier using data reflecting long-term change of the reference user's motion.
  • the motion change may be due to physiological or environmental changes associated with the user. For instance, a user may be injured and as a result develop a different motion cycle profile. In some instances, the user may have different characteristics of motion at different times of a day. In some instances, a same user may have different cycle profile during different kinds of activities.
  • the continue training of the LDA classifier may improve the classifier's ability to account for these different factors.
  • the continual training of the LDA classifier can improve the confidence of the classification. For instance, two or more classifiers may be generated for the same user depending on certain factors, such as time of the day or activity types. In some implementations, values of additional factors may be included into the feature vector used to train the LDA classifier or other types of classifier.
  • decision 1222 operates to continue to apply the LDA classifier to multiple sets of cycle profiles.
  • each set of cycle profiles may be classified by the classifier.
  • multiple classification results can be combined by probabilistic methods to obtain a final classification result.
  • the multiple classification results may be combined by Bayesian Fusion using Na ⁇ ve Bayes. See block 1223 outlined by dashed lines indicating an optional operation.
  • the final classification result can then be used to determine if a user should be authenticated.
  • the authentication of 1218 is modified to be contingent on the final classification result meeting an authentication criterion.
  • process 1200 can determine not to continue to train or use the LDA classifier, and the process comes to an end at block 1224 .
  • the data set is randomly divided into training and testing sets.
  • the graph shows the performance of the classifier on these 8 subjects.
  • the wearable fitness monitor is one component of a system that comprises a secondary device capable of communicating with the wearable fitness monitor.
  • the secondary device may be a smart phone, a PDA, a tablet, or a computer.
  • the secondary device may have a shape and mechanical and/or magnetic interface to accept the wearable fitness monitor for safe keeping, communication, and/or charging.
  • the communication between the wearable fitness monitor and the secondary device may be provided through wireless communication techniques/methods and protocols mentioned elsewhere herein.
  • a secondary device performs the biometric matching between a wearer's motion signature and a user's reference feature.
  • the secondary device may comprise sensors to assist in biometric or environmental monitoring such as, for example, sensors that measure ambient light, noise and/or sound (e.g., to detect snoring), temperature, humidity, and air quality (pollen, dust, CO2, etc.).
  • the secondary device may communicate with an external service such as www.fitbit.com or server (e.g., personal computer). Communication may be achieved through wired or wireless circuitry and protocols to transfer data to and/or from the secondary device. As examples, any of the wireless technologies described above for the fitness monitor may be used.
  • the secondary device may also act as a relay to transfer data to and/or from the wearable fitness monitor to an external service such as www.fitbit.com or other service (e.g., news, social network updates, email, calendar notifications).
  • Calculation of the user's fitness data may be executed on one or both devices or an external service (e.g., a cloud server) using data from one or both devices.
  • one or more of the operations performed to identify a user are performed on a secondary device. It should be understood that some or all of the operations may be performed on the wearable fitness monitor. Often, the operations are divided between the wearable fitness monitor and the secondary device.
  • the wearable fitness monitor may be used as a proxy to authorize a credit card or other secure service.
  • data of the wearable fitness monitor for an authorized user may be compared to the user's data of a smart phone. The comparison results may be used to authorize a user.
  • a user may be authorized, deauthorized, and or reauthorized.
  • the authorization, deauthorization, and reauthorization may be performed “online” on the wearable fitness monitor.
  • the classifier logic may execute on the wearable fitness monitor.
  • the authorization, deauthorization, and reauthorization may be performed on the smart phone. In such cases, the classifier logic may execute on the smart phone.
  • the authorization, deauthorization, and reauthorization may be performed “offline” on a back end server over a network.
  • the classification logic may be implemented on the back-end servers.
  • the authorization, deauthorization, and reauthorization may be performed using the wearable fitness monitor, the smart phone, and the back end server. In such cases, portions of the classification logic is split among those devices/systems.
  • a wearable fitness monitor may be implemented in a wearable fitness monitor as machine-readable instruction sets, either as software stored in memory, as application-specific integrated circuits, field-programmable gate-arrays, or other mechanisms for providing system control.
  • Such instruction sets may be provided to a processor or processors of a wearable fitness monitor to cause the processor or processors to control other aspects of the wearable fitness monitor to provide the functionality described above.
  • the present disclosure is neither limited to any single aspect nor implementation, nor to any single combination and/or permutation of such aspects and/or implementations.
  • each of the aspects of the present disclosure, and/or implementations thereof may be employed alone or in combination with one or more of the other aspects and/or implementations thereof.
  • many of those permutations and combinations will not be discussed and/or illustrated separately herein.

Landscapes

  • Health & Medical Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Biophysics (AREA)
  • Pathology (AREA)
  • Biomedical Technology (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Physiology (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Dentistry (AREA)
  • Radar, Positioning & Navigation (AREA)
  • General Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Cardiology (AREA)
  • Remote Sensing (AREA)
  • Human Computer Interaction (AREA)
  • Computer Hardware Design (AREA)
  • Business, Economics & Management (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Mathematical Physics (AREA)
  • Computer Security & Cryptography (AREA)
  • Physical Education & Sports Medicine (AREA)
  • Evolutionary Computation (AREA)
  • Fuzzy Systems (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Human Resources & Organizations (AREA)

Abstract

The disclosure relates to methods, devices, and systems to identify a user of a wearable fitness monitor using data obtained using the wearable fitness monitor. Data can be obtained from motion sensors of the wearable fitness monitor. A first motion signature and a second motion signature may be obtained from the motion data and used to identify the user.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims benefits under 35 U.S.C. §119(e) to U.S. Provisional Patent Application No. 62/202,773, entitled: USER IDENTIFICATION VIA COLLECTED FITNESS DATA, filed Aug. 7, 2015, which is herein incorporated by reference in its entirety for all purposes.
  • BACKGROUND
  • Recent consumer interest in personal health has led to a variety of personal health monitoring devices being offered on the market. Recent advances in sensor, electronics, and power source miniaturization have allowed the size of fitness monitoring devices, also sometimes referred to as “biometric tracking devices,” or “biometric monitoring devices,” “wearable fitness monitors,” “fitness monitors,” etc. to be offered in small sizes that were previously impractical.
  • While such monitors have gained widespread acceptance and commercial success, the use of fitness data and other data obtained by such claims is still limited.
  • SUMMARY
  • Methods and systems are provided for determining the identification of a user of a wearable fitness monitors or authenticating the user of the wearable fitness monitors.
  • One aspect of the disclosure relates to methods for using two motion signatures to determine an identity of a wearer of a wearable fitness monitor. In some implementations, a method comprising: (a) obtaining a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a second motion signature obtained using data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor; (c) comparing the first and second motion signatures or a combination thereof to a reference motion feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • In some implementations, the one or more first motion sensors comprise an accelerometer, a gyroscope, a magnetometer, an altimeter, a GPS receiver, or any combination thereof. In some implementations, the one or more second motion sensors comprise one or more of the first motion sensors. In some implementations, the first motion signature is a time domain representation of a periodic motion of the movement experienced by the wearable fitness monitor. In some implementations, the periodic motion is produced by a movement of the wearer, wherein the movement is a step rate, a metric derived from an amplitude of the time domain representation, a biking cadence, a rowing rate, a resistance-based repetition rate, a typing speed, a zero crossing rate, a peak-to-peak time, an arm swing rate, and any combination thereof.
  • In some implementations, the first motion signature comprises a frequency domain representation of a periodic motion of the movement experienced by the wearable fitness monitor. In some implementations, the frequency domain representation of the periodic motion of the movement experienced by the wearable fitness monitor comprises a spectral component in the periodic motion. In some implementations, the spectral component is a harmonic in the frequency domain representation of the periodic motion. In some implementations, the first motion signature comprises a property of two or more harmonics in the frequency domain representation of the periodic motion.
  • In some implementations, the first motion signature comprises a motion periodicity and the second motion signature comprises a metric derived from an amplitude of the motion periodicity. In some implementations, the first motion signature comprises a time domain representation of a periodic motion and the second motion signature comprises a frequency domain representation of the periodic motion. In some implementations, the user's reference motion feature comprises a model of typical motion of the user. In some implementations, the user's reference motion feature is generated using data obtained from the one or more first motion sensors and the one or more second motion sensors when the user wears the wearable fitness monitor. In some implementations, the user's reference motion feature comprises a profile of a step by the user.
  • In some implementations, operation (c) of the method comprises comparing a combination of the first and second motion signatures to the user's reference motion feature, and wherein the user's reference motion feature comprises a line, a curve, or a look up table relating the first and second motion signatures for the user. In some implementations, (c) comprises (i) determining a difference between the user's reference motion feature and the first and second motion signatures or the combination thereof, and (ii) determining whether the difference is greater than a threshold. In some implementations, (c) comprises performing a linear discriminant analysis on the first and second motion signatures or the combination thereof with respect to the user's reference motion feature. In some implementations, (c) comprises determining that at least one of the first and second motion signatures is an invalid motion for a human wearer.
  • In some implementations, the method further comprises determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user, wherein the one or more second motion sensors are located on a device that is separate from the wearable fitness monitor. In some implementations, the separate device is a mobile phone. In some implementations, determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user comprises determining whether the first and second motion signatures represent a characteristic of a periodic motion. In some implementations, the first motion signature comprises a step count or a step rate and the second motion signature comprises a GPS or Bluetooth signature.
  • In some implementations of the methods above, at least one of the first and second motion signatures comprises a cycle profile of a periodic motion performed by the user, and wherein the reference motion feature is a predetermined typical cycle for the user's periodic motion. In some implementations, the cycle profile comprises a time varying amplitude of an output from the one or more first motion sensors. In some implementations, the user's periodic motion is selected from the group consisting of running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, and riding an animal.
  • In some implementations, the method further includes repeating (a)-(d) at multiple times. In some implementations, wherein repeating (a)-(d) is performed automatically, without triggering by the wearer of the fitness monitor. In some implementations, the method further includes, responsive to determining that the identity of the wearer of the fitness monitor is not the user, preventing the wearable fitness monitor from allowing a transaction. In some implementations, the transaction comprises accessing a secure item or providing the user with an award for meeting an activity threshold.
  • In some implementations, the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is not the user, requiring the user to authenticate himself or herself. In some implementations, requiring the user to authenticate comprises requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing.
  • In some implementations, the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is not the user, discrediting a fitness metric obtained for the user via the wearable fitness monitor.
  • In some implementations, the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is the user, crediting a fitness metric obtained for the user via the wearable fitness monitor.
  • In some implementations, the method further involves, responsive to determining that the identity of the wearer of the fitness monitor is the user, allowing the wearable fitness monitor to facilitate a transaction. In some implementations, the transaction comprises accessing a secure item or providing the user with an award for meeting an activity threshold.
  • In some implementations, at least one of the one or more first motion sensors and the one or more second motion sensors are the same sensors.
  • In some implementations, (c) comprises obtaining a function between the first motion signature and the second motion signature, and comparing the function to a reference function based on the reference motion feature.
  • In some implementations, (c) comprises obtaining an average motion signature by averaging the first motion signature and the second motion signature, and comparing the average motion signature to the reference motion feature.
  • In some implementations, (c) comprises: extracting features from the first motion signature and the second motion signature; forming a feature vector using the extracted features, and applying a classifier to the feature vector to determine whether the feature vector belongs to a class corresponding to the reference motion feature.
  • In some implementations, the obtaining the first motion signature obtained using the data from the one or more first motion sensors comprises: low-pass filtering the data from the one or more first motion sensors; and obtaining a cycle profile from the low-pass filtered data. In some implementations, the obtaining the cycle profile from the low-pass filtered data comprises: obtaining local minima from the low-passed filtered data; dividing the low-passed filtered data into two or more segments using the local minima; and obtaining the cycle profile from the two or more segments. In some implementations, the obtaining the cycle profile from the two or more segments comprises: (i) rejecting one or more outliers among the two or more segments that deviate from the mean of the two or more segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: repeating (i) one or more times among remaining segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: averaging remaining segments to obtain the cycle profile.
  • In some implementations, the obtaining the first motion signature obtained using the data from the one or more first motion sensors further comprises: extracting one or more features from the cycle profile or values derived from the cycle profile. In some implementations, each feature is selected from the group consisting of: a slope, an inflection, a zero crossing, a derivative, a moment, a cumulant, and any combination thereof. In some implementations, comparing the first motion signature to the reference motion feature for the user comprises: obtaining a classifier using motion data obtained from the user; and applying the classifier to the extracted one or more features, wherein the classifier takes the one or more features as inputs and provides a classification of the wearer being the user or not the user as an output.
  • In some implementations, the classifier comprises a linear discriminant analysis classifier. In some implementations, the classifier comprises a neural network classifier. In some implementations, the classifier is trained using at least one cycle profile derived from motion data obtained from the user. In some implementations, the method further involves updating the classifier using additional motion data obtained from the user.
  • Another aspect of the disclosure relates to methods for using a motion signature and a heartbeat waveform signature or a body characteristic to determine an identity of a wearer of a wearable fitness monitor. In some implementations, a method involves: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a heartbeat waveform signature obtained using data from one or more heartbeat waveform sensors, wherein the heartbeat waveform signature characterizes a detected heartbeat waveform of a wearer of the wearable fitness monitor; (c) comparing the motion signature and the heartbeat waveform signature or a combination thereof to one or more reference features of a user; and (d) based on the comparison in (c), determining whether an identity of the wearer of the fitness monitor is the user.
  • In some implementations, a method of identifying a user is provided. The method involves: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a body characteristic obtained using data from one or more body characteristic sensors, wherein the body characteristic characterizes the body of a person wearing the wearable fitness monitor; (c) comparing the motion signature and the body characteristic or a combination thereof to at least one reference feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • In some implementations, the body characteristic is a detected response of the one or more body characteristic sensors to the wearer's skin. In some implementations, at least one of the one or more body characteristic sensors comprises a light pulse emitter and a light pulse detector configured to determine a variable response of the detector to a variable intensity of light pulses from the emitter.
  • In some implementations, the body characteristic is body composition determined through bioelectrical impedance. In some implementations, at least one of the one or more body characteristic sensors is disposed on the wearable fitness monitor.
  • A further aspect of the disclosure relates to methods for using a motion signature to determine an identity of a wearer of a wearable fitness monitor. In some implementations, a method includes: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) comparing the motion signature to a reference motion feature for a user; and (c) based on the comparison in (b), determining whether an identity of a wearer of the fitness monitor is the user. In some implementations, the motion signature characterizes a cycle of periodic movement of the person wearing the wearable fitness monitor.
  • In some implementations, the reference motion feature is a reference cycle for periodic movement of a user. In some implementations, the reference cycle is a predetermined typical cycle for the user's periodic motion. In some implementations, the motion signature comprises a time-varying amplitude of an output from the one or more motion sensors.
  • In some implementations, the user's periodic motion is selected from the group consisting of running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, and riding an animal. In some implementations, the reference motion feature is a characteristic of a periodic motion. In some implementations, the reference motion feature is a metric derived from an amplitude.
  • In some implementations, the obtaining the motion signature obtained using the data from the one or more motion sensors comprises: low-pass filtering the data from the one or more motion sensors; and obtaining a cycle profile from the low-pass filtered data. In some implementations, the obtaining the cycle profile from the low-pass filtered data comprises: obtaining local minima from the low-passed filtered data; dividing the low-passed filtered data into two or more segments using the local minima; and obtaining the cycle profile from the two or more segments.
  • In some implementations, the obtaining the cycle profile from the two or more segments comprises: (i) rejecting one or more outliers among the two or more segments that deviate from the mean of the two or more segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: repeating (i) one or more times among remaining segments. In some implementations, the obtaining the cycle profile from the two or more segments further comprises: averaging remaining segments to obtain the cycle profile.
  • In some implementations, the obtaining the motion signature obtained using the data from the one or more motion sensors further comprises: extracting one or more features from the cycle profile or values derived from the cycle profile. In some implementations, each feature is selected from the group consisting of: a slope, an inflection, a zero crossing, a derivative, a moment, a cumulant, and any combination thereof.
  • In some implementations, comparing the motion signature to the reference motion feature for the user comprises: obtaining a classifier using motion data obtained from the user; and applying the classifier to the extracted one or more features, wherein the classifier takes the one or more features as inputs and provides a classification of the wearer being the user or not the user as an output. In some implementations, the classifier comprises a linear discriminant analysis classifier.
  • An additional aspect of the disclosure relates to methods for using a motion signature to determine that the motion of a wearable fitness monitor is generated by non-human, and preventing a transaction based on the determination. In some implementations, a method includes: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a body movement of the person wearing the wearable fitness monitor; (b) determining whether the motion signature corresponds to an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human; and (c) based on the determination in (b), preventing the wearable fitness monitor from allowing a transaction.
  • In some implementations, the transaction comprises accessing a secure item or providing an award for meeting an activity threshold to a user associated with the wearable fitness monitor. In some implementations, the method further involves, responsive to determining that the motion signature corresponds to the invalid motion feature, requiring a wearer of the fitness monitor to authenticate himself or herself In some implementations, requiring the wearer to authenticate comprises requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing.
  • In some implementations, determining whether the motion signature corresponds to an invalid motion feature in (b) comprises: obtaining an additional signature using data from one or more additional sensors, and determining that the motion signature, the additional signature, or a combination thereof is inconsistent with at least one human activity. In some implementations, the motion signature is step rate or step count and the additional signature is a heart rate or a heartbeat waveform.
  • In some implementations, determining whether the motion signature corresponds to the invalid motion feature in (b) comprises determining whether a periodicity for the motion signature is within a threshold periodicity for a given time period. In some implementations, the invalid motion feature is a periodic motion having a cycle-to-cycle consistency greater than a threshold. In some implementations, the invalid motion feature comprises one or more periodic motion contributions from one or more spatial dimensions that is less than a threshold.
  • Yet another aspect of the disclosure relates to methods for using a motion signature obtained from a wearable fitness monitor and an additional signature obtained from another device to determine if a wearer of the monitor is a particular user. In some implementations, a method involves: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a user, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining an additional signature obtained using the data from one or more additional sensors located on a device that is separate from the wearable fitness monitor, wherein the additional motion signature further characterizes the movement; (c) comparing the motion signature to the additional signature; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • In some implementations, the motion signature and the additional signature are obtained from the data collected at the same time. In some implementations, the comparing in (c) comprises determining whether the motion signature and the additional signature represent the same activity or activity level of the user. In some implementations, determining whether the motion signature and the additional signature represent the same activity or activity level of the user comprises determining whether the motion signature and the additional signature represent a characteristic of the user's gait.
  • In some implementations, the separate device is a mobile phone. In some implementations, the motion signature comprises a step count or a step rate and the second motion signature comprises a GPS or Bluetooth signature.
  • Another aspect of the disclosure relates to systems and devices for implementing various methods described above. In some implementations, a system includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, one or more second motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic. The classification logic is configured to: (a) obtain a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtain a second motion signature obtained using data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor; (c) compare the first and second motion signatures or a combination thereof to a reference motion feature for a user; and (d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
  • In some implementations, a system includes (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, one or more heartbeat waveform sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic. The classification logic is configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor, (b) obtain a heartbeat waveform signature obtained using data from the one or more heartbeat waveform sensors, wherein the heartbeat waveform signature characterizes a detected heartbeat waveform of a wearer of the wearable fitness monitor, (c) compare the motion signature and the heartbeat waveform signature or a combination thereof to one or more reference features of a user, and (d) based on the comparison in (c), determine whether an identity of the wearer of the fitness monitor is the user.
  • In some implementations, a system for identifying a user includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, one or more body characteristic sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic. The classification logic is configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor, (b) obtain a body characteristic obtained using data from the one or more body characteristic sensors, wherein the body characteristic characterizes the body of a wearer of the wearable fitness monitor, (c) compare the motion signature and the body characteristic or a combination thereof to at least one reference feature for a user, and (d) based on the comparison in (c), determine whether an identity of the wearer of the wearable fitness monitor is the user.
  • In some implementations, a system includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor, (b) compare the motion signature to a reference motion feature for a user, and (c) based on the comparison in (b), determine whether an identity of a wearer of the fitness monitor is the user.
  • In some implementations, a system includes: (A) a wearable fitness monitor configured to be worn by a person and comprising: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and (B) classification logic configured to: (a) obtain a motion signature obtained using data from the one or more motion sensors, wherein the motion signature characterizes a body movement of a wearer of the wearable fitness monitor, (b) comparing the motion signature to an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human; and (c) based on the determination in (b), prevent the wearable fitness monitor from allowing a transaction.
  • In some implementations, a system includes: a wearable fitness monitor configured to be worn by a user and comprising: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; one or more additional sensors located on a device that is separate from the wearable fitness monitor; and classification logic. The classification logic is configured to: (a) obtain a motion signature obtained using data from the one or more first motion sensors, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtain an additional signature obtained using data from the one or more additional sensors, wherein the additional motion signature further characterizes the movement; (c) compare the motion signature to the additional signature; and (d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
  • These and other objects and features of the present disclosure will become more fully apparent from the following description and appended claims, or may be learned by the practice of the disclosure as set forth hereinafter.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The various implementations disclosed herein are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings, in which like reference numerals may refer to similar elements.
  • FIG. 1 depicts a generalized schematic of an example wearable fitness monitor with which various operations described herein may be executed.
  • FIG. 2 shows one implementation of a wearable fitness monitor having an accelerometer, illustrating a coordinate system of the accelerometer.
  • FIG. 3 shows a representative 3-axis accelerometer signal from a wearable fitness monitor worn on the wrist of a user who is walking.
  • FIG. 4 depicts the step rate and signal energy for 4 unique users.
  • FIG. 5 depicts the step rate and signal energy for 3 unique users for a run.
  • FIG. 6 depicts power spectral density of motion data of two unique users who are otherwise indistinguishable in running by step rate and signal energy.
  • FIG. 7 shows examples of heart beat waveform data obtained from post-processing.
  • FIG. 8 shows a schematic illustration of a PQRST heart beat waveform.
  • FIG. 9 illustrates a number of time domain features of a PQRST waveform.
  • FIGS. 10A-10D depict representative features of the PPG waveform that may be used to identify the user.
  • FIG. 11 shows a flowchart of a method for determining whether an identity of an instant wearer of a fitness monitor matches that of a user.
  • FIG. 12 shows a process 1200 for training an LDA classifier and using the classifier to authenticate a wearer based on the classification result.
  • FIG. 13 shows an example of acceleration data as a function of time.
  • FIG. 14 shows motion data depicting multiple stride profiles from a same subject.
  • FIG. 15 shows eight mean stride cycle profiles for eight different subjects.
  • FIG. 16 shows an example of classification results for two-minute walking data for eight subjects.
  • DETAILED DESCRIPTION
  • Numeric ranges are inclusive of the numbers defining the range. It is intended that every maximum numerical limitation given throughout this specification includes every lower numerical limitation, as if such lower numerical limitations were expressly written herein. Every minimum numerical limitation given throughout this specification will include every higher numerical limitation, as if such higher numerical limitations were expressly written herein. Every numerical range given throughout this specification will include every narrower numerical range that falls within such broader numerical range, as if such narrower numerical ranges were all expressly written herein.
  • The headings provided herein are not intended to limit the disclosure.
  • Unless defined otherwise herein, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skill in the art. Various scientific dictionaries that include the terms included herein are well known and available to those in the art. Although any methods and materials similar or equivalent to those described herein find use in the practice or testing of the embodiments disclosed herein, some methods and materials are described.
  • Context and Overview
  • This disclosure relates to methods, devices, and systems to recognize a user of a wearable fitness monitor using information obtained using the wearable fitness monitor.
  • Example Contexts which May Identify Users Via Wearable Fitness Monitors
  • In cases where a wearable fitness monitor is shared amongst several users, the wearable fitness monitor can provide data to the correct user's digital account.
  • When used to access secure resources, it may be useful for some embodiments that the wearable fitness monitor can estimate the identity of the user with some amount of certainty.
  • When data from the wearable fitness monitor is used to provide monetary incentives for user behavior, it may be useful for some embodiments that the wearable fitness monitor knows the identity of the user with some certainty. Additionally, trust of the user data itself (e.g., that a “step” is a true user step rather than a fake motion, hereinafter referred to as a “fake”) can be an important aspect to some embodiments.
  • When data from the wearable fitness monitor is used to compete with other users of wearable fitness monitors, the veracity of the data and the identity of the user can be an important feature to verify, which can ensure that the data used to compete is not faked.
  • Overview
  • In one aspect, this disclosure presents a method including the following operations: (a) obtaining a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a second motion signature obtained using the data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor; (c) comparing the first and second motion signatures or a combination thereof to a reference motion feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user. In some embodiments, the one or more first motion sensors include an accelerometer, a gyroscope, a magnetometer, an altimeter, a GPS receiver, or any combination thereof.
  • A “combination” of motion signatures in (c) may be used in cases where the signatures are not analyzed separately, but together as in the case of a point or a curve in a multi-dimensional signature space. In other embodiments, each of the first and second signatures is separately compared against the reference motion signature.
  • In certain embodiments, the “movement” experienced by the wearable fitness monitor is a voluntary body movement such as the wearer's intentional movement of her head, neck, eyelid, mouth, shoulder, arm, wrist, finger, torso, hips, knee, ankle, and/or toe. In various implementations, the movement is characterized by a wearer's gait or stride when walking or running. The movement may be associated with a particular activity type such as running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc. In some cases, the movement does not include involuntary motions such as heartbeats which may be determined using, e.g., sensor technology include photoplethysmography (PPG), electrocardiography (ECG), etc.
  • The comparison of the signature(s) and the reference motion signature may indicate that the wearer is not the user. Such cases may arise where the wearer is a different individual than the user and cases where the “wearer” is a robot or other automaton. The “reference motion feature” used in operation (c) may be a user reference motion feature (a template built using historical motion data of the user).
  • It should be understood that determining whether an identity of the wearer of the fitness monitor is the user may be a matter of probability or a prediction of whether the user is likely wearing the monitor. As explained elsewhere herein, the determination may be made by mathematical or other logical techniques that determine a distance or difference between the wearer's current motion signature(s) and the user's reference motion feature in movement signature space. As such, the determination may rely on a classification technique which provides a likelihood, rather than certainty, that the wearer is the user. For instance, classifiers using one or more of the following techniques may be used to determine the likelihood that the wearer is the user: linear discriminant analysis, neural network, clustering techniques, support vector machine, logistic regression, naive Bayes, random forest, decision tree, etc.
  • The sensor producing the data for the first and second motion signatures may be the same or different or they may overlap, with some first motion sensors being the same as some second motion sensors. In certain embodiments, the one or more second motion sensors include at least one motion sensor from the one or more first motion sensors.
  • In certain embodiments, the first motion signature is a time domain representation of a periodic motion of the movement experienced by the wearable fitness monitor, which periodic motion may be a person's gait, which may be, for example, a step rate, a metric derived from an amplitude of the wearer's periodic motion, a biking cadence, a rowing rate, a resistance-based repetition rate (for, e.g., weightlifting), a typing speed, a zero-crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof. In certain embodiments, the first motion signature is a frequency domain representation of a person's periodic motion (e.g., gait), which may include, for example, contributions of a spectral component (e.g., the fundamental frequency or a harmonic thereof) in the first motion signature. In some examples, the spectral component includes a combination of the harmonics or the fundamental frequency and one or more harmonics. As an example, a combination of the contributions of harmonics may be a property of two or more harmonics (or the fundamental frequency) such as a ratio of the powers of the individual harmonics.
  • Various combinations of the first and second motion signatures may be used. For example, the first motion signature includes a motion periodicity parameter (e.g., step rate) and the second motion signature includes a metric derived from an amplitude of the wearer's periodic motion. In another example, the first motion signature includes a time domain representation of a periodic motion (e.g., a wearer's gait) and the second motion signature includes a frequency domain representation of the periodic motion.
  • The user's reference motion feature typically has characteristics that facilitate comparison with the motion signatures. For example, the user's reference motion feature may include a predetermined typical motion signature for the user. Such feature may be obtained in various ways such as by using data obtained from the one or more first motion sensors and the one or more second motion sensors when the user wears the wearable fitness monitor. In such cases, it should be established that the user is actually wearing a fitness monitor when capturing data for generating her reference motion feature. In certain embodiments, the user's reference motion feature comprises a profile of a step by the user.
  • In some implementations, comparing the motion signature(s) to the user's reference motion feature in (c) includes comparing a combination of the first and second motion signatures to the user's reference motion feature. In such implementations, the user's reference motion feature may be a relationship between the first and second motion signatures for the user. As examples, the relationship may be a line, a curve, a look up table, etc. relating the first and second motion signatures for the user. In some examples, comparing the motion signature(s) to the reference motion feature includes (i) determining a distance (or difference) between the user's reference motion feature and the first and second motion signatures or the combination thereof, and (ii) determining whether the distance (or difference) is greater than a threshold. As explained elsewhere, the distance is a minimum separation between two points, lines, surfaces, etc. in multidimensional analysis. It may be viewed as a type of “difference.”
  • In some implementations, the comparison in (c) includes performing a linear discriminant analysis (LDA) on the first and second motion signatures or the combination thereof with respect to the user's reference motion feature. The operations and applications of LDA are further explained hereinafter.
  • In certain embodiments, the comparison in (c) includes determining that at least one of the first and second motion signatures is an invalid motion for a human user (e.g., it is unnatural for a human user). As an example, a machine provides movements from which the fitness monitor generates data having unnaturally high consistency over time or from cycle to cycle. Also, data produced from machine movements may be unnaturally limited to contributions from one or a few axes (e.g., one axis of a three-axis accelerometer or gyroscope). In various implementations, the comparison between motion signatures and a reference motion feature may be performed by various classification techniques such as LDA, neural network, clustering, logistic regression, support vector machine, naive Bayes, etc. In some embodiments, the method includes an operation of determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user. In some implementations, determining whether the first and second motion signatures, taken at the same time, represent the same activity or activity level of the user includes determining whether the first and second motion signatures represent a characteristic of a periodic motion. For example, the motion signatures may be compared to determine whether they represent the same gait of a wearer. As an example, the first motion signature includes a step count or a step rate and the second motion signature includes a GPS or Bluetooth signature. This approach may be particularly relevant where the one or more second motion sensors are located on a device that is separate from the wearable fitness monitor, such as a smart phone or a second monitoring device worn or carried by the wearer of the fitness monitor. In some examples, the separate device is a mobile phone or other portable device with one or more sensors.
  • The motion signatures may be or include representations of multiple cycles of the wearer's movement or even a single cycle. For example, at least one of the first and second motion signatures may include a cycle profile of a periodic motion performed by the user. In such cases, the reference motion feature may be a predetermined typical cycle for the user's periodic motion. In some implementations, the cycle profile is a time varying amplitude of an output from the one or more first motion sensors. In some implementations, the user's periodic motion is running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc.
  • In some cases, the process of obtaining the wearer's motion signatures and comparing them with the user's reference motion feature is performed repeatedly, sometime continuously. For example, the above method may involve repeating operations (a)-(d) at multiple times. The repeating of operations (a)-(d) may be performed automatically, without triggering by the wearer of the fitness monitor.
  • Determining whether or not the wearer of the fitness monitor is the user can be used in various contexts, often in the same or similar way as biometric information is conventionally used. When the system/method determines that the wearer is not the user, various actions may be taken to block a transaction involving the wearer, require the wearer to take additional steps to authenticate or otherwise identify herself, etc. In some cases, responsive to determining that the identity of the wearer of the fitness monitor is not the user, the method prevents the wearable fitness monitor from allowing a transaction. As examples, the transaction may be accessing a secure item or providing the user with an award for meeting an activity threshold, which may be determined from quantifiable biometric information. In certain embodiments, responsive to determining that the identity of the wearer of the fitness monitor is not the user, the method requires the user to authenticate himself or herself. As examples, requiring the user to authenticate may include requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing. In certain embodiments, responsive to determining that the identity of the wearer of the fitness monitor is not the user, the method discredits a fitness metric obtained for the user via the wearable fitness monitor.
  • When the system/method determines that the wearer is the user, various actions may be taken to credit the user or allow the user to engage in a transaction. In certain embodiments, responsive to determining that the identity of the wearer of the fitness monitor is the user, the method credits a fitness metric obtained for the user via the wearable fitness monitor. In certain embodiments, responsive to determining that the identity of the wearer of the fitness monitor is the user, the method allows the wearable fitness monitor to facilitate a transaction. Examples of such transactions include accessing a secure item or providing the user with an award for meeting an activity threshold.
  • In certain embodiments, at least one of the one or more first motion sensors and the one or more second motion sensors are the same sensors. In certain embodiments, the data from the first and second motions sensors includes at least a first datum from the one or more first motion sensors and a second datum from the one more second motion sensors.
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a heartbeat waveform signature obtained using data from one or more heartbeat waveform sensors, wherein the heartbeat waveform signature characterizes a detected heartbeat waveform of a wearer of the wearable fitness monitor; (c) comparing the motion signature and the heartbeat waveform signature or a combination thereof to one or more reference features of a user; and (d) based on the comparison in (c), determining whether an identity of the wearer of the fitness monitor is the user. In certain contexts, a wearer is a non-human wearer and, in certain cases, a “heartbeat waveform” is detected where the non-human lacks a heartbeat. The method may be employed to identify the user, deauthenticate the user, etc. as described above.
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining a body characteristic obtained using data from one or more body characteristic sensors, wherein the body characteristic characterizes the body of a person wearing the wearable fitness monitor; (c) comparing the motion signature and the body characteristic or a combination thereof to at least one reference feature for a user; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user. The method may be employed to identify the user, deauthenticate the user, etc. as described above.
  • Various types of body characteristic may be employed, some related to body morphology, some to body composition, some to body color, etc. In certain embodiments, the body characteristic is one or more characteristics of the wearer's skin (e.g., the wearer's skin color). Skin color or another skin characteristic can be determined (or approximated) by various techniques. In one implementation, at least one of the one or more body characteristic sensors includes a light pulse emitter and a light pulse detector configured to determine a variable response of the detector to a intensity of light pulses from the emitter. The variability of the response of the detector is can be influenced by the user's skin color. Thus, this response may be used as a signature that is compared against a reference feature of a user. In certain embodiments, the body characteristic is body composition determined through bioelectrical impedance. In some implementations, at least one of the one or more body characteristic sensors is disposed on the wearable fitness monitor.
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, where the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) comparing the motion signature to a reference motion feature for a user; and (c) based on the comparison in (b), determining whether an identity of a wearer of the fitness monitor is the user. In some implementations, data from the one or more motion sensors are preprocessed before the motion signature is obtained. In some implementations the reference motion feature may be updated continuously or periodically after the identity of the wearer has been determined. In some implementations, the comparison in operation (b) is implemented using an LDA classifier.
  • In some embodiments, the motion signature characterizes a cycle of periodic movement of the person wearing the wearable fitness monitor. A cycle contains information about a wearer's step or other unit of periodic motion. Examples of the user's periodic motion include running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, riding an animal, etc. In certain embodiments, a motion signature for a cycle includes a time-varying amplitude of an output from the one or more motion sensors. The motion signature for a cycle may be obtained for data generated during a single instance of the cycle or from multiple instances, with the instances being averaged or otherwise combined to provide the “cycle” used in this method. In embodiments where the motion signature is a cycle, the reference motion feature may be a reference cycle for periodic movement of a user. In some cases, such reference cycle is a predetermined typical cycle for the user's periodic motion. The method may be employed to identify the user, deauthenticate the user, etc. as described above. In certain embodiments, the reference motion feature is a characteristic of a periodic motion. In certain embodiments, the reference motion feature is a metric derived from an amplitude (e.g., an amplitude a reference cycle).
  • Another aspect of the disclosure pertains to methods including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, where the motion signature characterizes a body movement of the person wearing the wearable fitness monitor; (b) determining whether the motion signature corresponds to an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human; and (c) based on the determination in (b), preventing the wearable fitness monitor from allowing a transaction (or deauthenticating a user associated with the fitness monitor). In certain embodiments, the transaction includes accessing a secure item or providing an award for meeting an activity threshold to a user associated with the wearable fitness monitor. In some implementations, the invalid motion feature is simply an unnatural (for humans or non-machines) level of consistency or repetition in the motion signature. In a further example, an invalid motion feature has limited dimensional range; for example, the motion feature unnaturally (for a human or non-machine) emphasizes one or two spacial dimensions. The user identification logic can identify this level of consistency or limited dimensional range by comparing the relevant motion signature component to a defined threshold.
  • As mentioned unnatural motion signatures might be generated by automatons or other machines. Additionally, an unnatural motion signature may be generated by a non-human animal wearing the fitness monitor. In certain embodiments, an approach as described here allows a system or entity to deauthenticate a user, discredit a user, prevent access to a secure item, and/or prevent a transaction involving the user. For example, in certain embodiments, responsive to determining that the motion signature corresponds to an invalid motion feature, the method requires a wearer of the fitness monitor to authenticate himself or herself. In some implementations, requiring the wearer to authenticate includes requiring the wearer of the fitness monitor to input a passcode, a fingerprint, an iris image, an ECG, a facial image, a vocal message, or any combination of the foregoing. In certain embodiments, operation (c) includes determining that the motion signature matches an invalid motion feature, the invalid motion feature characterizing motion likely to be performed by a non-human.
  • In certain embodiments, determining whether the motion signature corresponds to an invalid motion feature in (b) includes obtaining an additional signature using data from one or more additional sensors, and determining that the motion signature and/or the additional signature are/is inconsistent with a human activity. In one example, the motion signature is step rate or step count and the additional signature is a heart rate or a heartbeat waveform. In certain embodiments, determining whether the motion signature corresponds to the invalid motion feature in (b) includes determining whether a periodicity for the motion signature is within a threshold periodicity for a given time period. In some in some embodiments, determining whether the motion signature corresponds to the invalid motion feature in (b) includes determining a relation (e.g., a function) between step rate and heart rate for human activities, and determining that a combination of the motion signature (e.g., measured step rate) and the additional signature (e.g., measured heart rate) is inconsistent with the relation determined for one or more human activities. In some implementations, a measured step rate is compared to a reference step rate range associated with human activities to determine whether the measure step rate is outside of the step rate range of human activities. In some implementations, measured heart rate is compared to a heart rate range of human activities to determine whether the measured heart rate is outside of the heart rate range of human activities.
  • Another aspect of the disclosure concerns a method including the following operations: (a) obtaining a motion signature obtained using data from one or more motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the motion signature characterizes a movement experienced by the wearable fitness monitor; (b) obtaining an additional signature obtained using data from one or more additional sensors located on a device that is separate from the wearable fitness monitor, wherein the additional motion signature further characterizes the movement experienced by the wearable fitness monitor, or characterizes movement experienced by the device that is separate from the wearable fitness monitor; (c) comparing the motion signature to the additional signature; and (d) based on the comparison in (c), determining whether an identity of a wearer of the fitness monitor is the user.
  • In certain embodiments, the motion signature and the additional signature are obtained from the data collected at the same time. In some implementations, the comparing in (c) includes determining whether the motion signature and the additional signature represent the same activity or activity level of the user. In some implementations, determining whether the motion signature and the additional signature represent the same activity or activity level of the user comprises determining whether the motion signature and the additional signature represent a characteristic of the user's gait. As an example, the separate device is a mobile phone, a second fitness monitor, a headset, or other portable device. In some cases, the motion signature includes a step count or a step rate and the second motion signature includes a GPS or Bluetooth signature.
  • One aspect of the disclosure relates to systems that use sensor data to verify the identity of a wearer of a wearable fitness monitor. In various implementations, the sensor data include data from one or more motion sensors of the wearable fitness monitor. Various implementations of the systems are configured to perform any of the methods described above.
  • In some implementations, the system includes a wearable fitness monitor configured to be worn by a person. The wearable fitness monitor includes one or more first motion sensors, one or more second motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor. The system also includes classification logic configured to perform operations to implement one or more methods described above. For example, in one implementation, the classification logic is configured to: (a) obtain a first motion signature obtained using data from the one or more first motion sensors of the wearable fitness monitor, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor, (b) obtain a second motion signature obtained using the data from the one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor, (c) compare the first and second motion signatures or a combination thereof to a reference motion feature for a user, and (d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
  • In other implementations, the wearable fitness monitor is configured to be worn by a person and includes: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor. In some implementations, the wearable fitness monitor also includes one or more heartbeat waveform sensors. In other implementations, the wearable fitness monitor also includes one or more body characteristic sensors.
  • In further implementations, the wearable fitness monitor includes: one or more first motion sensors, one or more body characteristic sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor.
  • In some implementations, the system includes (a) a wearable fitness monitor configured to be worn by a person. The wearable fitness monitor includes: one or more first motion sensors, and a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor. The system also includes (b) classification logic, and (c) one or more sensors located on a device that is separate from the wearable fitness monitor.
  • User Identification from Motion Signatures
  • As explained, the methods and systems described herein identify users of fitness monitors that produce data from a motion sensor that responds to the user's movements (e.g., motion of a limb, head, torso, wrist, etc.). Such voluntary movements are typically produced for purposes other than user identification or authentication, typically for the purpose of activity tracking and monitoring. They may be generated from the user's normal walking or fitness activities.
  • Motion signatures may include any data relating motion periodicity (otherwise referred to as periodic motion) obtained from the motion sensor are analyzed by, e.g., classification logic to determine whether the detected motions are those of the user. Examples of data relating to motion periodicity include step rate, a metric derived from the amplitude of the motion signal, a biking cadence, a rowing rate, a resistance-based repetition rate (e.g., weightlifting repetition rates), a typing speed, a zero crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof. A metric derived from the amplitude of the motion signal may be referred to herein as “motion signal energy.” Motion energy may be calculated through various methods, such as classical energy formulas as well as simplified calculations, such as finding the absolute difference of motion signals.
  • One or more motion signatures may be used to identify a user. When multiple motion signatures are used, they may be used separately or in combination to identify a user. One example of a combination is a line or curve relating two or more signature types for a user. For example, a user's motion signal energy may vary as a function of step rate in a reproducible manner. Another example of a combination is where the values of multiple motion signatures are hashed or otherwise used to index a lookup table. In some cases, sensors other than motion sensors in the wearable fitness monitor are used to help identify the user wearing the fitness monitor. Such other sensors include heartbeat waveform sensors (PPG sensors and ECG sensors), bioimpedance sensors, and EEGs, the like. In some cases, such sensors measure involuntary body motions such a heart beats and respiration. An example of combination of a motion signature and another signature is a relationship between heartrate and step rate for a user.
  • A positive or negative identification of a user may be applied in various contexts such as insurance (e.g., an insured's fitness level factors into her premium), automatically switching users sharing a wearable fitness device, accessing secure devices such as automobile ignition systems, door locks, media systems, etc., and fitness competitions where the method credits or discredits the user's fitness results.
  • Timing and Use of User Identification from the Wearable Fitness Monitor (Use Cases)
  • Identifying a user as described herein may be performed at various times and frequencies. In some cases, the identifying is performed at a time when she needs to enter into a transaction such as accessing a secure device. For example, the system may evaluate a current motion signature from a wearable fitness monitor to confirm that a user can execute a transaction (such as accessing a secure device, or receiving an insurance benefit) at the time when the user wants to execute the transaction. Such can occur when the activity tracker detects an interaction with a payment system, as may occur via an NFC protocol, a Bluetooth connection, or a triggering event initiated by an application executing on the activity tracker.
  • In some implementations, a user who wears a fitness monitor regularly over a period of days, weeks, months, years, etc. may have her identity checked periodically, with or without notifying the user. The classification logic may check the identity of the wearer of the fitness monitor automatically, without prompting from the user. Changes in status may occur with or without notifying the user, e.g., user identification logic may determine that a fitness monitor that was identified as being worn by the user is no longer being worn by the user. Or the classification logic may determine that it can no longer conclude with a requisite level of confidence that the monitor is being worn by the user. Such checks and adjustments may be made in the “background,” that is to say that they are performed without the user's input and/or without notifying the user.
  • In some implementations, evaluating motion signatures from a wearable fitness monitor results in deauthenticating a previously authenticated wearable fitness monitor and blocks execution of a transaction. In one sequence, the process or classification logic starts by authenticating a user by using a technique other than a continuous or quasi-continuous fitness tracking measurement based on gait, heartbeat waveform, etc. Examples of the initial authentication techniques include fingerprint capture, ECG measurement, personal identification number (PIN) entry, and/or bioimpedance measurement. Such techniques often require only a short duration (e.g., a minute or less) to authenticate a user. After the classification logic authenticates the user, the wearable fitness monitor motion signatures are continually or periodically evaluated by the classification logic to determine whether to maintain authentication or deauthenticate the user/device. Such evaluations may be conducted in the background by the classification logic during normal operation of the wearable fitness monitor. When the classification logic determines that the user is authenticated or deauthenticated, it may or may not notify user. In some embodiments, the user is not notified until she attempts to execute a transaction.
  • In some cases, the user identification logic initially authenticates the user using the short duration technique (e.g., fingerprint, ECG, bioimpedance, etc.), and the logic then acquires the motion signatures produced by the voluntary and/or involuntary actions of the user wearing the fitness monitor, and uses these motion signatures to train a profile or other relationship between the motion signatures and a specific user.
  • In some cases, the device or logic may prompt the user to re-authenticate by verifying a code provided by a trusted device (e.g., a mobile phone in which authentication is achieved via fingerprint or PIN code) or any of the motion, heartbeat, or other fitness monitor-based authentication methods described in this disclosure. Prompting may occur by a vibration or haptic interaction from the wearable fitness monitor. Prompting may occur at the next moment in which the wearable fitness monitor's user interface is engaged (e.g., upon pressing a button or providing a motion gesture such as moving a wrist wearable toward the face).
  • Wearable Fitness Monitors—Structure and Operation
  • Wearable fitness monitors suitable for use with the methods and systems described herein collect data for the user of the device such as activity, sleep, and physiological measures. Examples include steps, distance traveled, calories burned, pace, floors climbed, elevation, number of active minutes, the start and stop of a sleep period, the duration of sleep, the number of awakenings during sleep, sleep disturbances from external stimuli, sleep stages, apnea hypopnea index, heart rate, resting heart rate, maximum heart rate, heart rate variability, time spent in at a specific exertion level (e.g., “cardio” zone), blood pressure, arterial stiffness, cardiovascular fitness, blood glucose, stress and/or relaxation levels, power output on a bicycle, number of swimming laps in a pool, number of swimming strokes, type of swimming strokes, lap splits, running mile splits, path walked or run overland (e.g., via GNSS tracking), location, time spent at work, home or the gym, number and/or length of sedentary periods in a day, start and stop of exercise activities such as walking, running, elliptical, swimming, bicycling, cardio workout, VO2max, SpO2, proximity to and interactions with other wearable fitness monitor users, risk of arrhythmia, lactate threshold, hydration level, water loss during an exercise, body fat, number of reps, sets, and types of exercises performed in a resistance training session, yoga poses performed and duration, respiration rate, etc., etc. Please note that “wearable fitness monitors” used herein may be a standalone wearable fitness monitor, multiple wearable fitness monitors, or one or more wearable fitness monitors in communication with an external server(s) used together to track user activity.
  • Wearable fitness monitors are devices that are worn on or carried by a person. They come in many form factors: wrist band, watch, clip, shoe pod, shoe, pendant, earbuds, clothing (shirt, socks, pants, undergarments), belt, cuff links, glasses, ring, earring (nose ring, studs, etc.), helmet, hat, hair clip, and socks. For the sake of simplicity, a handheld mobile phone will also be classified as a wearable fitness monitor because it is either held by or worn on the person for significant periods.
  • In some implementations, a set of protective, attachable and/or wearable cases (herein referred to simply as “cases”) that enable a user to wear a single wearable fitness monitor in multiple fashions or body locations may be provided. For example, in some implementations, a wearable fitness monitor may be designed such that it may be inserted into, and removed from, a plurality of compatible cases. In other implementations, the wearable fitness monitors may be permanently or semi-permanently mounted into (or joined to) straps, clips, clasps, bands, or other attachments for wear. Generally speaking, the various individual elements of the various example cases and/or biometric tracking devices shown herein may also be combined with elements from other example cases and/or biometric tracking devices shown herein, e.g., a necklace or pendant case for a removable wearable fitness monitor may also be provided for a permanently-mounted wearable fitness monitor. Such combinations of elements are considered to be within the scope of this disclosure. Generally speaking, a wearable fitness monitor or biometric tracking device combined with a case or some other means allowing it to be worn or easily carried by a person may be referred to herein as a “biometric monitoring system” or “biometric tracking system.”
  • FIG. 1 depicts a generalized schematic of an example wearable fitness monitor or other device with which the various operations described herein may be executed. The wearable fitness monitor 102 may include a processing unit 106 having one or more processors, a memory 108, an operator interface 104, one or more biometric sensors 110, and input/output 112. The processing unit 106, the memory 108, the operator interface 104, the one or more biometric sensors 110, and the input/output 112 may be communicatively connected via communications path(s) 114 (it is to be understood that some of these components may also be connected with one another indirectly).
  • The wearable fitness monitor (also referred to herein as “the device”) may collect one or more types of biometric data, e.g., data pertaining to physical characteristics of the human body (such as heartbeat, perspiration levels, etc.) and/or data relating to the physical interaction of that body with the environment (such as accelerometer readings, gyroscope readings, etc.), from the one or more biometric sensors 110 and/or external devices (such as an external heart rate monitor, e.g., a chest-strap heart rate monitor) and may then store such information for later use, e.g., for communication to another device via the I/O 112, e.g., a smartphone or to a server over a wide-area network such as the Internet. The processing unit 106 may also perform an analysis on the stored data and may initiate various actions depending on the analysis. For example, the processing unit 106 may determine that the data stored in the memory 108 indicates that a goal threshold has been reached and may then display content on a display of the portable biometric tracking device celebrating the achievement of the goal. The display may be part of the operator interface 104 (as may be a button or other control, not pictured, that may be used to control a functional aspect of the wearable fitness monitor).
  • In general, a wearable fitness monitor may incorporate one or more types of user interfaces including but not limited to visual, auditory, touch/vibration, or combinations thereof. The wearable fitness monitor may, for example, display the state of one or more of the data types available and/or being tracked by the wearable fitness monitor through, for example, a graphical display or through the intensity and/or color of one or more LEDs. The user interface may also be used to display data from other devices or internet sources. The device may also provide haptic feedback through, for instance, the vibration of a motor or a change in texture or shape of the device. In some implementations, the biometric sensors themselves may be used as part of the user interface, e.g., accelerometer sensors may be used to detect when a person taps the housing of the biometric monitoring unit with a finger or other object and may then interpret such data as a user input for the purposes of controlling the wearable fitness monitor. For example, double-tapping the housing of the wearable fitness monitor may be recognized by the wearable fitness monitor as a user input that will cause the display of the wearable fitness monitor to turn on from an off state or that will cause the wearable fitness monitor to transition between different monitoring states, e.g., from a state where the wearable fitness monitor may interpret data according to rules established for an “active” person to a state where the wearable fitness monitor may interpret data according to rules established for a “sleeping” person.
  • In another example, while the user is wearing the wearable fitness monitor 102, the wearable fitness monitor 102 may calculate and store a user's step count while the user is wearing the wearable fitness monitor 102 and then subsequently transmit data representative of step count to the user's account on a web service like www.fitbit.com, to a mobile phone paired with the portable biometric monitoring unit, and/or to a standalone computer where the data may be stored, processed, and visualized by the user. Indeed, the device may measure, calculate, or use a plurality of other physiological metrics in addition to, or in place of, the user's step count. These include, but are not limited to, caloric energy expenditure, floors climbed or descended, heart rate, heart rate variability, heart rate recovery, location and/or heading (e.g., through GPS technology including a GPS receiver), elevation, ambulatory speed and/or distance traveled, swimming lap count, bicycle distance and/or speed, blood pressure, blood glucose, skin conduction, skin and/or body temperature, electromyography data, electroencephalographic data, weight, body fat, and respiration rate. Some of this data may be provided to the wearable fitness monitor from an external source, e.g., the user may input their height, weight, and stride in a user profile on a fitness-tracking website and such information may then be communicated to the biometric tracking device and used to evaluate, in tandem with data measured by the biometric sensors 110, the distance traveled or calories burned of the user. The device may also measure or calculate metrics related to the environment around the user such as barometric pressure, weather conditions, light exposure, noise exposure, and magnetic field.
  • As mentioned previously, collected data from the wearable fitness monitor may be communicated to external devices through the communications interface. The communications interface may include wireless communication functionality so that when the wearable fitness monitor comes within range of a wireless base station or access point, the stored data automatically uploads to an Internet-viewable source such as a website, e.g., www.fitbit.com. The wireless communications functionality may be provided using one or more communications technologies known in the art, e.g., Bluetooth, RFID, Near-Field Communications (NFC), Zigbee, Ant, optical data transmission, etc. Some of these communications technologies such as Bluetooth and NFC may be characterized as low power and/or short range in comparison to some other wireless communications technologies such as cellular and Wifi. In some embodiments, the wearable fitness monitor also contains wired communication capability, e.g., USB.
  • Other implementations regarding the use of short range wireless communication are described in U.S. patent application Ser. No. 13/785,904, titled “Near Field Communication System, and Method of Operating Same” filed Mar. 5, 2013 which is hereby incorporated herein by reference in its entirety.
  • It is to be understood that FIG. 1 illustrates a generalized implementation of a wearable fitness monitor 102 that may be used to implement a portable wearable fitness monitor or other device in which the various operations described herein may be executed. It is to be understood that in some implementations, the functionality represented in FIG. 1 may be provided in a distributed manner between, for example, an external sensor device and communication device, e.g., a chest-strap heart rate sensor that may communicate with a wearable fitness monitor.
  • Moreover, it is to be understood that in addition to storing program code for execution by the processing unit to affect the various methods and techniques of the implementations described herein, the memory 108 may also store configuration data or other information used during the execution of various programs or instruction sets or used to configure the wearable fitness monitor. It is to be further understood that the processing unit may be implemented by a general or special purpose processor (or set of processing cores) and thus may execute sequences of programmed instructions to effectuate the various operations associated with sensor device syncing, as well as interaction with a user, system operator or other system components. In some implementations, the processing unit may be an application-specific integrated circuit (ASIC) or programmable hardware, such as a FPGA.
  • Though not shown, numerous other functional blocks may be provided as part of the wearable fitness monitor 102 according to other functions it may be required to perform, e.g., environmental sensing functionality, etc. Other functional blocks may provide wireless telephony operations with respect to a smartphone and/or wireless network access to a mobile computing device, e.g., a smartphone, tablet computer, laptop computer, etc. The functional blocks of the wearable fitness monitor 102 are depicted as being coupled by the communication path 114 which may include any number of shared or dedicated buses or signaling links. More generally, however, the functional blocks shown may be interconnected using a variety of different architectures and may be implemented using a variety of different underlying technologies and architectures. With regard to the memory architecture, for example, multiple different classes of storage may be provided within the memory 108 to store different classes of data. For example, the memory 108 may include non-volatile storage media such as fixed or removable magnetic, optical, or semiconductor-based media to store executable code and related data and/or volatile storage media such as static or dynamic RAM to store more transient information and other variable data.
  • The various methods and techniques disclosed herein may be implemented through execution of one or more sequences of instructions, e.g., software programs, by the processing unit 106 (e.g., a generalized or specialized processor) or by a custom-built hardware ASIC (application-specific integrated circuit) or a programmable hardware device such as an FPGA (field-programmable gate array), or any combination thereof within or external to the processing unit 106.
  • Further implementations and implementations of wearable fitness monitors can be found in U.S. patent application Ser. No. 13/156,304, titled “Portable Biometric Monitoring Devices and Methods of Operating Same” filed Jun. 8, 2011, which is hereby incorporated herein by reference in its entirety.
  • In some implementations, the wearable fitness monitor may include computer-executable instructions for controlling one or more processors of the wearable fitness monitor to obtain biometric data from one or more biometric sensors. The instructions may also control the one or more processors to receive a request, e.g., an input from a button or touch interface on the wearable fitness monitor, a particular pattern of biometric sensor data (e.g., a double-tap reading), etc., to display an aspect of the obtained biometric data on a display of the wearable fitness monitor. The aspect may be a numerical quantity, a graphic, or simply an indicator (a goal progress indicator, for example). In some implementations, the display may be an illuminable display so as to be visible when displaying data but otherwise invisible to a casual observer. The instructions may also cause the one or more processors to cause the display to turn on from an off state in order to display the aspect of the biometric data.
  • Motion Sensors
  • Motion sensors provide an output a signal responsive to motion experienced. Examples of motion sensors include accelerometers, gyroscopes, compasses, switches (for example, mechanical), GPS modules, piezoelectric film and/or pedometers to determine, calculate and/or detect one or more steps of the user; notably, the exemplary motion sensor may be incorporated into portable monitoring devices such as wearable fitness monitors.
  • The portable monitoring device may estimate, calculate and/or determine, calorie consumption, burn and/or expenditure using data which is representative of the intensity of user motion—for example, as provided or determined by one or more single axis or multi-axis accelerometers. In one embodiment, the signals from the one or more accelerometers may be filtered using time domain or frequency domain filtering techniques to produce a parameter indicative of the intensity of user motion, often referred to as a “count”. A count may be computed as the sum of the rectified filtered accelerometer output taken over a suitable time epoch, for example, 10 seconds, with or without additional processing such as thresholding and/or saturation. The portable monitoring device may calculate, determine and/or estimate calorie consumption, burn and/or expenditure as a function of the current count value or a sequence of count values. Further descriptions of various motion sensors are provided in U.S. Patent Application Publication No. 2015/0134268 (Attorney Reference: FTBTP014D15C2US), titled PORTABLE MONITORING DEVICES AND METHODS OF OPERATING SAME, filed on Jan. 22, 2015, which is incorporated by reference in its entirety.
  • An accelerometer is often used as a motion sensor. For clarity and ease of discussion, this disclosure adopts a coordinate system as outlined in FIG. 2. It is to be understood that coordinate systems, as a rule, are a matter of convenience and may be arbitrarily defined—for example, a tri-axial accelerometer may be flipped upside down to reverse the orientations of two of the three axes of the tri-axial accelerometer or may be subjected to two 90° rotations about mutually-perpendicular axes to cause all three axes to be aligned differently from the conventions used herein.
  • To that end, it is to be understood that the techniques discussed herein may be practiced using tri-axial accelerometers (and their corresponding measurement outputs) that are aligned with coordinate systems different from the convention used in this disclosure (as outlined in FIG. 2). It is to be further understood that the data from tri-axial accelerometers that are aligned with other coordinate systems may still be used to perform the techniques discussed herein if the data from such tri-axial accelerometers is transformed in order to align with the coordinate system convention adopted herein or if the techniques outlined herein are adapted, e.g., transformed, to account for the shift in coordinate systems (for example, if an axis is reversed from the convention used herein, a condition stating that an acceleration along that axis be less than −0.125 g may have an equivalent condition in the new coordinate system that the acceleration along that axis be more than 0.125 g). Further descriptions of accelerometers are provided in U.S. Provisional Patent Application No. 62/054,345 (Attorney Reference: FTBTP003X1PUS), titled WRIST-WEARABLE DEVICE WITH WATCH-CHECK GESTURE ACTIVATION, filed on Sep. 23, 2014, which is incorporated by reference in its entirety.
  • It is also to be understood that the accelerations obtained from the accelerometer(s) may be first subjected to one or more pre-processing steps prior to being used in the present techniques. For example, the accelerations may be used in raw form (counts or accelerations converted from counts, for example) or may first be smoothed or otherwise processed (e.g., by using a moving average filter) to reduce noise and produce a more stable signal.
  • As discussed above, the techniques of concepts presented herein are intended to provide more reliable, more responsive recognition of motion signatures while simultaneously having a low impact on battery life. As a result, in some cases herein, various operations that are discussed may be performed slightly differently in actual practice. For example, as part of one technique, the magnitude of acceleration measured by a tri-axial accelerometer may be evaluated to see if it exceeds a threshold acceleration.
  • Another type of motion sensor is an angular motion measurement system. A detailed description of angular motion measurement systems is provided in U.S. Provisional Patent Application No. 62/054,341 (Attorney Reference: FTBTP002X1KP), titled HIGH-DYNAMIC RANGE ANGULAR MOTION SENSING SYSTEM, filed on Sep. 23, 2014, which is incorporated by reference in its entirety. Such systems may obtain angular motion measurement data using a hybrid system incorporating two different, non-gyroscopic angular motion measurement sensors. Such a system includes, at a minimum, a multi-accelerometer angular rate sensor (MAARS), an accelerometer/magnetometer angular rate sensor (AMARS), and logic for determining which of the two angular rate sensors (ARS's) were to actively used to collect data at any given instant in time.
  • Applications of Wearable Fitness Monitors Requiring User Identification
  • Control Secure Devices and Other Appliances or “Things”
  • In certain embodiments, wearable fitness monitors are used to control or facilitate control of electronic and/or digital devices and systems such as household appliances, automobiles, door locks, and the like. Such devices may be secure in the sense that they cannot be controlled or otherwise accessed without a form of authentication (e.g., a password) or other user identification. In various implementations, the fitness monitor serves to identify a user and allow access a secure electronic device or system. This form of identification may have other applications such as enabling access to secure physical areas or property, and customizing user experiences/interfaces for a service. Further examples include controlling a television, unlocking and/or opening the door of a residence, office, car, or other locked space, providing a digital ID and password (pin or otherwise) to access a computer, banking account, online shopping site, or other computer account, changing the music played in a room for a specific user, selecting the goods to display to a user on a shopping site, etc.
  • Rewards and Incentives for Meeting Activity Thresholds
  • Similarly, wearable fitness monitors may be used to authenticate or otherwise identify a user when monitoring the user's behavior toward a reduction of insurance premiums or related incentive rewards programs. For example, a user who on average walks more than 10,000 steps per day for a duration of 6 months may receive a cash reward. A user who performs medium intensity exercise for 10 minutes (or more) for 150 minutes (or more) a week for 1 month may pay a discounted medical insurance premium. A user who performs vigorous intensity exercise for 10 minutes (or more) for 75 minutes (or more) a week for 1 month may receive discount coupons or cash cards at a selected retailer. A user who performs on average more than N exercises per week of a minimum duration T every 4 months may receive a cash reward (e.g., N=5, T=5 min). A user who increases daily activity by 10% over a baseline daily activity value for 6 months may obtain a reduced insurance premium for the following 6 months. In each of these and other cases, the wearer of the fitness monitor must be identified as the user before such rewards are distributed to the user. This identification can occur at different points in time as the user is wearing the device and making progress towards the goal. The points in time may be selected according to a set frequency, period, schedule (with different intervals between identification points to appear random) or triggered based on a condition that considers progress toward a goal (e.g., if the wearable fitness monitor detects activity of a given step rate, time period of activity, or an activity being performed).
  • Fitness Competition
  • The data from wearable fitness monitor may also be used to compete against friends (e.g., in a social network), coworkers, or in a game. For example, a user who runs a specific path the fastest may be granted a title (e.g., “king of the hill”). A user who has the most steps in a week (or any other time period) may be placed at the top of a leaderboard. A user in a digital race game may be ranked against peers on the basis of running distance in a month. In each of these and other cases, the wearer of the fitness monitor must be identified as the user before the fitness monitor's measured metric is ascribed to the competing user.
  • In some cases the competition may be individually based where the metrics of an individual are ranked or compared against other individuals. In other cases, the competitions may be group based, where groups are formed from multiple individuals and the metrics of the members of the group are aggregated (e.g., summed, averaged, etc.) and the aggregated metric of a group is compared against the aggregated metric of the other groups. Embodiments may allow groups to be formed based on an attribute such as a department, company, geography (state, city), a school, a dorm, a social group, a family connection, friend connection, favorite team, or any other suitable attribute. In other cases, embodiments may allow groups to be formed based on an invitation scheme.
  • Shared Devices
  • In some cases a wearable fitness monitor may be shared amongst multiple users. For example, a gym, household, school, work place, or any other community may provide a wearable fitness monitor to members of that community. The use of the wearable fitness monitor may be time-sliced among the member. Rather than requiring each user initiating a session with the wearable fitness monitor, the classification logic may authenticate the wearer of the classification logic and, once authenticate, correlate tracked activity data from the wearer's use of the wearable fitness monitor with the digital account of the wearer.
  • Motion Signatures and Features
  • In certain embodiments, the wearer of a fitness monitor is identified as the user by obtaining one or more motion signatures from the fitness monitor and determining whether such signature(s) are produced by the user's movement. Individual humans have movement characteristics which individually or collectively be used as a biometric identifier for the user. Such motion signatures are described further herein. These signatures and their comparison to reference features linked to the user may allow the wearable fitness monitor to authenticate or otherwise identify the wearer of the wearable fitness monitor. A motion signature may be obtained from a motion sensor such as one of the sensor types mentioned above. It may be viewed as a form of biometric information that can be collected or presented at any time to authenticate or otherwise identify the wearer of the wearable fitness monitor. The wearer of a fitness monitor is identified as the user when the motion signature (or the motion signature in combination with other wearer information) matches, to a degree required by an appropriate user classification procedure, the information in one or more reference features.
  • Examples of Motion Signatures
  • Many types of motion signatures may be obtained via data from motion sensors and may be used by the user identification logic. Some motion signatures may represent motion in the time domain (e.g., amplitude, power, intensity, phase, field strength, etc., each as a function of time). Others represent motion in the frequency domain (e.g., amplitude, power, intensity, phase, field strength, etc., each as a function of frequency). Other types of motion signatures are provided only in association with certain activity types such as running, walking, swimming, bicycling, rowing, weightlifting, etc. For example, curl repetition count is a motion signature that is associated with weightlifting but not running or bicycling.
  • Some other types of motion signature employ a single cycle which may be characterized by its “profile,” which may have either time or frequency as an independent variable. In use, the cycle motion signature is compared against a reference cycle feature, with the comparison matching features of the cycle such as total magnitude and/or duration (e.g., peak-to-peak), maximum or minimum magnitude, maximum positive or negative slope, the number and relative locations of inflection points, envelope, etc. The comparison can be performed by many different techniques including pattern recognition algorithms or classification algorithms known to those of skill in the art.
  • Other examples of motion signatures include step count, step rate, cadence, a metric derived from an amplitude of a periodic motion, a biking cadence, a rowing rate, a resistance-based repetition rate, a typing speed, a zero crossing rate, a peak-to-peak time, an arm swing rate, or a combination thereof.
  • As explained, motion signatures are obtained from data taken from motion sensors. The data may be processed little or substantially to obtain motion signatures. Further, the motion sensor data used to obtain the motion signatures may be the raw “absolute” signal, or may be obtained after filtering (by, e.g., bandpass filtering, low pass filtering), scaling, and the like. In some examples, the 2-norm of the 3-axis accelerometer motion signal may be used in lieu of or in combination with the (x, y, z) signal features to provide information used in a motion signature.
  • Motion signatures representing repetitive movement (not just a cycle) may be obtained using, e.g., data processed to obtain peak counts, zero crossings, spectral information from, e.g., a FFT, and the like. A Fourier series decomposition may be performed to extract the contributions of multiple periodic motions to sensor signal. These contributions may be harmonics associated with, e.g., steps and arm motion. Each harmonic may be a motion signature and/or the ratio of the harmonics' powers may be a motion signature. For example, it has been found that a user's step impact has a big effect on the power observed in higher harmonics. It is to be appreciated that although some embodiments have been described in the context of harmonics, other embodiments can operate on any contribution of a spectral component.
  • It should be understood that the wearing configuration of a fitness monitor affects the resulting motion signature. For example, a single user motion may produce one motion signature when obtained using a fitness monitor clipped to the user's hip and a second motion signature, different from the first motion signature, when obtained using a fitness monitor worn on the user's wrist. The motion signature analysis may therefore account for the type and worn location of the fitness monitor.
  • Examples of Reference Motion Features
  • As explained, a user may be recognized by comparing a user's reference motion feature(s) to a motion signature measured by a wearable fitness monitor. A few examples of reference motion features will now be provided. FIG. 3 shows a representative 3-axis accelerometer signal from a wearable fitness monitor worn on the wrist of a user who is walking. Using an accelerometer, motion signatures such as the step rate (e.g., steps/min) and signal energy may typify a user. For example, for a user who is walking, the accelerometer signal energy increases with step rate. A user may be identified, distinguished from another user, or determined to be a fake by comparing the pair of motion signatures (step/min, signal energy) from the motion signal provided by the wearable fitness monitor to data characterizing the user's walk (e.g., features from signals previously supplied by the user or an entity responsible for enrolling the user and characterizing the user's walk). In various embodiments, trusted information characterizing movement typical of the user is referred to as a reference motion feature for the user. FIG. 4 depicts the step rate and signal energy for 4 unique users. The curve may be approximated as a line and a fitness monitor or the wearer of the fitness monitor may be classified by the nearest line to the data from the wearable fitness monitor (i.e., to one or more motion signatures). If the data does not lie within a reasonable limit to the line, the data may be considered fake (e.g., in the sense of a linear discriminant). A line is used in this example, but any model may be employed such as an arbitrary polynomial, lookup table, etc. Also, the classifier employed may be a neural network, support vector machine, random forest, decision tree, or other machine learning or heuristic algorithm.
  • Similarly, the step rate and signal energy during a run may identify the user, distinguish the user from another, or determined to be a fake. In embodiments where there is no enrollment for a user, default curves or lookup tables for these quantities may be employed. FIG. 5 depicts the step rate and signal energy for 3 unique users for a run.
  • In other embodiments, one user is distinguished from another by the spectral characteristics of the motion data. For example, where the spectral characteristic is a harmonic, the ratios of the second to first harmonic, third to first harmonic, and third to second harmonic observed in the 2-norm of a 3-axis accelerometer may be used. Higher harmonic content corresponds to more impact in walking. Similarly, the approach can be used for running. FIG. 6 depicts two unique users who are otherwise indistinguishable in running by step rate and signal energy, but are clearly distinguished by motion harmonics. Harmonic features may also be used in combination with the aforementioned walking/running features.
  • In other embodiments, the accelerometer motion signal is split into “cycles” (e.g., periods between two steps), aligned, time warped, and used to construct a composite or typical step profile for a user by which features such as the (x, y, z) axis peak-to-peak heights, envelopes, and peak-to-peak duration may be used to build a model of the user's typical motion. A “cycle-type” motion signature and associated reference motion feature may be used to identify a user in the manner described elsewhere herein. For example, a machine learning algorithm such as an LDA classifier, an artificial neural network, decision tree, support vector machine, or the like may be used to classify the user.
  • The aforementioned examples with an accelerometer are presented for illustrative purposes. In several other embodiments the wearable fitness monitor may have a gyroscope and similar or identical approaches may be employed in lieu of or in combination with an accelerometer. In all mention of signal data and signal processing operations performed, they may be performed on the raw “absolute” signal, or after filtering (e.g., bandpass filter, low pass filter), scaling, and the like.
  • The wearable fitness monitor may store one or more invalid motion features that each individually or collectively characterize motion likely to be faked (e.g., performed by a non-human). To detect fakes, the user's identity may be rejected based on an invalid motion feature that characterizes a detected step rate being too consistent (e.g., not varying by more than 5 steps/min from minute to minute over a 10 minute time window) or the motion being too periodic (e.g., each “cycle” of motion corresponding to a step being nearly identical to the previous), the signal energy being contained to nearly one axis of motion (e.g., through principal component analysis with over a threshold value (e.g., 50%) of the 2-norm of a 3-axis accelerometer signal being comprised of one motion axis), the duration of motion being too long (e.g., over 1 hour with no breaks), the motion observed on an accelerometer lacking high harmonic structure (e.g., being too smooth or sinusoidal, no clear presence of integer harmonics from the fundamental step frequency), or too erratic (e.g., 50% or more in peak-to-peak amplitude on an accelerometer over several “step” cycles). The wearable fitness monitor may store one or more invalid motion features that each individually or collectively characterize motion likely to be performed by a non-human.
  • Accordingly, a feature may thus be viewed broadly as data or logic that defines an expected signature of a user, either in the positive (e.g., reference motion signatures) or in the negative (e.g., invalid motion features). Such features may be expressed in a data driven manner (such as through a line, curve, graph, data point), functionally (e.g., logic that defines an expected threshold around a given motion signature), or some combination thereof. And illustrative example of a functional expression of a feature is an invalid feature that specifies that motion signatures that do not exceed a minimal variance are to result in deauthenticating the wearer, as may be the case where a mechanical device may be causing the motion data detected by the wearable fitness monitor. Thus, to execute this invalid feature, the classification logic may analyze the motion signature to determine whether the cycles represented by the motion signature vary beyond a threshold.
  • Location and Proximity Signatures and Features
  • In embodiments where the wearable fitness monitor comprises a location sensor (e.g., GPS, WiFi interface) or is in communication with a device that has a location sensor (e.g., a smartphone), location may be used to determine that the wearable is not with the intended user. For example, if the activity data provided by the wearable fitness monitor does not spend significant periods of time at the user's registered home and/or workplace (e.g., at least 8 hours at home), then the activity data may be classified as fake in relation to a cash incentive program or competitive challenge. In another embodiment, if the wearable fitness monitor is linked to the user's smartphone (e.g., via an account on an app running on the phone) but is not in proximity of the phone or does not “sync” data with the user's phone for a period of time (e.g., 1 week), then the activity data provided by the wearable fitness monitor may be rejected as fake for the purposes of a cash incentive program or competitive challenge.
  • In another embodiment, the wearable fitness monitor comprises a wireless communication system such as Bluetooth, Bluetooth Low Energy, Near Field Communication, Wifi, Body Area Network (e.g., communication routed through the body), and the like. In a manner similar to using a location sensor (e.g., GPS), the device or a system incorporating the device (e.g., a mobile phone and the wearable fitness monitor or a cloud-based server and the wearable fitness monitor) may reject data from the device as a fake or de-authenticate the user based on an inference of the user's location from communication over the wireless communication system. For example, the names of enrolled or commonly observed Wifi networks in the user's typical areas of movement (e.g., home, office, coffee shops, city/town) may be used to determine if the wearable fitness monitor is in a foreign environment, which may trigger a de-authentication. In such case, the device may prompt the user to re-authenticate by verifying a code provided by a trusted device (e.g., a mobile phone in which authentication is achieved via fingerprint or PIN code) or any of the authentication methods described in this disclosure. Prompting may occur by a vibration or haptic interaction from the wearable fitness monitor. Prompting may occur at the next moment in which the wearable fitness monitor's user interface is engaged (e.g., upon pressing a button or providing a motion gesture such as moving a wrist wearable toward the face). Notably, in the present embodiments, it is not necessary that the system infer the geographical location (e.g., address, latitude and longitude) of the user based on the communication data—it is sufficient to maintain a list of networks and/or devices that are in communication with the user.
  • In an embodiment, the wearable fitness monitor includes a Body Area Network communication system. When the user wears the wearable fitness monitor (e.g., on a wrist, clipped to a bra, on clothing), the device may transmit data through the user's body to another device that is in close proximity or in contact with the user. The user may touch a door knob that likewise includes a Body Area Network communication system and the door knob may unlock in response to the user's touch. Similarly, the user may touch or come in close proximity (e.g., less than 1 cm) to an automobile door handle, secured door in an office, etc., and the lock may disengage and/or open in response. The same touch and/or proximity may engage “syncing” of the wearable fitness monitor in the sense of transmitting activity, sleep, and other biometric data to the device that is in proximity and, thereafter, a cloud-based service (e.g., www.fitbit.com). For example, a user may have a Body Area Network enabled lamp at home and touching the lamp may set the color and/or intensity of the light based on the user's preferences (including time of day) and also initiate the transmission of the wearable fitness monitor's activity data to the lamp (if so enabled to receive this data) or other communication hub (e.g., a computer in the user's residence).
  • Heartbeat Waveform Signatures and Features
  • In one embodiment, a user may authenticate to a wearable fitness monitor through features of a heart rate sensor, such as an electrocardiogram (ECG).
  • In an embodiment where the wearable is a bracelet, the device may contain a two electrode system where one electrode (lead 1) is in contact with the wrist and an electrode on the outer surface of the bracelet (lead 2) that is accessible by the opposite hand. When the opposite arm/hand makes contact with the bracelet (lead 2), a conductive path across the user's torso is created and an ECG signal can be collected for multiple heart beat cycles or for some duration of time (for example 10 heart beats or 10 seconds).
  • The ECG waveform collected is then post-processed into a signature waveform. This may involve techniques such as overlapping the periodic components of the waveform (e.g., individual PQRST sections), alignment, normalization, outlier removal, and linear/nonlinear filtering, outputting a composite or typical PQRST waveform signature.
  • FIG. 7 shows examples of data obtained from post-processing. Panel (a) of FIG. 7 shows the raw ECG waveform acquired from the wearable fitness monitor. Panel (b) of FIG. 7 shows the ECG waveform after filtering. Panel (c) of FIG. 7 shows overlapping multiple repeating PQRST waveforms. Panel (d) of FIG. 7 shows the final ECG signature waveform used for feature extraction.
  • Signatures are then extracted from the final ECG signature waveform. These signatures may include pathological characteristics of the PQRST wave (shown in FIG. 8 below) use in cardiac monitoring such as the time and magnitude features of the PR interval, PR segment, QRS complex, ST segment, and QT interval. Signatures may also include other time domain (shown in FIG. 9 below) such as the slopes, areas, curvatures, and polynomial fits that may not directly have any direct physical or medical significance, but present uniqueness useful for authentication. Frequency domain characteristics can also be used as features such as Fourier and cosine transforms of the PQRST.
  • Authentication is determined through comparing the features of a reference ECG vs. that of a the ECG signal detected from the current wearer. Techniques for authentications could involve a combination of techniques such as neural network, support vector machine, logistic regression, naive Bayes, random forest, decision tree, or other machine learning or heuristic algorithms. After authentication is successful, the wearer can remain authenticated until the bracelet is determined to be off wrist, such as due to device removal or loss of contact.
  • In an embodiment where the wearable is a shirt, the device may provide continuous ECG measurements of the user. Similar to the embodiments described above, the ECG may be split into “beats”, aligned, and used to construct a composite PQRST waveform over a moving window, which is then used to extract features to compare against a template for the user. In this embodiment, authentication may be continuous and the user may have a trust score that degrades if the features do not match for a period of time longer than the moving window. If the trust score goes below a designated threshold for a designated period of time, the user may be de-authenticated from the device. Additional features in this embodiment are that the presence of a live user may be determined both by a continuous (or nearly continuous) heart rate signal and that the user is wearing the shirt continuously. Using clothing to house a wearable fitness monitor is described in US Patent Application Publication No. 2010/0292599, which is incorporated herein by reference in its entirety.
  • In another embodiment, the wearable fitness monitor may include another type of heart rate sensor, such as a photoplethysmograph (PPG) sensor, for instance, to monitor the heart rate of the user when it is worn against the wrist. FIGS. 10A-10D below depict representative features of the PPG waveform (and its first derivative) that may be used to identify the user. The shape of the PPG waveform (e.g., in the sense of a template), which is correlated to the age of the user, may also be used as a feature after appropriate temporal and spatial normalization.
  • In addition, because a PPG is a light-based sensor and because many wavelengths of light are absorbed by skin pigmentation, a PPG sensor may be used to characterize the level of skin pigmentation for the wearer of the device based on a response of the PPG sensor. This may be used as an additional feature in a user authentication system. For example, if the intensity level of the PPG's light emitter is high but the return signal to the PPG's light detector is low, then the user has higher skin pigmentation. The ratio of the return signal to the emitter output may be used as a feature to characterize a user. Similarly, testing with different light intensities and/or wavelengths may provide a transfer function (or table lookup) for the user that may be used to identify the user. The wearable PPG sensor may have multiple wavelength LEDs that can be set to different intensities and/or multiple wavelength (e.g., spectral response) photodiodes to do this characterization.
  • In a related embodiment, the PPG sensor may be used to determine if the wearable fitness monitor is being worn. That is, if the PPG response is not representative of human skin (e.g., the return signal is low relative to the emitted output because there is nothing against the sensor), then the device can determine that it is not being worn, at least not on the wrist. This may be determined in combination with a motion sensor (e.g., that the device is stationary on a surface). Similarly, the absence of a heart rate signal in the PPG data, or lack of heart rate variability (e.g., the duration between heart beats is too consistent), may be used to determine that activity data from the device is faked. Moderate to vigorous activity (e.g., walking or running) in the presence of low heart rate (e.g., below 60 bpm for walking and 90 bpm for running) in the PPG data may likewise indicate a fake.
  • In an embodiment, the wearable fitness monitor contains a motion sensor and an ECG. The user authenticates to the device with the ECG and the authentication may be lost if the motion signature is not representative of the registered user. In another embodiment, the wearable fitness monitor comprises a contact sensor (e.g., in the clasp, a PPG, or capacitive sensor against the wrist of the user) that detects when the sensor is removed from the user's wrist and then authentication is lost.
  • In an embodiment, the wearable fitness monitor contains a motion sensor and a PPG. Activity data from the device may be rejected as fake if the motion signature or the PPG-derived data do not correspond to the registered user or to human activity.
  • In an embodiment, the wearable fitness monitor contains an ECG and PPG. The user authenticates not only by matching the ECG morphological signatures and PPG morphological signatures with those previously enrolled by the user (e.g., the reference features), but also in comparing the heart rate and heart rate variability of the two signals to each other.
  • In another embodiment, the wearer of the device may be determined to not be the authorized user of the device based on the heart rate exertion of the user observed during an exercise. For example, the user may walk at a moderate pace and if the heart rate divided by the pace (as observed, say, by GPS) or step cadence is significantly higher or lower than is characteristic for the authorized user, the current user of the device is determined to not be the authorized user. In other embodiments where the wearable fitness monitor automatically tracks exercises such as elliptical, bicycling, and the like, equivalent metrics of heart rate per unit activity (e.g., elliptical strokes, bicycling pace) may be used to compare the current wearer of the device to the authorized user of the device.
  • Worn Detection
  • In some embodiments, optical monitors are used in the wearable monitor, implementing different modes of operation by emitting pulses of light and detecting light after it interacts with the user's skin or other tissue, to thereby capture data that may be used to obtain the user's heartbeat waveform, worn state, user characteristics, etc. In various embodiments, the optical monitor is used as a heartbeat waveform monitor, and while much of the following description refers to such monitors as heartbeat waveform monitors, such monitors need not be configured or designed to measure heartbeat waveforms. It is sufficient that the monitor emit and detect pulses and interpret the pulsing information to accomplish the described results.
  • In some embodiments, the current disclosure provides methods for operating a wearable fitness monitoring device having a heart rate monitor (HRM) in a low power state when the device determines that the device is not worn by a user, or is “off-wrist” when implemented in a wrist-worn device. This feature of the HRM is also referred to as an “automatic off” function. In some embodiments, the automatic off function is implemented by operating the HRM in an “unworn” (or “off-wrist”) detection mode, and the automatic off function automatically turns off the heart rate monitoring operations of the HRM to conserve energy if the device determines that it is not being worn by the user. Other benefits of the automatic off function include providing more accurate heart rate estimation. For example, when an automatic off or automatic on (described below) is performed a heart rate detection algorithm may reset. In one implementation, the algorithm stops running when off-wrist is detected, and restarts when on-wrist is detected. When the heart rate monitor restarts, it resets.
  • In some embodiments, the current disclosure provides methods for operating a wearable fitness monitoring device having a heart rate monitor in a normal power state when the device is worn by the user, or “on-wrist” when implemented in a wrist-worn device. This feature of the HRM is also referred to as an “automatic on” function. In some embodiments, the automatic on function is implemented by operating the HRM in a “worn” (or “on-wrist”) detection mode. The automatic on function automatically takes the HRM out of a low power state and turns on the heart rate monitoring operations of the HRM if the device detects motion and determines that it is worn by the user.
  • In some embodiments, the unworn (or off-wrist) and worn (or on-wrist) detection may be implemented by light (e.g., LED) probing, which emits light pulses and detects signals after the light pulses interact with the user's skin and tissues. In some embodiments, the unworn and worn probing may share some hardware, firmware, software, and/or parameters for light emission, light detection, and analyses of detected signals. In other embodiments, the two probing modes employ different hardware, firmware, software, and/or parameters for light emission, light detection, and analyses may be used for unworn and worn detection.
  • In some embodiments, the wearable fitness monitoring device goes in and out of the low power state regulated by a probe light (e.g., LED) and a motion detector, implementing automatic off and on functions. In the low power state, the heart rate monitor saves power by turning off, or scaling back operation of, its LED light source and its photodetector. In some embodiments, other light sources and light detectors (e.g., photodiodes, photomultiplier tubes, CCD, or CMOS) may be used to implement the automatic off and on functions.
  • Some embodiments provide a method of operating a heart rate monitor of a wearable fitness monitoring device having a plurality of sensors. The method includes: (a) operating the heart rate monitor in a first mode while also operating in a second mode configured to detect near proximity of the wearable fitness monitoring device to a user's skin, where the first mode is configured to determine one or more characteristics of a user's heartbeat waveform when the wearable fitness monitoring device is in near proximity to the user; (b) from information collected in the second mode, determining that the heart rate monitor is not proximate to the user's skin; and (c) in response to determining that the heart rate monitor is not proximate to the user's skin, ending operating the heart rate monitor in the first mode. In some embodiments, the one or more characteristics of the user's heartbeat waveform include the user's heart rate.
  • In some embodiments, the wearable fitness monitor includes a motion sensor, and the method further involving: prior to (c), determining from information output by the motion detecting sensor that the wearable fitness monitoring device has had been still for at least a defined period; and in response to detecting that the wearable fitness monitoring device has had been still for at least the defined period, performing (c). In some embodiments, prior to (a) while the first mode is not operating, the device (i) detects motion of the wearable fitness monitoring device using a motion detecting sensor and/or detecting proximity of the heart rate monitor to the user′ skin by operating the heart rate monitor in a third mode; and (ii) initiates operation of the first mode of the heart rate monitor when the wearable fitness monitoring device is determined to be in near proximity to the user. Further implementations of operating heart rate monitors are provided in U.S. Pat. No. 8,948,832 (Attorney Reference: FTBTP002X1GUS), titled WEARABLE HEART RATE MONITOR, filed on May 30, 2014, which is incorporated by reference in its entirety.
  • Other User Signatures and Features
  • In yet other embodiments, the wearable fitness monitor has a bioimpedance sensor (possibly sharing the same electrodes as the ECG) and the bioimpedance of the user is further used with the ECG and/or PPG to determine the current user of the wearable fitness monitor.
  • In yet another embodiment, the wearable fitness monitor has a fingerprint sensor (e.g., capacitive, ultrasound) that images the pattern of skin ridges on the finger(s) of a user to authenticate the user to a device (e.g., when the device is put on, when the device is used as a proxy for a credit card). The device may include an ECG, PPG, and/or bioimpedance sensor to further enhance the authentication system with user-specific biometric data. The device may maintain authentication through motion signatures of the user and PPG-based signatures of the user. Removal of the device (e.g., as detected by a capacitive sensor mounted on the back of a wrist wearable fitness monitor, as detected optically using an optical sensor (perhaps the same as a PPG sensor or an independent optical sensor), as detected by a sensor in the clasp of a wrist wearable fitness monitor) may de-authenticate the wearer of the device.
  • In another embodiment, the authorized user of the wearable fitness monitor may also have a smartphone that tracks user activity such as walking. If the smartphone walking activity does not match the time and approximate count of the wearable fitness monitor walking activity data for more than a threshold percentage (e.g., 50% of time over some period), then the data of the wearable fitness monitor is considered fake. If the wearable fitness monitor is used as a proxy for a credit card or other secure service, the authorization of the user is deactivated. Reauthentication may be established by displaying a code or image on the wearable fitness monitor and entering the same on the user's mobile phone. Likewise, if the mobile phone is used as a proxy for a credit card or other secure service, the authorization of it may be deactivated until reauthentication by matching a code or image to the wearable is performed. If the wearable fitness monitor comprises a location sensor (e.g., GPS), it may activate the sensor and broadcast its location when it next synchronizes to a cloud-based service.
  • In another embodiment, the wearable fitness monitor tracks the sleep of the user based on sensor data generated by the wearable fitness monitor. For example, the wearable may use motion data and/or heart rate data to infer when a user is asleep. Other examples include the use of skin temperature, galvanic skin response, SpO2, blood pressure, and time of day in combination with or in lieu of the preceding data to infer when the user is asleep. If the user is determined to be asleep by the wearable fitness monitor (or a system in communication with the wearable fitness monitor such as a mobile phone or cloud-based service), the authentication may be “blacked out” so that the wearable fitness monitor cannot be used to access secure digital accounts, open secure areas, etc. Upon waking, the user may retain authentication.
  • Skin Calibration
  • Skin color may be used to define one or more features of a user. In addition, skin color may affect heart rate measurements. This section discloses techniques for measuring skin color as a user feature and using skin measurements to improve measurements of other features such as heart rate.
  • Ambient light and skin color may make it difficult to extract a user's heart rate from a PPG signal. The effect of ambient light may be reduced by subtracting a value of the received detected light signal when the PPG light source is off from the value of the received detected light signal when the PPG light source is on (assuming that both signals are obtained in close temporal proximity to each other). The effect of skin color may be reduced by changing the intensity of the PPG light source, the wavelength of the light emitted from the light source, and/or by using the ratio or difference of received signal corresponding to two different wavelengths. Skin color may be determined by using user input (e.g. the user entering their skin color), an image of the person's face, etc., and may then subsequently be used to calibrate the algorithm, light source brightness, light source wavelength, and the receiver gain. The effect of skin color (and tightness with which the user is wearing the device) on the raw PPG signal may also be measured by sending in a signal of known amplitude to the light source(s) and then measuring the received signal from the photodetector(s). Such a signal may be sent for a prolonged period of time (so as to capture data through multiple expected heart beats) and then averaged to produce a steady-state data set that is not heart-rate dependent. This amplitude may then be compared to a set of values stored in a table to determine algorithm calibration, transmitter amplitude and the receiver gain.
  • In some embodiments, the disclosure provides methods and devices to accurately measure heartbeat waveform for different user characteristics, such as skin colors, motion, sweat, position, and physiologic state (e.g., skin thickness, body fat, etc.) of the users. Because darker skin has lower reflectance of light, the relations between photodetector reading and light pulse intensity, e.g., DAC, tends to have a lower slope than for paler skin. In some embodiments, the signals for skin characterization may operate intermittently at higher frequency than the light pulses of the first mode for heart rate monitoring.
  • Some embodiments provide a method for adjusting at least one setting for operating a heart rate monitor in a wearable fitness monitoring device. The method involves: (a) pulsing a light source in the heart monitor in a skin characterization mode by emitting a succession of light pulses, at least some having variable intensity with respect to one another; (b) detecting a variation in intensity of light from the light pulses emitted in the skin characterization mode after the light has interacted with the user's skin; (c) determining a response characteristic of the user's skin from the variation in intensity of light detected in (b); and (d) using the response characteristic of the user's skin to adjust a gain and/or light emission intensity of the heart rate monitor operating in a first mode for detecting one or more characteristics of the user's heartbeat waveform.
  • In some embodiments, the response characteristic is dependent on an opacity value of the user's skin. In some embodiments, operating in the first mode and operating in the skin characterization mode are performed concurrently. In some embodiments, operating in the first mode and operating in the skin characterization mode concurrently involves periodically determining a response characteristic of the user's skin while continuously operating in the first mode.
  • In some embodiments, operating in the first mode involves pulsing the light source in the heart rate monitor at a first frequency and detecting light from the light source, after the light has interacted with the user's skin, at the first frequency. Furthermore, operating in the skin characterization mode involves pulsing the light source in the heart rate monitor at a second frequency and detecting light from the light source at the second frequency.
  • In some embodiments, as described above, the wearable fitness monitor may store a reference feature relating to the response of a sensor to the skin of a user to later verify the user based on a current response of the sensor to the skin of the wearer of the wearable fitness monitor using, for example, the aforementioned techniques.
  • Comparing User Signatures to User Reference Features
  • Various classification and identification techniques may be applied to compare motion and/or other signatures to user reference features. Generally, such techniques determine whether or not it is likely that the motion signature obtained from a fitness monitor was created by a user in question wearing the device. In this way, a wearer of the fitness monitor can be authenticated or otherwise identified.
  • The logic used to compare a signature to a reference feature may be a classifier or other routine implemented on the fitness monitor and/or a secondary device as described elsewhere herein. As examples, the classifier employed may be an LDA classifier, neural network, support vector machine, random forest, decision tree, or other machine learning or heuristic algorithm.
  • As mentioned, the motion signal may be split into “cycles” (e.g., periods between two steps), aligned, time warped, and used to construct a composite or typical step profile for a user by which features such as the (x, y, z) axis peak-to-peak heights, envelopes, and peak-to-peak duration may be used to build a model of the user's typical motion. As with other types of signatures, this may be used by a machine learning algorithm such as a neural network, decision tree, support vector machine, and the like to classify the user.
  • In certain embodiments, the classification logic compares signatures to reference features and applies a level of confidence (provided directly or indirectly by the comparison algorithm) for authenticating or otherwise identifying the user. The confidence level for identifying the user may be set as appropriate for the application (insurance versus fitness credits in casual competition). The level may also be set for the type of classification algorithm used to compare the wearer's signature(s) to the user's reference feature.
  • FIG. 11 shows a flowchart of a method for determining whether an identity of an instant wearer of a fitness monitor matches that of a user. If an instant wearer's identity matches that of the user, the instant wearer is authenticated as the user. For example, the user may be an owner or an authorized user of the wearable fitness monitor. Because the motion features of the user are provided as a reference against which a wearer's data are compared, the user is also referred to as the reference user herein.
  • In the implementations shown in FIG. 11, the operations of process 1100 are performed by a single wearable fitness monitor. In other implementations, some of the operations can be performed by the wearable fitness monitor, while others operations can be performed by an external device associated with the wearable fitness monitor such as smart phone, a personal computer, a tablet, or a webserver that is associated or communicatively linked to the wearable fitness monitor.
  • Process 1100 involves obtaining a reference motion feature of the reference user (or user as used elsewhere herein) using one or more motion sensors on the wearable fitness monitor worn by the user. See block 1102. The motion sensors may be selected from accelerometers, gyroscopes, GPS sensors, and other motion sensors described herein elsewhere. In some implementations, the reference motion feature comprises a motion cycle profile as described above and hereinafter at block 1206 of FIG. 12.
  • At block 1104, process 1100 involves training a classifier using the reference motion feature. In some implementations, the classifier is a binary classifier. In some examples, the classifier is a linear discriminant (LDA) classifier. Although an LDA classifier is provided as an example below, in various implementations, other classifiers described herein or known in the field may be used instead of or in combination with an LDA. For instance, clustering methods, neural networks, support vector machines, linear and nonlinear models, decision trees, etc., may be used as classifiers to classify the wearer. In some implementations, as described above for the LDA classifier, test data are classified as two classes. In some implementations, the classifier may determine three or more classes. In such implementations, the classifier may be implemented to authenticate three or more users. In some implementations, a C-class LDA classifier may be used, where C is not fewer than three.
  • A binary LDA provides a method to classify data in a multidimensional space into two classes: a target class and a non-target. Data points from each of the two classes are provided to train the classifier. The LDA projects the data points into a new space that best separates the two classes of the data points. More specifically, the projected data points have an optimal combination of mean difference and class variance, with the largest difference in means of the two classes and the smallest variance within each class. The LDA determines a hyperplane that separate the two classes of data. The projections of data points from the same class are very close to each other and at the same time the projected means of the two classes are as far apart from each other as possible. After the LDA classifier has been trained, the classifier is applied to a test vector. The test vector belongs to the target class if the test vector is located on the same side of the hyperplane as the target class, and the location of the hyper plane is defined by a threshold value.
  • The process 1100 further involves obtaining a motion signature from motion data of an instant wearer, using the one or more motion sensors of the wearable fitness monitor. See block 1106. The process 1100 shown in FIG. 11 uses the same motion sensors and the same wearable fitness monitor to obtain both the reference motion feature and the motion signature. However, in some implementations, the reference motion feature may be provided by sensors or wearable fitness monitors different from those producing the motion signature. For instance, the reference motion feature may be imported from another wearable fitness monitor or computer, and then stored on the instant wearable fitness monitor or an instant computer performing one or more operations of process 1100. Then the imported reference motion feature may be compared with the motion signature on the instant wearable fitness monitor or the instant computing device.
  • Operations 1102 and 1104 represent a different phase of the process from operations 1106-1110, with 1102 and 1104 performed initially and the result used repeatedly in 1106-10. Training the classifier is qualitatively different from using the classifier. Of course, 1102 and 1104 can be performed more than once to update the classifier as described in the example below. But typically a single trained classifier can be used repeatedly to determine identity.
  • In some implementations, the process involves obtaining two or more motion signatures from motion data of an instant wearer. In some implementations, the two or more motion signatures comprise two or more motion cycle profiles. In some implementations, the two more motion signatures comprise two different motion features, such as step rate and motion signal power.
  • Process 1100 involves comparing the motion signature to the reference motion signature by applying the classifier (e.g., a LDA classifier) to the motion signature. See block 1108. Although an LDA classifier is provided as an example, other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA. In some implementations, a feature vector is extracted from the reference motion feature, which feature vector is then provided as a data point belonging to the target class to train the LDA classifier. The LDA classifier may also be trained by additional data points belonging to the target class and additional data points belonging to the nontarget class. When more data points are provided to train the LDA classifier, the confidence of classification may be improved.
  • In some implementations, the motion signature is analyzed to extract a feature vector, which can then be tested using the LDA classifier to determine whether the motion signature from the instant wearer matches the reference motion feature from the reference user. The LDA classifier takes a feature vector as an input and provides a target or a non-target classification as an output.
  • If the feature vector extracted from the motion signature of the instant wearer is classified as the target class, it means that the motion signature matches the reference motion feature. Therefore, the process can determine that the identity of the wearer of the fitness monitor if the reference user. See block 1110.
  • In some implementations, when two or more motion signatures are obtained, they can be combined into one value or function, which can then be compared to the reference motion feature. In some implementations, two motion cycle profiles may be averaged, and then the average profile may be compared to a reference cycle profile. In some implementations, values of two motion signatures can form a function (e.g., power as a function of step rate), which can then be compared to a reference function. In some implementations, it is possible to classify each of the motion signatures and combine the classification results, e.g., in a probabilistic framework such as Naïve Bayes. Such a probabilistic combination is also known as Bayesian Fusion. Other probabilistic approaches instead of or in addition to Naïve Bayes (e.g., a mixture model of multiple probability functions) may also be used to combine multiple motion signatures.
  • FIG. 12 shows an implementation of a process 1200 for training a classifier and using the classifier to authenticate a wearer based on the classification result. Although an LDA classifier is provided as an example, other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA. Operation 1102 of obtaining a reference motion feature of the user may be implemented according to operations 1202, 1204, and 1206 of FIG. 12. Operation 1104 of training a linear discriminant analysis classifier may be implemented as the operations 1208 and 1112 of FIG. 12. In some implementations, operation 1106 of obtaining a motion signature may be implemented as operations 1202, 1204, and 1206. The operation 1108 of comparing the motion signature to the reference motion feature may be implemented as operation 1208 and the 1214 in FIG. 12. Operation 1110 of determining whether the motion signature matches the reference motion feature may be implemented as operation 1216 of FIG. 12.
  • Process 1200 starts operating one or more motion sensors of a wearable fitness monitor to generate motion data. See block 1202. In some implementations, the one or more motion sensors are selected from accelerometers, gyroscopes, magnetometers, GPS sensors, etc. In some implementations, the motion data include, for instance, about 2 minutes of data sampled at 25 Hz. In some implementations, motion data include about 5 minutes of data. In some implementations, reference data may be obtained when the wearable fitness monitor is worn by a reference user in a training phase. Then in a testing phase, data provided by an instant wearer are compared to the reference data. When the instant wearer's data match the reference data, the wearer's identity is determined to match the identity of the user, thereby providing authentication to the wearer.
  • For example, the owner of the wearable fitness monitor is a reference user in this context. Training data are collected from the owner of the wearable fitness monitor. In a testing phase, the wearable fitness monitor may determine whether an instant wearer of the wearable fitness monitor has the same identity as the owner, thereby authenticating the wearer as the reference user.
  • In some implementations, the wearable fitness monitor may experience motion caused by walking or running at various speeds. In some implementations, data from different speeds or speed ranges are used to train and generate different classifiers. Although an LDA classifier is provided as an example, other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA. The different classifiers will be applied for data associated with different speeds. In some implementations, data generated by motion at different speeds or speed ranges may be used to train and generate a single LDA classifier. In some implementations, data are normalized on the time dimension to obtain a single cycle profile of the motion data. In some implementations, movement speed may be provided as a feature of a feature vector, and the feature vector being provided to train the LDA classifier or test the LDA classifier.
  • In some imitations, motion data includes data of one, two or three motion sensors. In some imitations, each motion sensor includes three axes. In some implementations, motion data from one axis of a motion sensor are used. Such implementations can provide effective classification and efficient analysis when signal is sufficiently strong.
  • Process 1200 involves preprocessing motion data generated by the motion sensors using various techniques. See block 1204. In some implementations, one or more of the preprocessing techniques described herein are optional. In some implementations, raw motion sensor data are low-pass filtered to smooth the data. In some implementations, data may be smoothed by a rolling time window. In some implementations, local minima are obtained from the smoothed data. FIG. 13 shows an example of acceleration data as a function of time. As shown in the example, the filtering and smoothing provide are more regular and cyclic data. The smoothed data can then be segmented into stride profiles or cycle profiles. See block 1206 of FIG. 12. FIG. 14 shows motion data depicting multiple stride profiles from a same subject. In some implementations, data of multiple stride profiles are normalized on the time dimension. In some implementations, a profile of a mean stride is calculated from multiple strides. In some implementations, outlier profiles that deviate from the mean stride over a criterion are removed. In some implementations, a new mean stride is obtained from stride profiles having the outliers removed. In some implementations, outlier removal and averaging are performed for additional iterations to further improved representativeness of the obtained mean stride profile.
  • In some implementations, one cycle profile of a mean stride is obtained at operation 1206. In some implementations, two or more cycle profiles may be obtained from the motion data to provide training data to train the classifier. In some plantations, cycle profiles from multiple users are obtained, which are then used to train the classifier. At least one cycle profile belongs to the target class of the reference user, and at least one cycle profile belongs to the nontarget class other than the reference user. FIG. 15 shows eight mean stride cycle profiles for eight different subjects.
  • Process 1200 further involves extracting at least one feature vector from the at least one cycle profile. See block 1208. In some implementations, signal values based on amplitude may be used to extract features such as slopes, minima, maxima, accelerations, relative distance between features, etc. In some implementations, features may be based on moments, cumulants, time frequency domain functions, etc. In some implementations, additional features may be added to the feature vector to train or test the classifier. In some implementations, the additional features include motion data not reflected by the cycle profile. For instance, speed of motion may be used as an additional feature. In other implementations, other biometric data may be used to provide additional features of the feature vector, which can then be used to train the classifier or be tested using the classifier. For instance, heart rate, blood pressure, respiration rate, skin color, and other metrics described herein may be included as additional features.
  • Process 1200 further involves determining whether to train the classifier or to apply the classifier to authenticate a wearer. See block 1210. An LDA classifier is an example of a suitable classifier. Other classifiers described herein or known in the field may be used in some implementations instead of or in combination with an LDA. When the process determines to train the LDA classifier, the motion data is configured to be obtained from the reference user. The reference user's identity is the reference identity, against which an instance wearer's identity is compared to. If the instant wearer's identity matches that of the reference user, the instant wearer is authenticated as the reference user.
  • If the process proceeds to train the LDA classifier, feature vectors are used to train the classifier. See block 1212. The at least one cycle profile obtained in 1206, along non-target data, is used to train the LDA classifier. The at least one cycle profile is equivalent to the reference motion feature in blocks 1104, 1108, and 1110 of FIG. 11. In various implementations, at least one feature vector of a target cycle profile and at least one feature vector from a non-target cycle profile are obtained to train the LDA classifier. The target cycle profile is obtained from the reference user. The non-target profile can be obtained from a person other than the reference user. In some implementations, the non-target profile may be obtained from the reference user performing a motion that is different from a target motion. In various applications, numerous feature vectors are obtained from the reference user to train the LDA classifier. In some implementations, the multiple feature vectors obtained from the reference user provide data points of the target class. In some implementations, one or more feature vectors obtained from motion data from individuals other than the reference user are also provided as data points of the nontarget class to train the LDA classifier.
  • In some implementations process 1200 then determines whether to continue to train the LDA classifier with additional data. See block 1222. If the decision is positive, process 1200 returns to operation 1202 to generate more motion data using the one or more motion sensors of the wearable fitness monitor. The training LDA classifier operations described above are repeated.
  • In some implementations, operation 1210 decides to apply the LDA classifier to determine whether the instant wearer is the reference user. In such a case, the at least one cycle profile obtained in block 1206 is equivalent to the motion signature in blocks 1106, 1108, and 1110 in FIG. 11. The process uses the LDA classifier by applying it to the at least one feature vector. See block 1214. In effect, this operation compares the feature vector obtained from the motion data of the instant wearer to that obtained from the reference user. The LDA classifier takes the feature vector as an input and provides an output of a classification of whether or not the data belongs to the target class, i.e., the reference user. If the data is classified as belonging to the reference user, the wearable fitness monitor provides authentication to the wearer, determining that the wearer has the same identity as the reference user. See block 1218. Otherwise, the process does not provide authentication to the wearer, or requires the wearer to provide an alternative method to authenticate the wearer's identity. For instance, the wearer may be required to provide a fingerprint, password, retina scan, a heart rate measurement, or other biometric data for authenticating the wearer. In some implementations, the wearer may be required to provide more motion data to repeat the authentication operations described above.
  • In some implementations, process 1200 proceeds to determine whether to continue to train or use the LDA classifier. If the decision is positive, the process loops back to operation 1202 to generate more motion data to train or use the LDA classifier.
  • In some implementations, decision 1222 determines to continue to train the LDA classifier. Such an implementation may provide continuous learning of the classifier using data reflecting long-term change of the reference user's motion. In some implementations, the motion change may be due to physiological or environmental changes associated with the user. For instance, a user may be injured and as a result develop a different motion cycle profile. In some instances, the user may have different characteristics of motion at different times of a day. In some instances, a same user may have different cycle profile during different kinds of activities. The continue training of the LDA classifier may improve the classifier's ability to account for these different factors. In some implementations, the continual training of the LDA classifier can improve the confidence of the classification. For instance, two or more classifiers may be generated for the same user depending on certain factors, such as time of the day or activity types. In some implementations, values of additional factors may be included into the feature vector used to train the LDA classifier or other types of classifier.
  • In some implementations, decision 1222 operates to continue to apply the LDA classifier to multiple sets of cycle profiles. In such implementations, each set of cycle profiles may be classified by the classifier. Then multiple classification results can be combined by probabilistic methods to obtain a final classification result. The multiple classification results may be combined by Bayesian Fusion using Naïve Bayes. See block 1223 outlined by dashed lines indicating an optional operation. The final classification result can then be used to determine if a user should be authenticated. In such an implementation, the authentication of 1218 is modified to be contingent on the final classification result meeting an authentication criterion.
  • Finally, process 1200 can determine not to continue to train or use the LDA classifier, and the process comes to an end at block 1224.
  • FIG. 16 shows example classification results for two-minute walking data (fs=25 Hz) for 8 subjects. The data set is randomly divided into training and testing sets. The graph shows the performance of the classifier on these 8 subjects.
  • Systems
  • In one embodiment, the wearable fitness monitor is one component of a system that comprises a secondary device capable of communicating with the wearable fitness monitor. In some implementations, the secondary device may be a smart phone, a PDA, a tablet, or a computer. In some implementations, the secondary device may have a shape and mechanical and/or magnetic interface to accept the wearable fitness monitor for safe keeping, communication, and/or charging. Notably, the communication between the wearable fitness monitor and the secondary device may be provided through wireless communication techniques/methods and protocols mentioned elsewhere herein. In some implementations, a secondary device performs the biometric matching between a wearer's motion signature and a user's reference feature.
  • In some implementations, the secondary device may comprise sensors to assist in biometric or environmental monitoring such as, for example, sensors that measure ambient light, noise and/or sound (e.g., to detect snoring), temperature, humidity, and air quality (pollen, dust, CO2, etc.). In one embodiment, the secondary device may communicate with an external service such as www.fitbit.com or server (e.g., personal computer). Communication may be achieved through wired or wireless circuitry and protocols to transfer data to and/or from the secondary device. As examples, any of the wireless technologies described above for the fitness monitor may be used. The secondary device may also act as a relay to transfer data to and/or from the wearable fitness monitor to an external service such as www.fitbit.com or other service (e.g., news, social network updates, email, calendar notifications). Calculation of the user's fitness data may be executed on one or both devices or an external service (e.g., a cloud server) using data from one or both devices.
  • In some implementations, one or more of the operations performed to identify a user are performed on a secondary device. It should be understood that some or all of the operations may be performed on the wearable fitness monitor. Often, the operations are divided between the wearable fitness monitor and the secondary device.
  • As mentioned above, the wearable fitness monitor may be used as a proxy to authorize a credit card or other secure service. Moreover, data of the wearable fitness monitor for an authorized user may be compared to the user's data of a smart phone. The comparison results may be used to authorize a user. Based on the data from the wearable fitness monitor and/or the smart phone, a user may be authorized, deauthorized, and or reauthorized. In some implementations, the authorization, deauthorization, and reauthorization may be performed “online” on the wearable fitness monitor. In such cases, the classifier logic may execute on the wearable fitness monitor. In other implementations, the authorization, deauthorization, and reauthorization may be performed on the smart phone. In such cases, the classifier logic may execute on the smart phone. In further implementations, the authorization, deauthorization, and reauthorization may be performed “offline” on a back end server over a network. In such cases, the classification logic may be implemented on the back-end servers. In yet other implementations, the authorization, deauthorization, and reauthorization may be performed using the wearable fitness monitor, the smart phone, and the back end server. In such cases, portions of the classification logic is split among those devices/systems.
  • The techniques and functions outlined above may be implemented in a wearable fitness monitor as machine-readable instruction sets, either as software stored in memory, as application-specific integrated circuits, field-programmable gate-arrays, or other mechanisms for providing system control. Such instruction sets may be provided to a processor or processors of a wearable fitness monitor to cause the processor or processors to control other aspects of the wearable fitness monitor to provide the functionality described above.
  • Unless the context (where the term “context” is used per its typical, general definition) of this disclosure clearly requires otherwise, throughout the description and the claims, the words “comprise,” “comprising,” and the like are to be construed in an inclusive sense as opposed to an exclusive or exhaustive sense; that is to say, in a sense of “including, but not limited to.” Words using the singular or plural number also generally include the plural or singular number respectively. Additionally, the words “herein,” “hereunder,” “above,” “below,” and words of similar import refer to this application as a whole and not to any particular portions of this application. When the word “or” is used in reference to a list of two or more items, that word covers all of the following interpretations of the word: any of the items in the list, all of the items in the list, and any combination of the items in the list. The term “implementation” refers to implementations of techniques and methods described herein, as well as to physical objects that embody the structures and/or incorporate the techniques and/or methods described herein.
  • There are many concepts and implementations described and illustrated herein. While certain features, attributes and advantages of the implementations discussed herein have been described and illustrated, it should be understood that many others, as well as different and/or similar implementations, features, attributes and advantages of the present inventions, are apparent from the description and illustrations. As such, the above implementations are merely exemplary. They are not intended to be exhaustive or to limit the disclosure to the precise forms, techniques, materials and/or configurations disclosed. Many modifications and variations are possible in light of this disclosure. It is to be understood that other implementations may be utilized and operational changes may be made without departing from the scope of the present disclosure. As such, the scope of the disclosure is not limited solely to the description above because the description of the above implementations has been presented for the purposes of illustration and description.
  • Notably, the present disclosure is neither limited to any single aspect nor implementation, nor to any single combination and/or permutation of such aspects and/or implementations. Moreover, each of the aspects of the present disclosure, and/or implementations thereof, may be employed alone or in combination with one or more of the other aspects and/or implementations thereof. For the sake of brevity, many of those permutations and combinations will not be discussed and/or illustrated separately herein.
  • None of the pending claims include limitations presented in “means plus function” or “step plus function” form. (See, 35 USC §112(f)). It is Applicant's intent that none of the claim limitations be interpreted under or in accordance with 35 U.S.C. §112(f).
  • The present disclosure may be embodied in other specific forms without departing from its spirit or essential characteristics. The described embodiments are to be considered in all respects only as illustrative and not restrictive. The scope of the disclosure is, therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims (43)

1. A method, implemented on a wearable fitness monitor comprising one or more first motion sensors and one or more processors, the method comprising:
(a) obtaining, by the one or more processors, a first motion signature obtained using data from the one or more first motion sensors of the wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor;
(b) obtaining, by the one or more processors, a second motion signature obtained using data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor;
(c) comparing, by the one or more processors, the first and second motion signatures or a combination thereof to a reference motion feature for a user; and
(d) determining, based on the comparison in (c) and by the one or more processors, whether an identity of a wearer of the fitness monitor is the user.
2. The method of claim 1, wherein the one or more first motion sensors comprise an accelerometer, a gyroscope, a magnetometer, an altimeter, a GPS receiver, or any combination thereof.
3. The method of claim 1, wherein the one or more second motion sensors comprise one or more of the first motion sensors.
4. The method of claim 1, wherein the first motion signature is a time domain representation of a periodic motion of the movement experienced by the wearable fitness monitor.
5. (canceled)
6. The method of claim 1, wherein the first motion signature comprises a frequency domain representation of a periodic motion of the movement experienced by the wearable fitness monitor.
7. The method of claim 6, wherein the frequency domain representation of the periodic motion of the movement experienced by the wearable fitness monitor comprises a spectral component in the periodic motion.
8-13. (canceled)
14. The method of claim 1, the user's reference motion feature comprises a profile of a step by the user.
15. The method of claim 1, wherein (c) comprises comparing a combination of the first and second motion signatures to the user's reference motion feature, and wherein the user's reference motion feature comprises a line, a curve, or a look up table relating the first and second motion signatures for the user.
16. (canceled)
17. The method of claim 1, wherein (c) comprises applying a linear discriminant analysis classifier to the first and second motion signatures or the combination thereof.
18. The method of claim 1, wherein (c) comprises determining that at least one of the first and second motion signatures is an invalid motion for a human wearer.
19-22. (canceled)
23. The method of claim 1, wherein at least one of the first and second motion signatures comprises a cycle profile of a periodic motion performed by the user, and wherein the reference motion feature is a predetermined typical cycle for the user's periodic motion.
24. The method of claim 23, wherein the cycle profile comprises a time varying amplitude of an output from the one or more first motion sensors.
25. The method of claim 23, wherein the user's periodic motion is selected from the group consisting of running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, and riding an animal.
26. The method of claim 1, further comprising repeating (a)-(d) at multiple times.
27. (canceled)
28. The method of claim 1, further comprising, responsive to determining that the identity of the wearer of the fitness monitor is not the user, preventing the wearable fitness monitor from allowing a transaction.
29. The method of claim 28, wherein the transaction comprises accessing a secure item or providing the user with an award for meeting an activity threshold.
30. The method of claim 1, further comprising, responsive to determining that the identity of the wearer of the fitness monitor is not the user, requiring the user to authenticate himself or herself.
31. (canceled)
32. The method of claim 1, further comprising, responsive to determining that the identity of the wearer of the fitness monitor is not the user, discrediting a fitness metric obtained for the user via the wearable fitness monitor.
33. The method of claim 1, further comprising, responsive to determining that the identity of the wearer of the fitness monitor is the user, crediting a fitness metric obtained for the user via the wearable fitness monitor.
34. The method of claim 1, further comprising, responsive to determining that the identity of the wearer of the fitness monitor is the user, allowing the wearable fitness monitor to facilitate a transaction.
35-52. (canceled)
53. A system comprising:
a wearable fitness monitor configured to be worn by a person and comprising:
one or more first motion sensors,
one or more second motion sensors, and
a communication interface configured for communicating data from the one or more first motion sensors to a device external to the wearable fitness monitor; and
classification logic configured to:
(a) obtain a first motion signature obtained using data from one or more first motion sensors of a wearable fitness monitor configured to be worn by a person, wherein the first motion signature characterizes a movement experienced by the wearable fitness monitor;
(b) obtain a second motion signature obtained using data from one or more second motion sensors, wherein the second motion signature further characterizes the movement experienced by the wearable fitness monitor;
(c) compare the first and second motion signatures or a combination thereof to a reference motion feature for a user; and
(d) based on the comparison in (c), determine whether an identity of a wearer of the fitness monitor is the user.
54. The system of claim 53, wherein the one or more first motion sensors comprise an accelerometer, a gyroscope, a magnetometer, an altimeter, a GPS receiver, or any combination thereof.
55-68. (canceled)
69. The system of claim 53, wherein (c) comprises applying a linear discriminant analysis classifier to the first and second motion signatures or the combination thereof.
70. The system of claim 53, wherein (c) comprises determining that at least one of the first and second motion signatures is unnatural for a human user.
71-73. (canceled)
74. The system of claim 53, wherein the movement experienced by the wearable fitness monitor is caused by a periodic motion selected from the group consisting of running, walking, cycling, swimming, weight lifting, climbing, rowing, a gymnastic exercise, dancing, an aerobic exercise, a yoga routine, golfing, swinging a club, swinging a racquet, striking a ball or another object, swimming, diving, surfing, skating, skiing, skate boarding, exercising on a machine, driving a vehicle, and riding an animal.
75-76. (canceled)
77. The system of claim 53, the classification logic is further configured to prevent, responsive to determining that the identity of the wearer of the fitness monitor is not the user, the wearable fitness monitor from allowing a transaction.
78. (canceled)
79. The system of claim 53, the classification logic is further configured to require, responsive to determining that the identity of the wearer of the fitness monitor is not the user, the user to authenticate himself or herself.
80. (canceled)
81. The system of claim 53, the classification logic is further configured to discredit, responsive to determining that the identity of the wearer of the fitness monitor is not the user, a fitness metric obtained for the user via the wearable fitness monitor.
82. The system of claim 53, classification logic is further configured to credit, responsive to determining that the identity of the wearer of the fitness monitor is the user, a fitness metric obtained for the user via the wearable fitness monitor.
83. The system of claim 53, the classification logic is further configured to allow, responsive to determining that the identity of the wearer of the fitness monitor is the user, the wearable fitness monitor to facilitate a transaction.
84-100. (canceled)
US15/231,620 2015-08-07 2016-08-08 User identification via collected sensor data from a wearable fitness monitor Abandoned US20170038848A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/231,620 US20170038848A1 (en) 2015-08-07 2016-08-08 User identification via collected sensor data from a wearable fitness monitor

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201562202773P 2015-08-07 2015-08-07
US15/231,620 US20170038848A1 (en) 2015-08-07 2016-08-08 User identification via collected sensor data from a wearable fitness monitor

Publications (1)

Publication Number Publication Date
US20170038848A1 true US20170038848A1 (en) 2017-02-09

Family

ID=58052472

Family Applications (8)

Application Number Title Priority Date Filing Date
US15/231,627 Active US9693711B2 (en) 2015-08-07 2016-08-08 User identification via motion and heartbeat waveform data
US15/231,641 Abandoned US20170039358A1 (en) 2015-08-07 2016-08-08 Transaction prevention using fitness data
US15/231,636 Abandoned US20170035328A1 (en) 2015-08-07 2016-08-08 User identification via data collected from sensors of a wearable fitness monitor
US15/231,620 Abandoned US20170038848A1 (en) 2015-08-07 2016-08-08 User identification via collected sensor data from a wearable fitness monitor
US15/603,305 Active US9851808B2 (en) 2015-08-07 2017-05-23 User identification via motion and heartbeat waveform data
US15/812,879 Active US10126830B2 (en) 2015-08-07 2017-11-14 User identification via motion and heartbeat waveform data
US16/153,618 Active US10503268B2 (en) 2015-08-07 2018-10-05 User identification via motion and heartbeat waveform data
US16/663,166 Active US10942579B2 (en) 2015-08-07 2019-10-24 User identification via motion and heartbeat waveform data

Family Applications Before (3)

Application Number Title Priority Date Filing Date
US15/231,627 Active US9693711B2 (en) 2015-08-07 2016-08-08 User identification via motion and heartbeat waveform data
US15/231,641 Abandoned US20170039358A1 (en) 2015-08-07 2016-08-08 Transaction prevention using fitness data
US15/231,636 Abandoned US20170035328A1 (en) 2015-08-07 2016-08-08 User identification via data collected from sensors of a wearable fitness monitor

Family Applications After (4)

Application Number Title Priority Date Filing Date
US15/603,305 Active US9851808B2 (en) 2015-08-07 2017-05-23 User identification via motion and heartbeat waveform data
US15/812,879 Active US10126830B2 (en) 2015-08-07 2017-11-14 User identification via motion and heartbeat waveform data
US16/153,618 Active US10503268B2 (en) 2015-08-07 2018-10-05 User identification via motion and heartbeat waveform data
US16/663,166 Active US10942579B2 (en) 2015-08-07 2019-10-24 User identification via motion and heartbeat waveform data

Country Status (2)

Country Link
US (8) US9693711B2 (en)
CN (1) CN106445101A (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9693711B2 (en) 2015-08-07 2017-07-04 Fitbit, Inc. User identification via motion and heartbeat waveform data
US20180225437A1 (en) * 2017-02-03 2018-08-09 Samsung Electronics Co., Ltd Electronic device for authenticating biometric data and system
IT201700041203A1 (en) * 2017-04-13 2018-10-13 S A T E Systems And Advanced Tech Engineering S R L METHOD OF BIOMETRIC IDENTIFICATION OF AN INDIVIDUAL THROUGH THE ANALYSIS OF A PHYSIOLOGICAL SIGNAL
EP3476672A1 (en) * 2017-10-27 2019-05-01 Continental Automotive GmbH Mobile transmitter unit and method for controlling a function of a technical system using such mobile transmitter unit
WO2019081975A1 (en) * 2017-10-24 2019-05-02 Orcam Technologies Ltd. Wearable camera systems and methods for authenticating identity
EP3525122A1 (en) * 2018-02-13 2019-08-14 Koninklijke Philips N.V. Using ultrasound for identification
EP3584743A1 (en) * 2018-06-18 2019-12-25 Telefonica S.A. Improved biometric user's authentication
US10624561B2 (en) 2017-04-12 2020-04-21 Fitbit, Inc. User identification by biometric monitoring device
US10713874B2 (en) * 2018-08-24 2020-07-14 TruU, Inc. Machine learning-based platform for user identification
WO2020254291A1 (en) * 2019-06-21 2020-12-24 Novartis Ag Systems and methods for user verification based on actigraphy data
US20210065014A1 (en) * 2019-08-29 2021-03-04 Nihon Kohden Corporation Subject discriminating device, subject discriminating method, and non-transitory computer-readable medium
EP3824809A1 (en) * 2019-11-19 2021-05-26 Luxshare-ICT Co., Ltd. Animal physical activity tracking device and system having the same
WO2021190907A1 (en) * 2020-03-27 2021-09-30 Sony Group Corporation Method and system for providing an electronic credential associated with electronic identification information
WO2023061565A1 (en) * 2021-10-13 2023-04-20 Huawei Technologies Co., Ltd. Apparatus, system and method for determining whether a person is asleep
US20230143628A1 (en) * 2021-11-08 2023-05-11 Penumbra, Inc. Systems and methods of classifying movements for virtual reality activities
WO2023156762A1 (en) * 2022-02-21 2023-08-24 Prevayl Innovations Limited Method and system for verifying an activity metric
US11903680B2 (en) 2015-06-14 2024-02-20 Facense Ltd. Wearable-based health state verification for physical access authorization
WO2024039511A1 (en) * 2022-08-19 2024-02-22 Cardiac Pacemakers, Inc. Managing an operating mode of a wearable health monitor

Families Citing this family (129)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9811818B1 (en) * 2002-10-01 2017-11-07 World Award Academy, World Award Foundation, Amobilepay, Inc. Wearable personal digital device for facilitating mobile device payments and personal use
US9704154B2 (en) * 2002-10-01 2017-07-11 World Award Academy, World Award Foundation, Amobilepay, Inc. Wearable personal digital device for facilitating mobile device payments and personal use
US8989837B2 (en) 2009-12-01 2015-03-24 Kyma Medical Technologies Ltd. Methods and systems for determining fluid content of tissue
EP3063832B1 (en) 2013-10-29 2022-07-06 Zoll Medical Israel Ltd. Antenna systems and devices and methods of manufacture thereof
US10561321B2 (en) * 2013-12-12 2020-02-18 Alivecor, Inc. Continuous monitoring of a user's health with a mobile device
US11013420B2 (en) 2014-02-05 2021-05-25 Zoll Medical Israel Ltd. Systems, apparatuses and methods for determining blood pressure
JP2016047155A (en) * 2014-08-27 2016-04-07 セイコーエプソン株式会社 Biological information measurement device
US9942222B1 (en) * 2014-09-02 2018-04-10 Amazon Technologies, Inc. Authentication with wearable device
WO2016040337A1 (en) 2014-09-08 2016-03-17 KYMA Medical Technologies, Inc. Monitoring and diagnostics systems and methods
US10548485B2 (en) 2015-01-12 2020-02-04 Zoll Medical Israel Ltd. Systems, apparatuses and methods for radio frequency-based attachment sensing
KR20180081717A (en) * 2015-09-21 2018-07-17 데스티니 헬스 인코포레이티드 Compensation System and Method
WO2017156577A1 (en) * 2016-03-14 2017-09-21 National Ict Australia Limited Energy harvesting for sensor systems
US10617356B2 (en) 2016-03-15 2020-04-14 Anhui Huami Information Technology Co., Ltd. Garment and cardiac data processing
US10123741B2 (en) * 2016-11-30 2018-11-13 Huami Inc. Cardiac condition detection
US10163282B2 (en) * 2016-03-30 2018-12-25 Intermec, Inc. Systems and methods for authentication
US10827952B2 (en) * 2016-04-29 2020-11-10 Arizona Board Of Regents On Behalf Of Arizona State University Electrocardiographic biometric authentication
US10497191B2 (en) * 2016-08-10 2019-12-03 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10037641B2 (en) * 2016-08-10 2018-07-31 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10424407B2 (en) 2016-08-10 2019-09-24 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US9905063B1 (en) * 2016-08-10 2018-02-27 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10019859B2 (en) * 2016-08-10 2018-07-10 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10013832B2 (en) * 2016-08-10 2018-07-03 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US9779352B1 (en) 2016-08-10 2017-10-03 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10032109B2 (en) * 2016-08-10 2018-07-24 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10593137B2 (en) * 2016-08-10 2020-03-17 Elwha Llc Systems and methods for individual identification and authorization utilizing conformable electronics
US10277283B2 (en) * 2016-09-14 2019-04-30 Sony Corporation NFC device, reader/writer device and methods for authorizing and performing an update
US10361860B2 (en) * 2016-09-16 2019-07-23 Raytheon Bbn Technologies Corp. Method and apparatus for authentication and validation of persons using gait data
DK179471B1 (en) 2016-09-23 2018-11-26 Apple Inc. Image data for enhanced user interactions
US9961547B1 (en) * 2016-09-30 2018-05-01 EMC IP Holding Company LLC Continuous seamless mobile device authentication using a separate electronic wearable apparatus
US10888735B2 (en) * 2016-10-07 2021-01-12 William W. Clark Calibration of initial orientation and position of sports equipment and body segments for inertial sensors
KR102570787B1 (en) * 2016-10-27 2023-08-24 삼성전자주식회사 Recognition device for NFC tag and system including the same
WO2018081795A1 (en) 2016-10-31 2018-05-03 Zipline Medical, Inc. Systems and methods for monitoring physical therapy of the knee and other joints
CN107049324B (en) * 2016-11-23 2019-09-17 深圳大学 A kind of judgment method and device of limb motion posture
US10893415B2 (en) * 2016-11-29 2021-01-12 P&P Ultra G Ltd. Preventing unauthorized use of devices
JP2018093979A (en) * 2016-12-09 2018-06-21 セイコーエプソン株式会社 Exercise diagnostic device, exercise diagnosis system, program, recording medium and exercise diagnosis method
US10140440B1 (en) * 2016-12-13 2018-11-27 Symantec Corporation Systems and methods for securing computing devices that are not in users' physical possessions
CN108781221B (en) * 2017-01-28 2021-10-22 卫保数码有限公司 Apparatus for identifying a person and method thereof
US10880303B2 (en) 2017-03-21 2020-12-29 Global E-Dentity, Inc. Real-time COVID-19 outbreak identification with non-invasive, internal imaging for dual biometric authentication and biometric health monitoring
US10135822B2 (en) * 2017-03-21 2018-11-20 YouaretheID, LLC Biometric authentication of individuals utilizing characteristics of bone and blood vessel structures
CN106874057B (en) * 2017-04-11 2018-06-29 惠州学院 A kind of information input method and its equipment
CN107085231B (en) * 2017-04-24 2020-03-24 济南中景电子科技有限公司 Positioning calling equipment with self-adaptive function and method for detecting wearing mode of positioning calling equipment
CN107411753A (en) * 2017-06-06 2017-12-01 深圳市科迈爱康科技有限公司 A kind of wearable device for identifying gait
DE102017211631A1 (en) * 2017-07-07 2019-01-10 Bundesdruckerei Gmbh Electronic system and method for classifying a physiological condition
JP7056904B2 (en) * 2017-07-26 2022-04-19 ファミリーイナダ株式会社 Massage machine system and massage machines and wearable measuring devices used for it
US10075846B1 (en) * 2017-08-10 2018-09-11 The Florida International University Board Of Trustees Method for continuous user authentication with wearables
WO2019030746A1 (en) 2017-08-10 2019-02-14 Zoll Medical Israel Ltd. Systems, devices and methods for physiological monitoring of patients
US11257573B2 (en) 2017-08-16 2022-02-22 Disney Enterprises, Inc. System for adjusting an audio/visual device based on health and wellness data
GB2566101A (en) * 2017-09-05 2019-03-06 B Secur Ltd Wearable authentication device
EP3685288B1 (en) * 2017-09-21 2022-09-28 Nokia Technologies Oy Apparatus, method and computer program product for biometric recognition
US10776467B2 (en) 2017-09-27 2020-09-15 International Business Machines Corporation Establishing personal identity using real time contextual data
US10839003B2 (en) 2017-09-27 2020-11-17 International Business Machines Corporation Passively managed loyalty program using customer images and behaviors
US10803297B2 (en) 2017-09-27 2020-10-13 International Business Machines Corporation Determining quality of images for user identification
US10795979B2 (en) 2017-09-27 2020-10-06 International Business Machines Corporation Establishing personal identity and user behavior based on identity patterns
CN111542285A (en) 2017-10-13 2020-08-14 皇家飞利浦有限公司 Method and system for characterizing a user of a personal care device
WO2019079776A1 (en) * 2017-10-20 2019-04-25 iNmotion Wellness, Inc. Method and apparatus for providing interactive fitness equipment via a cloud-based networking
US10225737B1 (en) * 2017-10-31 2019-03-05 Konica Minolta Laboratory U.S.A., Inc. Method and system for authenticating a user using a mobile device having plural sensors
WO2019095055A1 (en) * 2017-11-15 2019-05-23 Uti Limited Partnership Method and system utilizing pattern recognition for detecting atypical movements during physical activity
US10878072B2 (en) * 2017-11-20 2020-12-29 Ppip, Llc Systems and methods for biometric identity and authentication
US10565432B2 (en) 2017-11-29 2020-02-18 International Business Machines Corporation Establishing personal identity based on multiple sub-optimal images
CN108280332B (en) * 2017-12-15 2021-08-03 创新先进技术有限公司 Biological characteristic authentication, identification and detection method, device and equipment of mobile terminal
WO2019127088A1 (en) * 2017-12-27 2019-07-04 深圳和而泰数据资源与云技术有限公司 Snore recognition method and snore-stopping device
CN108245869B (en) * 2017-12-29 2020-03-06 北京顺源开华科技有限公司 Swimming information detection method and device and electronic equipment
CA2992333C (en) * 2018-01-19 2020-06-02 Nymi Inc. User access authorization system and method, and physiological user sensor and authentication device therefor
JP7069771B2 (en) * 2018-02-06 2022-05-18 カシオ計算機株式会社 Exercise data display device and exercise data display method
WO2019161857A1 (en) * 2018-02-23 2019-08-29 Sens Innovation Aps Device and method for providing safe data transfer via a non-secure communication protocol
CN108759813A (en) * 2018-04-10 2018-11-06 广东小天才科技有限公司 Three-dimensional motion trajectory description method, device, equipment and storage medium
PL425374A1 (en) * 2018-04-26 2019-11-04 Invis Spolka Z Ograniczona Odpowiedzialnoscia Elastic measuring insert
JP7384154B2 (en) * 2018-05-16 2023-11-21 ソニーグループ株式会社 Information processing device, information processing method, information processing program, terminal device, terminal device control method and control program
CN108920228B (en) * 2018-05-28 2021-01-15 云谷(固安)科技有限公司 Control instruction input method and input device
EP3807795A4 (en) * 2018-06-14 2021-08-11 Facebook Technologies, LLC. User identification and authentication with neuromuscular signatures
CN108875667B (en) * 2018-06-27 2021-03-02 北京字节跳动网络技术有限公司 Target identification method and device, terminal equipment and storage medium
CN109036541B (en) * 2018-07-12 2021-06-25 广东小天才科技有限公司 Wearable device use reminding method and wearable device
US11278238B2 (en) * 2018-09-14 2022-03-22 Warsaw Orthopedic, Inc. Wearable sensor device and analysis platform for objective outcome assessment in spinal diseases
GB2574074B (en) 2018-07-27 2020-05-20 Mclaren Applied Tech Ltd Time synchronisation
CN110494079B (en) * 2018-08-03 2022-09-02 广东高驰运动科技有限公司 Heart rate detection method and device, detection equipment and storage medium
USD899605S1 (en) 2018-09-21 2020-10-20 MedHab, LLC Wrist attachment band for fall detection device
US10380866B1 (en) 2018-09-21 2019-08-13 Med Hab, LLC. Dual case system for fall detection device
USD866957S1 (en) 2018-09-21 2019-11-19 MedHab, LLC Belt clip for fall detection device
US10860096B2 (en) 2018-09-28 2020-12-08 Apple Inc. Device control using gaze information
US20200117780A1 (en) * 2018-10-15 2020-04-16 Ca, Inc. Multi-factor biometric authentication
US11576618B2 (en) 2018-10-16 2023-02-14 Cardionomous Llc Heart signal waveform processing system and method
CA3118245A1 (en) * 2018-10-29 2020-05-07 Intellisports Inc. System and method for real-time activity classification and feedback
CN109718528B (en) * 2018-11-28 2021-06-04 浙江骏炜健电子科技有限责任公司 Identity recognition method and system based on motion characteristic parameters
CN109766751B (en) * 2018-11-28 2022-02-01 西安电子科技大学 Steady-state vision-evoked electroencephalogram identity recognition method and system based on frequency domain coding
CN109480825B (en) * 2018-12-13 2021-08-06 武汉中旗生物医疗电子有限公司 Electrocardio data processing method and device
CN109656137A (en) * 2018-12-19 2019-04-19 南通大学 The daily behavior prediction technique of smart home environment servant
CN109684812B (en) * 2018-12-29 2021-09-03 西安电子科技大学 Continuous identity authentication method for mobile equipment to acquire user motion behaviors
WO2020178724A1 (en) * 2019-03-04 2020-09-10 Lampros Kourtis Method and system to pair an article to a user.
US20220156356A1 (en) * 2019-03-26 2022-05-19 Nec Corporation Biometric-information authentication system, biometricinformation authentication method, authentication device, authentication method, measurement device, measurement method, and computer-readable recording medium having program recorded thereon
CN110151187B (en) * 2019-04-09 2022-07-05 缤刻普达(北京)科技有限责任公司 Body-building action recognition method and device, computer equipment and storage medium
CA3042361A1 (en) * 2019-05-06 2020-11-06 Nymi Inc. Live user authentication device, system and method and fraud or collusion prevention using same
US11252336B2 (en) * 2019-05-31 2022-02-15 Samsung Electronics Co.. Ltd. Electronic device obtaining skin image and method of controlling the same
CN110197172B (en) * 2019-06-10 2021-10-19 清华大学 Method and device for identity authentication based on photoelectric blood vessel volume information
CN110270079B (en) * 2019-06-26 2020-11-06 深圳市悦动天下科技有限公司 Method and device for detecting shake-shake state during movement, intelligent terminal and storage medium
CN110288802A (en) * 2019-06-28 2019-09-27 北京小米移动软件有限公司 A kind of reminding method, device and medium
EP3782547B1 (en) * 2019-08-21 2024-04-10 The Swatch Group Research and Development Ltd Method and system for gait detection of a person
CN110515088B (en) * 2019-08-27 2021-07-09 中国人民解放军国防科技大学 Odometer estimation method and system for intelligent robot
GB2588236B (en) 2019-10-18 2024-03-20 Mclaren Applied Ltd Gyroscope bias estimation
WO2021100197A1 (en) * 2019-11-22 2021-05-27 日本電信電話株式会社 Heartbeat detection device, heartbeat detection method, and heartbeat detection program
CN111124860B (en) * 2019-12-16 2021-04-27 电子科技大学 Method for identifying user by using keyboard and mouse data in uncontrollable environment
EP3838141A1 (en) * 2019-12-17 2021-06-23 Koninklijke Philips N.V. System and method to determine if a device is worn on a subject's dominant limb
US11482047B2 (en) * 2020-01-06 2022-10-25 Kaia Health Software GmbH ML model arrangement and method for evaluating motion patterns
US11106772B2 (en) * 2020-01-31 2021-08-31 Dell Products, Lp System and method for continuous user identification via piezo haptic keyboard and touchpad dynamics
CN111248922B (en) * 2020-02-11 2022-05-17 中国科学院半导体研究所 Human body respiration condition acquisition paste based on accelerometer and gyroscope and preparation method thereof
CN111276208A (en) * 2020-03-06 2020-06-12 中国人民解放军陆军军医大学第一附属医院 Health analysis system based on big data
WO2021188018A2 (en) * 2020-03-16 2021-09-23 Vladimir Savchenko Ecg, heart and lungs sound monitoring system with wireless earphones
EP3890236B1 (en) 2020-03-30 2022-07-20 Tata Consultancy Services Limited Continuously validating a user during an established authenticated session using photoplethysmogram and accelerometer data
CN111481191A (en) * 2020-03-30 2020-08-04 上海赛族网络科技有限公司 Adjusting system based on electrocardio sensor parameter
US11861752B2 (en) 2020-04-06 2024-01-02 Agt International Gmbh Method of scoring a motion of a user and system thereof
GB202006303D0 (en) * 2020-04-29 2020-06-10 Xplora Tech As Method and apparatus for motion data analysis
EP3910646A1 (en) * 2020-05-11 2021-11-17 Fresenius Medical Care Deutschland GmbH Medical treatment system for identifying and authorising medical staff
US11637835B2 (en) * 2020-06-17 2023-04-25 Irdeto B.V. System and method for context-sensitive access control
CN112036572B (en) * 2020-08-28 2024-03-12 上海冰鉴信息科技有限公司 Text list-based user feature extraction method and device
US20220080262A1 (en) * 2020-09-14 2022-03-17 Train121 Inc. Method and apparatus to generate motion data of a barbell and to process the generated motion data
US20220088460A1 (en) * 2020-09-23 2022-03-24 Sensor Maestros, LLC Visual Or Audible Indicators Of Sensed Motion In A Hockey Puck
US20240108233A1 (en) * 2020-12-18 2024-04-04 Nitto Denko Corporation Method and Device for Collecting Physiological Data of a Wearer
WO2022137571A1 (en) * 2020-12-25 2022-06-30 三菱電機株式会社 Authentication device, authentication method, and authentication program
EP4264460A1 (en) * 2021-01-25 2023-10-25 Apple Inc. Implementation of biometric authentication
US12111898B2 (en) * 2021-02-09 2024-10-08 The Trustees Of Princeton University Devices and methods for smartphone impostor detection using behavioral and environmental data
CN113713353B (en) * 2021-05-12 2022-05-31 北京冰锋科技有限责任公司 Method and system for acquiring technical actions of ski-jump skiers
SE2130254A1 (en) * 2021-09-23 2023-03-24 Rths Ab A sensing arrangement for obtaining data from a body part using accurate reference values
GB2613591A (en) * 2021-12-07 2023-06-14 Prevayl Innovations Ltd Electronic module, a controller for an electronics module and a method performed by a controller
CN115062639B (en) * 2022-04-06 2024-09-06 青岛迈金智能科技股份有限公司 Pedal frequency resolving method based on accelerometer
CN114602155B (en) * 2022-05-11 2023-02-21 荣耀终端有限公司 Swimming information statistical method, computer-readable storage medium and electronic device
WO2024145231A1 (en) * 2022-12-30 2024-07-04 Verily Life Sciences Llc Hybrid compute methods for wearable on-body detection utilizing multiple sensing modalities
CN117577266B (en) * 2024-01-15 2024-04-30 南京信息工程大学 Hand rehabilitation training monitoring system based on force touch glove
CN117883076B (en) * 2024-01-23 2024-06-18 北京邦尼营策科技有限公司 Human movement energy consumption monitoring system and method based on big data
CN118078270B (en) * 2024-04-25 2024-07-26 华南理工大学 ECG (electrocardiogram) and PPG (pulse-g) signal-based electrocardiographic motion monitoring system and motion state evaluation method
CN118445788A (en) * 2024-07-04 2024-08-06 支付宝(杭州)信息技术有限公司 Heartbeat signal-based identity verification method, device, medium and equipment

Family Cites Families (138)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8049597B1 (en) * 2000-01-10 2011-11-01 Ensign Holdings, Llc Systems and methods for securely monitoring an individual
US6695207B1 (en) 2000-02-04 2004-02-24 Carroll Boyd Norris, Jr. System for secure, identity authenticated, and immediate financial transactions as well as activation of varied instrumentalities
US7536557B2 (en) * 2001-03-22 2009-05-19 Ensign Holdings Method for biometric authentication through layering biometric traits
US20020013717A1 (en) * 2000-12-28 2002-01-31 Masahiro Ando Exercise body monitor with functions to verify individual policy holder and wear of the same, and a business model for a discounted insurance premium for policy holder wearing the same
US6583369B2 (en) 2001-04-10 2003-06-24 Sunbeam Products, Inc. Scale with a transiently visible display
US6973445B2 (en) 2001-05-31 2005-12-06 Contentguard Holdings, Inc. Demarcated digital content and method for creating and processing demarcated digital works
JP2003051012A (en) * 2001-08-03 2003-02-21 Nec Corp Method and device for authenticating user
US7269737B2 (en) * 2001-09-21 2007-09-11 Pay By Touch Checking Resources, Inc. System and method for biometric authorization for financial transactions
JP3838141B2 (en) * 2002-04-09 2006-10-25 オムロンヘルスケア株式会社 Blood pressure measuring device and exercise equipment
US20060136744A1 (en) * 2002-07-29 2006-06-22 Lange Daniel H Method and apparatus for electro-biometric identity recognition
US7171680B2 (en) * 2002-07-29 2007-01-30 Idesia Ltd. Method and apparatus for electro-biometric identity recognition
US7020508B2 (en) 2002-08-22 2006-03-28 Bodymedia, Inc. Apparatus for detecting human physiological and contextual information
US7606790B2 (en) * 2003-03-03 2009-10-20 Digimarc Corporation Integrating and enhancing searching of media content and biometric databases
IL155955A0 (en) * 2003-05-15 2003-12-23 Widemed Ltd Adaptive prediction of changes of physiological/pathological states using processing of biomedical signal
EP1521161A3 (en) * 2003-09-25 2006-03-15 Matsushita Electric Industrial Co., Ltd. An apparatus and a method for preventing unauthorized use and a device with a function of preventing unauthorized use
US20090131759A1 (en) * 2003-11-04 2009-05-21 Nathaniel Sims Life sign detection and health state assessment system
WO2005083546A1 (en) 2004-02-27 2005-09-09 Simon Richard Daniel Wearable modular interface strap
US8109858B2 (en) * 2004-07-28 2012-02-07 William G Redmann Device and method for exercise prescription, detection of successful performance, and provision of reward therefore
JP2006059282A (en) * 2004-08-24 2006-03-02 Fujitsu Ltd Authenticating method based on biological feature information
US7620819B2 (en) * 2004-10-04 2009-11-17 The Penn State Research Foundation System and method for classifying regions of keystroke density with a neural network
EP1646254B1 (en) * 2004-10-11 2008-04-16 Swisscom Mobile AG Identification and/or authentification method by means of fingerprints
JP4672357B2 (en) * 2004-12-20 2011-04-20 富士フイルム株式会社 Authentication system
JP4469746B2 (en) * 2005-03-29 2010-05-26 株式会社東芝 Heart rate measuring device and method of operating heart rate measuring device
US20070002141A1 (en) * 2005-04-19 2007-01-04 Objectvideo, Inc. Video-based human, non-human, and/or motion verification system and method
US8656458B2 (en) * 2005-08-25 2014-02-18 Guy Heffez Method and system for authenticating internet user identity
JP5395429B2 (en) 2005-06-03 2014-01-22 シナプティクス インコーポレイテッド Method and system for detecting capacitance using sigma delta measurement
US8079079B2 (en) * 2005-06-29 2011-12-13 Microsoft Corporation Multimodal authentication
US7733224B2 (en) * 2006-06-30 2010-06-08 Bao Tran Mesh network personal emergency response appliance
US20070142715A1 (en) * 2005-12-20 2007-06-21 Triage Wireless, Inc. Chest strap for measuring vital signs
US7848811B2 (en) 2005-12-21 2010-12-07 Cardiac Pacemakers, Inc. Posture sensor
US8762733B2 (en) * 2006-01-30 2014-06-24 Adidas Ag System and method for identity confirmation using physiologic biometrics to determine a physiologic fingerprint
US20070260511A1 (en) * 2006-05-03 2007-11-08 Image Products, Inc. System and method for providing rewards including visible indicia thereon to a user to encourage physical activity and to encourage the accurate reporting of physical activity
US7558622B2 (en) 2006-05-24 2009-07-07 Bao Tran Mesh network stroke monitoring appliance
US7844081B2 (en) 2006-05-15 2010-11-30 Battelle Memorial Institute Imaging systems and methods for obtaining and using biometric information
US8684900B2 (en) 2006-05-16 2014-04-01 Bao Tran Health monitoring appliance
US8562526B2 (en) 2006-06-01 2013-10-22 Resmed Sensor Technologies Limited Apparatus, system, and method for monitoring physiological signs
US20080147502A1 (en) * 2006-11-06 2008-06-19 Baker Steve G Exercise incenting methods and devices
US11196811B2 (en) * 2006-12-01 2021-12-07 Fitistics, Llc Data communications between an exercise device and a personal content device
US7742995B2 (en) 2007-03-23 2010-06-22 Mastercard International, Inc. Pre-authenticated identification token
US7884727B2 (en) * 2007-05-24 2011-02-08 Bao Tran Wireless occupancy and day-light sensing
US9757043B2 (en) * 2007-12-06 2017-09-12 Los Angeles Biomedical Research Institute At Harbor-Ucla Medical Center Method and system for detection of respiratory variation in plethysmographic oximetry
US20090241174A1 (en) 2008-02-19 2009-09-24 Guru Rajan Handling Human Detection for Devices Connected Over a Network
US20090232362A1 (en) 2008-03-12 2009-09-17 Hitachi Maxell, Ltd. Biometric information acquisition apparatus and biometric authentication apparatus
CN101533938B (en) 2008-03-14 2012-06-20 鸿富锦精密工业(深圳)有限公司 Low-pass filter
US20090239709A1 (en) * 2008-03-21 2009-09-24 Shen Yi Wu Health management feedback method using fitness equipments
AU2009246442B2 (en) * 2008-05-14 2015-02-12 Heartmiles, Llc. Physical activity monitor and data collection unit
US8996332B2 (en) 2008-06-24 2015-03-31 Dp Technologies, Inc. Program setting adjustments based on activity identification
US8447704B2 (en) * 2008-06-26 2013-05-21 Microsoft Corporation Recognizing gestures from forearm EMG signals
US8615290B2 (en) * 2008-11-05 2013-12-24 Apple Inc. Seamlessly embedded heart rate monitor
US20100203967A1 (en) * 2009-02-06 2010-08-12 Inventec Corporation Exercising game system and method thereof
US8289130B2 (en) * 2009-02-19 2012-10-16 Apple Inc. Systems and methods for identifying unauthorized users of an electronic device
US8655441B2 (en) * 2009-04-16 2014-02-18 Massachusetts Institute Of Technology Methods and apparatus for monitoring patients and delivering therapeutic stimuli
CN103955131B (en) 2009-04-26 2017-04-12 耐克创新有限合伙公司 GPS features and functionality in an athletic watch system
US9141087B2 (en) 2009-04-26 2015-09-22 Nike, Inc. Athletic watch
US8140156B2 (en) 2009-06-30 2012-03-20 Medtronic, Inc. Heart sound sensing to reduce inappropriate tachyarrhythmia therapy
US7924212B2 (en) * 2009-08-10 2011-04-12 Robert Bosch Gmbh Method for human only activity detection based on radar signals
US9174123B2 (en) * 2009-11-09 2015-11-03 Invensense, Inc. Handheld computer systems and techniques for character and command recognition related to human movements
US8922342B1 (en) * 2010-02-15 2014-12-30 Noblis, Inc. Systems, apparatus, and methods for continuous authentication
US8747330B2 (en) * 2010-04-19 2014-06-10 Sotera Wireless, Inc. Body-worn monitor for measuring respiratory rate
TR201903706T4 (en) 2010-04-22 2019-04-22 Koninklijke Philips Nv Skin contact detector.
US9030294B2 (en) * 2010-09-20 2015-05-12 Pulsar Informatics, Inc. Systems and methods for collecting biometrically verified actigraphy data
US9167991B2 (en) * 2010-09-30 2015-10-27 Fitbit, Inc. Portable monitoring devices and methods of operating same
US8849610B2 (en) * 2010-09-30 2014-09-30 Fitbit, Inc. Tracking user physical activity with multiple devices
US9011292B2 (en) 2010-11-01 2015-04-21 Nike, Inc. Wearable device assembly having athletic functionality
US9298886B2 (en) * 2010-11-10 2016-03-29 Nike Inc. Consumer useable testing kit
US8475367B1 (en) 2011-01-09 2013-07-02 Fitbit, Inc. Biometric monitoring device having a body weight sensor, and methods of operating same
US8886578B2 (en) 2011-01-11 2014-11-11 The Royal Institution For The Advancement Of Learning/Mcgill University Method and system for automatically classifying and identifying vestibulo-ocular responses
US9378336B2 (en) * 2011-05-16 2016-06-28 Dacadoo Ag Optical data capture of exercise data in furtherance of a health score computation
US20130002435A1 (en) 2011-06-10 2013-01-03 Aliphcom Sleep management method and apparatus for a wellness application using data from a data-capable band
EP2718080A2 (en) 2011-06-10 2014-04-16 Aliphcom Motion profile templates and movement languages for wearable devices
CA2817145A1 (en) 2011-06-10 2012-12-13 Aliphcom Determinative processes for wearable devices
US20120316455A1 (en) * 2011-06-10 2012-12-13 Aliphcom Wearable device and platform for sensory input
US20140089672A1 (en) * 2012-09-25 2014-03-27 Aliphcom Wearable device and method to generate biometric identifier for authentication using near-field communications
US8682821B2 (en) * 2011-08-08 2014-03-25 Robert Bosch Gmbh Method for detection of movement of a specific type of object or animal based on radar signals
US8660322B2 (en) * 2011-08-25 2014-02-25 King Saud University Passive continuous authentication method
US20130158686A1 (en) * 2011-12-02 2013-06-20 Fitlinxx, Inc. Intelligent activity monitor
US20150119728A1 (en) 2011-12-02 2015-04-30 Fitlinxx, Inc. Health monitor
US9734304B2 (en) * 2011-12-02 2017-08-15 Lumiradx Uk Ltd Versatile sensors with data fusion functionality
DE102011088817A1 (en) 2011-12-16 2013-06-20 Robert Bosch Gmbh Device designed to measure vital parameters of a patient
US9643050B2 (en) * 2011-12-22 2017-05-09 Adidas Ag Fitness activity monitoring systems and methods
US9352207B2 (en) * 2012-01-19 2016-05-31 Nike, Inc. Action detection and activity classification
US20140067494A1 (en) * 2012-01-24 2014-03-06 Blake Squires Activity monitor system and method
US20130204410A1 (en) * 2012-02-03 2013-08-08 Frank Napolitano System and method for promoting and tracking physical activity among a participating group of individuals
KR101971697B1 (en) * 2012-02-24 2019-04-23 삼성전자주식회사 Method and apparatus for authenticating user using hybrid biometrics information in a user device
US9442100B2 (en) 2013-12-18 2016-09-13 Medibotics Llc Caloric intake measuring system using spectroscopic and 3D imaging analysis
US9641239B2 (en) * 2012-06-22 2017-05-02 Fitbit, Inc. Adaptive data transfer using bluetooth
US9044171B2 (en) * 2012-06-22 2015-06-02 Fitbit, Inc. GPS power conservation using environmental data
US8948832B2 (en) * 2012-06-22 2015-02-03 Fitbit, Inc. Wearable heart rate monitor
US8983637B2 (en) * 2012-07-30 2015-03-17 Mapmyfitness, Inc. Determining authenticity of reported fitness-related activities
US20140059066A1 (en) * 2012-08-24 2014-02-27 EmoPulse, Inc. System and method for obtaining and using user physiological and emotional data
US20140085050A1 (en) * 2012-09-25 2014-03-27 Aliphcom Validation of biometric identification used to authenticate identity of a user of wearable sensors
WO2014058894A1 (en) 2012-10-08 2014-04-17 Lark Technologies, Inc. Method for delivering behavior change directives to a user
WO2014088621A1 (en) * 2012-12-03 2014-06-12 Google, Inc. System and method for detecting gestures
US9053308B2 (en) * 2012-12-12 2015-06-09 Intel Corporation Multi electro-biometric user recognition
US20150371023A1 (en) * 2013-03-13 2015-12-24 University Of Pittsburgh - Of The Commonwealth System Of Higher Education Usage modeling
US20140327515A1 (en) * 2013-03-15 2014-11-06 AlipCom Combination speaker and light source responsive to state(s) of an organism based on sensor data
US20140266860A1 (en) * 2013-03-14 2014-09-18 Gaddi BLUMROSEN Method and system for activity detection and classification
US9314666B2 (en) * 2013-03-15 2016-04-19 Ficus Ventures, Inc. System and method for identifying and interpreting repetitive motions
US9782075B2 (en) * 2013-03-15 2017-10-10 I2Dx, Inc. Electronic delivery of information in personalized medicine
CN103445766B (en) * 2013-05-09 2016-01-20 陈飞 The state monitoring method contrasted based on history physiological data and physical data and device
US9606721B2 (en) 2013-07-22 2017-03-28 Lg Electronics Inc. Mobile terminal and control method thereof
KR102127927B1 (en) * 2013-07-30 2020-07-09 엘지전자 주식회사 Mobile terminal, smart watch and security authorizing method therbetween
US9892576B2 (en) * 2013-08-02 2018-02-13 Jpmorgan Chase Bank, N.A. Biometrics identification module and personal wearable electronics network based authentication and transaction processing
US9996803B2 (en) * 2013-09-03 2018-06-12 Arizona Board Of Regents On Behalf Of Arizona State University Systems and methods for authenticating a user through an unobservable re-authentication system
US9898880B2 (en) 2013-09-10 2018-02-20 Intel Corporation Authentication system using wearable device
CA2931973A1 (en) * 2013-11-29 2015-06-04 Motiv Inc. Wearable computing device
US9547972B2 (en) * 2013-12-10 2017-01-17 Sal Castillo Methods and systems for emergency alerts
US20150169854A1 (en) * 2013-12-16 2015-06-18 Iowa State University Research Foundation, Inc. Capturing cognitive fingerprints from keystroke dynamics for active authentication
WO2015126095A1 (en) * 2014-02-21 2015-08-27 삼성전자 주식회사 Electronic device
US20150242605A1 (en) * 2014-02-23 2015-08-27 Qualcomm Incorporated Continuous authentication with a mobile device
WO2015127119A2 (en) 2014-02-24 2015-08-27 Sony Corporation Body position optimization and bio-signal feedback for smart wearable devices
WO2015134908A1 (en) 2014-03-07 2015-09-11 Thalchemy Corporation Learn-by-example systems and methods
US10575760B2 (en) 2014-03-26 2020-03-03 GestureLogic Inc. Systems, methods and devices for activity recognition
US20150305674A1 (en) * 2014-04-02 2015-10-29 Corewave LLC Systems and Methods for Measuring Hydration in a Human Subject
KR102244856B1 (en) * 2014-04-22 2021-04-27 삼성전자 주식회사 Method for providing user interaction with wearable device and wearable device implenenting thereof
CN103955824A (en) * 2014-05-14 2014-07-30 金陵科技学院 High-security wearable collection and payment method
US10602981B2 (en) 2014-05-30 2020-03-31 Microsoft Technology Licensing, Llc Optical pressure sensor
US9414761B2 (en) 2014-06-02 2016-08-16 Indian Institute Of Technology Delhi QRS complex identification in electrocardiogram signals
US9943241B2 (en) 2014-06-12 2018-04-17 PhysioWave, Inc. Impedance measurement devices, systems, and methods
WO2016003269A1 (en) 2014-06-30 2016-01-07 Scint B.V. Body worn measurement device
US9667353B2 (en) * 2014-07-11 2017-05-30 Sony Corporation Methods of providing body area network communications when a user touches a button of a wireless electronic device, and related wireless electronic devices and wearable wireless electronic devices
US9159213B1 (en) * 2014-07-17 2015-10-13 National Chung Shan Institute Of Science And Technology Motion detection method and device
US20170231519A1 (en) * 2014-08-19 2017-08-17 The General Hospital Corporation System and method for annotating and analyzing eeg waveforms
KR102257300B1 (en) * 2014-09-01 2021-05-27 삼성전자주식회사 Method and apparatus for authenticating user using ecg signal
US10114935B2 (en) * 2014-12-23 2018-10-30 Intel Corporation Technologies for login pattern based multi-factor authentication
KR102297193B1 (en) * 2014-12-24 2021-09-02 삼성전자 주식회사 Electronic device having user identifying function and user authentication method
KR102299819B1 (en) * 2014-12-30 2021-09-08 삼성전자주식회사 Method and apparatus for authenticating user using bio signal
US9836896B2 (en) * 2015-02-04 2017-12-05 Proprius Technologies S.A.R.L Keyless access control with neuro and neuro-mechanical fingerprints
US20160313176A1 (en) 2015-04-21 2016-10-27 Salutron, Inc. User-wearable devices including uv light exposure detector with calibration for skin tone
US20160352727A1 (en) * 2015-05-26 2016-12-01 Reticle Ventures Canada Incorporated System and method for asset authentication and management
KR101657005B1 (en) * 2015-06-11 2016-09-12 전문석 Method for electrocardiogram authentication
US9696822B2 (en) * 2015-06-29 2017-07-04 International Business Machines Corporation Rotating ring for device control
US9913591B2 (en) * 2015-07-02 2018-03-13 Verily Life Sciences Llc Wrist-mounted device with integrated electronics
US10318958B2 (en) * 2015-07-28 2019-06-11 Jong Ho Kim Smart watch and operating method using the same
US9693711B2 (en) 2015-08-07 2017-07-04 Fitbit, Inc. User identification via motion and heartbeat waveform data
US10181021B2 (en) 2016-02-01 2019-01-15 Fitbit, Inc. Method and apparatus for off-body detection for wearable device
US10888278B2 (en) 2016-07-20 2021-01-12 Hall Labs, Llc Method of monitoring health while using a toilet
US10624561B2 (en) 2017-04-12 2020-04-21 Fitbit, Inc. User identification by biometric monitoring device

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11903680B2 (en) 2015-06-14 2024-02-20 Facense Ltd. Wearable-based health state verification for physical access authorization
US9851808B2 (en) 2015-08-07 2017-12-26 Fitbit, Inc. User identification via motion and heartbeat waveform data
US10942579B2 (en) 2015-08-07 2021-03-09 Fitbit, Inc. User identification via motion and heartbeat waveform data
US10126830B2 (en) 2015-08-07 2018-11-13 Fitbit, Inc. User identification via motion and heartbeat waveform data
US9693711B2 (en) 2015-08-07 2017-07-04 Fitbit, Inc. User identification via motion and heartbeat waveform data
US10503268B2 (en) 2015-08-07 2019-12-10 Fitbit, Inc. User identification via motion and heartbeat waveform data
US20180225437A1 (en) * 2017-02-03 2018-08-09 Samsung Electronics Co., Ltd Electronic device for authenticating biometric data and system
US10977349B2 (en) * 2017-02-03 2021-04-13 Samsung Electronics Co., Ltd. Electronic device for authenticating biometric data and system
US10624561B2 (en) 2017-04-12 2020-04-21 Fitbit, Inc. User identification by biometric monitoring device
US10806379B2 (en) 2017-04-12 2020-10-20 Fitbit, Inc. User identification by biometric monitoring device
US11382536B2 (en) 2017-04-12 2022-07-12 Fitbit, Inc. User identification by biometric monitoring device
IT201700041203A1 (en) * 2017-04-13 2018-10-13 S A T E Systems And Advanced Tech Engineering S R L METHOD OF BIOMETRIC IDENTIFICATION OF AN INDIVIDUAL THROUGH THE ANALYSIS OF A PHYSIOLOGICAL SIGNAL
US11526590B2 (en) 2017-10-24 2022-12-13 Orcam Technologies Ltd. Automatic low radiation mode for a wearable device
WO2019081975A1 (en) * 2017-10-24 2019-05-02 Orcam Technologies Ltd. Wearable camera systems and methods for authenticating identity
EP3476672A1 (en) * 2017-10-27 2019-05-01 Continental Automotive GmbH Mobile transmitter unit and method for controlling a function of a technical system using such mobile transmitter unit
WO2019158561A1 (en) * 2018-02-13 2019-08-22 Koninklijke Philips N.V. Using ultrasound for identification
EP3525122A1 (en) * 2018-02-13 2019-08-14 Koninklijke Philips N.V. Using ultrasound for identification
EP3584743A1 (en) * 2018-06-18 2019-12-25 Telefonica S.A. Improved biometric user's authentication
US11069165B2 (en) * 2018-08-24 2021-07-20 TruU, Inc. Machine learning-based platform for user identification
US11734977B2 (en) * 2018-08-24 2023-08-22 TruU, Inc. Machine learning-based platform for user identification
US20210312738A1 (en) * 2018-08-24 2021-10-07 TruU, Inc. Machine Learning-Based Platform For User Identification
US10713874B2 (en) * 2018-08-24 2020-07-14 TruU, Inc. Machine learning-based platform for user identification
US11514739B2 (en) * 2018-08-24 2022-11-29 TruU, Inc. Machine learning-based platform for user identification
US20230033188A1 (en) * 2018-08-24 2023-02-02 TruU, Inc. Machine Learning-Based Platform For User Identification
US20230030943A1 (en) * 2018-08-24 2023-02-02 TruU, Inc. Machine Learning-Based Platform For User Identification
US11861947B2 (en) * 2018-08-24 2024-01-02 TruU, Inc. Machine learning-based platform for user identification
US20220245227A1 (en) * 2019-06-21 2022-08-04 Novartis Ag Systems and methods for user verification based on actigraphy data
WO2020254291A1 (en) * 2019-06-21 2020-12-24 Novartis Ag Systems and methods for user verification based on actigraphy data
US20210065014A1 (en) * 2019-08-29 2021-03-04 Nihon Kohden Corporation Subject discriminating device, subject discriminating method, and non-transitory computer-readable medium
EP3824809A1 (en) * 2019-11-19 2021-05-26 Luxshare-ICT Co., Ltd. Animal physical activity tracking device and system having the same
WO2021190907A1 (en) * 2020-03-27 2021-09-30 Sony Group Corporation Method and system for providing an electronic credential associated with electronic identification information
WO2023061565A1 (en) * 2021-10-13 2023-04-20 Huawei Technologies Co., Ltd. Apparatus, system and method for determining whether a person is asleep
US20230143628A1 (en) * 2021-11-08 2023-05-11 Penumbra, Inc. Systems and methods of classifying movements for virtual reality activities
WO2023156762A1 (en) * 2022-02-21 2023-08-24 Prevayl Innovations Limited Method and system for verifying an activity metric
WO2024039511A1 (en) * 2022-08-19 2024-02-22 Cardiac Pacemakers, Inc. Managing an operating mode of a wearable health monitor

Also Published As

Publication number Publication date
US20170035327A1 (en) 2017-02-09
CN106445101A (en) 2017-02-22
US20170039358A1 (en) 2017-02-09
US10503268B2 (en) 2019-12-10
US10942579B2 (en) 2021-03-09
US9851808B2 (en) 2017-12-26
US20190050064A1 (en) 2019-02-14
US20170255273A1 (en) 2017-09-07
US20200167004A1 (en) 2020-05-28
US9693711B2 (en) 2017-07-04
US20180067565A1 (en) 2018-03-08
US20170035328A1 (en) 2017-02-09
US10126830B2 (en) 2018-11-13

Similar Documents

Publication Publication Date Title
US10942579B2 (en) User identification via motion and heartbeat waveform data
US12064260B2 (en) Methods and systems for providing a preferred fitness state of a user
CN112955973B (en) Electronic device for distinguishing mental behavior attributes based on deep neural network
Masood et al. Modeling mental stress using a deep learning framework
US20140089673A1 (en) Biometric identification method and apparatus to authenticate identity of a user of a wearable device that includes sensors
US20140085050A1 (en) Validation of biometric identification used to authenticate identity of a user of wearable sensors
KR20150095661A (en) Method and Device for Determining Vital Parameters
CN110099603A (en) The system and method for carrying out living things feature recognition using sleep physiology
Hill Wearables–the future of biometric technology?
US20220375590A1 (en) Sleep staging algorithm
Maiorana A survey on biometric recognition using wearable devices
US20230214470A1 (en) User authentication by a wearable device
Girish Rao Salanke et al. An enhanced intrinsic biometric in identifying people by photopleythsmography signal
US20240197199A1 (en) Tactile and auditory respiration biofeedback with a wearable device
US11914842B1 (en) Dynamic application icons
Shen et al. PupilHeart: Heart Rate Variability Monitoring via Pupillary Fluctuations on Mobile Devices
US20240094821A1 (en) Techniques for gesture recognition using wearable device data
US20230084205A1 (en) Techniques for menopause and hot flash detection and treatment
AU2022425520A1 (en) User authentication by a wearable device
GB2617258A (en) Method, computer readable medium and system
WO2023156762A1 (en) Method and system for verifying an activity metric
Tu High-Precision and Personalized Wearable Sensing Systems for Healthcare Applications
Eteke et al. Flow From Motion: A Deep Learning Approach
Li Accurate and Robust Heart Rate Sensor Calibration on Smart-watches using Deep Learning
WO2022245594A1 (en) Automatic sleep staging classification with circadian rhythm adjustment

Legal Events

Date Code Title Description
AS Assignment

Owner name: FITBIT, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YUEN, SHELTEN GEE JAO;PARK, JAMES;GHOREYSHI, ATIYEH;SIGNING DATES FROM 20160819 TO 20160822;REEL/FRAME:040068/0537

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION