WO2015006523A1 - Calibration of grab detection - Google Patents
Calibration of grab detection Download PDFInfo
- Publication number
- WO2015006523A1 WO2015006523A1 PCT/US2014/046072 US2014046072W WO2015006523A1 WO 2015006523 A1 WO2015006523 A1 WO 2015006523A1 US 2014046072 W US2014046072 W US 2014046072W WO 2015006523 A1 WO2015006523 A1 WO 2015006523A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- computing device
- sensor data
- determined
- vector
- real
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/169—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated pointing device, e.g. trackball in the palm rest area, mini-joystick integrated between keyboard keys, touch pads or touch stripes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
- G06N20/10—Machine learning using kernel methods, e.g. support vector machines [SVM]
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01D—MEASURING NOT SPECIALLY ADAPTED FOR A SPECIFIC VARIABLE; ARRANGEMENTS FOR MEASURING TWO OR MORE VARIABLES NOT COVERED IN A SINGLE OTHER SUBCLASS; TARIFF METERING APPARATUS; MEASURING OR TESTING NOT OTHERWISE PROVIDED FOR
- G01D18/00—Testing or calibrating apparatus or arrangements provided for in groups G01D1/00 - G01D15/00
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01D—MEASURING NOT SPECIALLY ADAPTED FOR A SPECIFIC VARIABLE; ARRANGEMENTS FOR MEASURING TWO OR MORE VARIABLES NOT COVERED IN A SINGLE OTHER SUBCLASS; TARIFF METERING APPARATUS; MEASURING OR TESTING NOT OTHERWISE PROVIDED FOR
- G01D21/00—Measuring or testing not otherwise provided for
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1694—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a single or a set of motion sensors for pointer control or gesture input obtained by sensing movements of the portable computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3215—Monitoring of peripheral devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3231—Monitoring the presence, absence or movement of users
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/30—Monitoring
- G06F11/3003—Monitoring arrangements specially adapted to the computing system or computing system component being monitored
- G06F11/3013—Monitoring arrangements specially adapted to the computing system or computing system component being monitored where the computing system is an embedded system, i.e. a combination of hardware and software dedicated to perform a certain function in mobile devices, printers, automotive or aircraft systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/30—Monitoring
- G06F11/3058—Monitoring arrangements for monitoring environmental properties or parameters of the computing system or of the computing system component, e.g. monitoring of power, currents, temperature, humidity, position, vibrations
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/30—Monitoring
- G06F11/34—Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
- G06F11/3438—Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment monitoring of user actions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03547—Touch pads, in which fingers can move on a surface
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/038—Control and interface arrangements therefor, e.g. drivers or device-embedded control circuitry
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72448—User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
- H04M1/72454—User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/40—Data acquisition and logging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2200/00—Indexing scheme relating to G06F1/04 - G06F1/32
- G06F2200/16—Indexing scheme relating to G06F1/16 - G06F1/18
- G06F2200/163—Indexing scheme relating to constructional details of the computer
- G06F2200/1637—Sensing arrangement for detection of housing movement or orientation, e.g. for controlling scrolling or cursor movement on the display of an handheld computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/033—Indexing scheme relating to G06F3/033
- G06F2203/0339—Touch strips, e.g. orthogonal touch strips to control cursor movement or scrolling; single touch strip to adjust parameter or to implement a row of soft keys
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/038—Indexing scheme relating to G06F3/038
- G06F2203/0381—Multimodal input, i.e. interface arrangements enabling the user to issue commands by simultaneous use of input devices of different nature, e.g. voice plus gesture on digitizer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2250/00—Details of telephonic subscriber devices
- H04M2250/12—Details of telephonic subscriber devices including a sensor for measuring a physical value, e.g. temperature or motion
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- This disclosure generally relates to mobile computing devices.
- a mobile computing device such as a smartphone, tablet computer, or laptop computer— may include functionality for determining its location, direction, or orientation, such as a GPS receiver, compass, or gyroscope. Such a device may also include functionality for wireless communication, such as BLUETOOTH communication, near-field communication (NFC), or infrared (IR) communication or communication with a wireless local area networks (WLANs) or cellular-telephone network. Such a device may also include one or more cameras, scanners, touchscreens, microphones, or speakers. Mobile computing devices may also execute software applications, such as games, web browsers, or social-networking applications. With social-networking applications, users may connect, communicate, and share information with other users in their social networks.
- wireless communication such as BLUETOOTH communication, near-field communication (NFC), or infrared (IR) communication or communication with a wireless local area networks (WLANs) or cellular-telephone network.
- WLANs wireless local area networks
- Mobile computing devices may also execute software applications, such
- certain user experience improvements may be enabled by inferring user intent from sensor input generated by transitions between physical states with respect to a human hand or other human body part (e.g., approaching the device, making contact with the device, grasping the device, moving the device, releasing the device, moving away from the device).
- detection of such a transition based on sensor input is dependent upon determining an accurate baseline— with respect to the action of a hand making contact with and grasping a device ("grabbing" the device), the baseline for the raw sensor data may vary depending on differences in the size of the user's hand, the orientation of the user's hand, temperature, humidity, etc.
- the transitions are the meaningful aspect, this issue may be addressed by treating the detection space as the derivative of the raw sensor data.
- the data points can be used to generate a set of training data.
- a support vector machine (SVM) model can be generated based on the training data and applied in real time to sensor input to classify detected transitions as a "grab” or as "not-a-grab".
- a mobile device having N touch sensors calculate a derivative of each sensor's output to generate a tuple comprising a vector in N-dimensional space (a support vector).
- Multiple support vectors (across multiple types of physical contexts, for multiple types of users) may be generated, and each support vector may be classified into one of two sets of support vectors (e.g. "grab” or "not-a-grab”).
- a separating hyperplane in the N- dimensional space may be calculated based on the two sets of support vectors.
- the SVM may be applied to map real-time sensor input into the N-dimensional space, calculate the dot product with respect to the hyperplane, and thereby classify the event triggering the sensor input.
- Improvements in the accuracy of detection of such transitions between states may be further correlated using input data from other types of sensors, e.g., (1) motion sensors (e.g., accelerometer(s) or gyroscope(s)), (2) proximity sensor(s) (optical or ambient), (3) pressure sensors (e.g., piezoresistive), (4) temperature sensors, etc. Such correlation may be used to help confirm detection of a "grab.”
- sensors e.g., (1) motion sensors (e.g., accelerometer(s) or gyroscope(s)), (2) proximity sensor(s) (optical or ambient), (3) pressure sensors (e.g., piezoresistive), (4) temperature sensors, etc.
- Such correlation may be used to help confirm detection of a "grab.”
- the communication device can more accurately detect the "grab,” the device may be able to infer that use of the device by the user is imminent, and thus initiate any processes to download and/or upload data in order to bring applications and/or data on the device up to date.
- FIGURE 1 illustrates an example mobile computing device.
- FIGURE 2 illustrates an example sensor configuration of an example mobile computing device.
- FIGURE 3 illustrates an example method for initiating a pre-determined function of a computing device based on an inferred intent of a user.
- FIGURES 4A-B illustrate example detection of a transition in example sensor data.
- FIGURE 5 illustrates an example network environment associated with a social- networking system.
- FIGURE 6 illustrates an example classification of sensor data.
- FIGURE 7 illustrates an example method for determining whether sensor data corresponds to a pre-determined use of a client system.
- FIGURE 8 illustrates an example isolation of components of sensor data through calculation of an example projection.
- FIGURE 9 illustrates example method of isolating a component of sensor data.
- FIGURE 10 illustrates an example computing system.
- FIGURE 1 illustrates an example mobile computing device.
- the client system may be a mobile computing device 10 as described above.
- mobile computing device 10 may be a computing system as described below.
- mobile computing device 10 may be a single- board computer system (SBC) (such as, for example, a computer-on-module (COM) or system- on-module (SOM)), a laptop or notebook computer system, a mobile telephone, a smartphone, a personal digital assistant (PDA), a tablet computer system, or a combination of two or more of these.
- SBC single- board computer system
- PDA personal digital assistant
- mobile computing device 10 may have a primary touch sensor 12 as an input component.
- capacitive touch sensors there may be three types of electrodes: transmitting, receiving, and loading. These electrodes may be connected to a controller designed to drive the transmitting electrodes with electrical pulses.
- touch sensor 12 is incorporated on a front surface of mobile computing device 10.
- one or more secondary touch sensors 14A-D may be incorporated into one or more surfaces of mobile computing device 10.
- one or more secondary touch sensors 14A-D may have coverage over a portion of multiple surfaces of mobile computing device 10, such as for example a portion of a side or bottom surface.
- the intent of the user associated with mobile computing device 10 may be inferred through transitions in sensor data detected by one or more touch sensors 12 and 14A-D or any combination of sensor types.
- Mobile computing device 10 many include a communication component for communicating with an Ethernet or other wire -based network or a wireless NIC (WNIC), wireless adapter for communicating with a wireless network, such as for example a WI-FI network or modem for communicating with a cellular network, such third generation mobile telecommunications (3G), or Long Term Evolution (LTE) network.
- WNIC wireless NIC
- This disclosure contemplates any suitable network and any suitable communication component for it.
- mobile computing device 10 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these.
- PAN personal area network
- LAN local area network
- WAN wide area network
- MAN metropolitan area network
- mobile computing device 10 may communicate with a wireless PAN (WPAN) (such as, for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM), 3G, or LTE network), or other suitable wireless network or a combination of two or more of these.
- WPAN wireless PAN
- GSM Global System for Mobile Communications
- 3G 3G, or LTE network
- Mobile computing device 10 may include any suitable communication component for any of these networks, where appropriate.
- mobile computing device 10 may have multiple operational states. As an example and not by way of limitation, when mobile computing device 10 has not being used by its user for a period of time (e.g.
- mobile computing device 10 may enter into a power-saving state.
- mobile computing device 10 may operate at a lower power level in order to save energy and prolong battery life.
- the display of mobile computing device 10 may become dim or be powered down.
- mobile computing device 10 may be in any suitable operational state, depending on, for example, whether the user is currently using mobile computing device 10, an amount of time that has elapsed since the most recent use of mobile computing device 10, the physical environment of mobile computing device 10 (e.g. in a carrying case, a pocket, or a drawer).
- an application executed by an application processor of mobile computing device 10 may prompt the user to perform specific actions within a predetermined period of time to provide sensor data that may function as training data for a machine learning algorithm, such as for example a support vector machine (SVM), neural network, belief propagation, or k-means algorithm.
- a machine learning algorithm such as for example a support vector machine (SVM), neural network, belief propagation, or k-means algorithm.
- the user may indicate to the application that a particular action is being performed, such as for example riding a bicycle, sitting with mobile computing device 10 in a pocket, or taking mobile computing device 10 from a pocket, and the training application may record sensor data corresponding to the particular action through one or more types of sensors.
- each of the actions may be classified into a particular one of a number of states associated with mobile computing device 10, such as for example, actions associated with making physical contact with mobile computing device 10 or actions not associated with physical contact with mobile computing device 10.
- mobile computing device 10 may send the sensor data as an array of measurement values and a state value corresponding to the particular state associated with each action.
- the training data may be an array of capacitance values from one or more touch sensors of mobile computing device 10.
- the training data may include the acceleration measured by the accelerometer while the particular action is being performed.
- the training data may also include indicator information associating the particular action with a particular state of mobile computing device 10, such as for example physical contact with mobile computing device 10.
- a "0" may be assigned to a state representing resting mobile computing device 10 on a surface, such as for example a table.
- a "1" may assigned to a state representing physical contact being made with mobile computing device 10, such as for example picking up from the table.
- this disclosure describes collecting training data for a particular number of particular states associated with the mobile computing device, this disclosure contemplates collecting training data for any suitable number of states associated with any suitable computing device.
- real-time sensor data may be determined to be an event corresponding to one or more pre-determined intended use of mobile computing device 10 based at least in part on the comparing the real-time sensor data to the training data.
- the training data may be used to classify sensor data into a number of predetermined uses of mobile computing device 10 and define a hyperplane separating sensor data into pre-determined uses of mobile computing device 10.
- parameters defining the hyperplane may be sent to mobile computing device 10 and a processor (e.g. sensor hub) of mobile computing device 10 may determine of the real-time sensor is an event corresponding to one of the pre-determined intended uses of mobile computing device 10 based at least in part on a comparison of the real-time sensor data relative to hyperplane, as described below.
- real-time sensor data may be determined to corresponding to an imminent use of mobile computing device 10 based at least in part on analyzing a projection of vector mapping of the real-time sensor data.
- a projection of a vector corresponding to the real-time sensor data on a vector corresponding to steady-state condition may reduce the linear dependence of the vectors.
- a processor e.g. sensor hub
- FIGURE 2 illustrates an example sensor configuration of an example mobile computing device.
- an sensor array 20 of mobile computing device 10 may include one or more types of sensors.
- the one or more types of sensors may include a touch sensor, accelerometer, gyroscope, optical proximity sensor, ambient light sensor, image sensor, microphone, or any combination thereof.
- Different sensor types of sensor array 20 may each measure a different type of data.
- this disclosure describes the collection of environmental data associated with the mobile computing device by particular types of sensors, this disclosure contemplates the collection of sensor data associated with the mobile computing device by any suitable type of sensor.
- One or more sensors of sensor array 20 may be coupled to a sensor hub 40 of mobile computing device 10.
- sensor hub 40 may be a low power-consuming processor that controls one or more sensors of sensor array 20, manages power for sensors, processes sensor inputs, aggregates sensor data, and performs certain sensor functions.
- one or more types of sensors of sensor array 20 may be connected to a controller 42.
- sensor hub 40 may be coupled to controller 42 that is in turn coupled to sensor array 20.
- a sensor monitor may manage sensor array 20.
- sensor hub 40 or application processor of mobile computing device 10 detect a transition in the data measured by one or more types of sensors of sensor array 20 and correlate to the transitions in the data from the different types of sensors determine an imminent intended use of mobile computing device 10, as described below.
- sensor array 20 of mobile computing device 10 may include an accelerometer in addition to one or more other types of sensors.
- the sensor data provided by the accelerometer may be used at least in part to infer whether the user intends to use mobile computing device 10.
- mobile computing device 10 When the mobile computing device 10 is stored in a user's pocket, mobile computing device 10 may move as the user moves. However, such movements occur over a relatively long period of time.
- the user makes physical contact with mobile computing device 10 and takes mobile computing device 10 out of the pocket to bring it in front of the user's face, there may be an increase in the movement speed of mobile computing device 10 within a relatively short period of time. This change in a movement speed of mobile computing device 10 may be detected based on the sensor data supplied by the accelerometer.
- sensor array 20 of mobile computing device 10 may include a gyroscope in addition to one or more other types of sensors.
- a gyroscope is a type of sensor configured to measure the angular velocity along one or more positional axes.
- a gyroscope may be used to measure the orientation of mobile computing device 10.
- mobile computing device 10 when mobile computing device 10 is stored in the user's pocket, it may remain substantially in place along a particular orientation. However, when the user makes physical contact with mobile computing device 10 and takes it out of the pocket to bring it in front of the user's face, there may be a change in the orientation of mobile computing device 10 that occurs in a relatively short period of time.
- the change in orientation of mobile computing device 10 may be detected and measured by the gyroscope. If the orientation of mobile computing device 10 has changed significantly, the change of orientation may be a corroborative indicator along with data from another type of sensor, such as for example touch sensor or accelerometer data, that the user may have made physical contact with mobile computing device 10.
- another type of sensor such as for example touch sensor or accelerometer data
- sensor array 20 of mobile computing device 10 may include an optical-proximity sensor.
- the sensor data supplied by the optical proximity sensor may be analyzed to detect when mobile computing device 10 is in close proximity to a specific object, such as the user's hand.
- mobile computing device 10 may have an optical-proximity sensor with an infrared light-emitting diode (IR LED) placed on its back side.
- IR LED infrared light-emitting diode
- IR LED infrared light-emitting diode
- determination of an object in proximity to mobile computing device 10 may be a corroborative indicator along with data from another type of sensor, such as for example touch sensor or accelerometer data, that the user may have made physical contact with mobile computing device 10.
- correlating individual types of sensor data may be used to infer an intention of the user with respect to mobile computing device 10 (e.g. whether the user really means grasp mobile computing device 10 and use it).
- using multiple types of sensor data in combination may yield a more accurate inference of the user's intention with respect to mobile computing device 10 compared to using data from a single type of sensor in isolation.
- use of mobile computing device 10 may be inferred based at least in part on detecting a significant increase in the speed of the movement of mobile computing device 10 through an accelerometer in addition to detecting a body part of the user in proximity to mobile computing device 10 through one or more touch sensors.
- use of mobile computing device 10 may be inferred based at least in part on detecting a change of orientation of mobile computing device 10 through a gyroscope in addition to detecting a body part of the user in proximity to mobile computing device 10 through an optical proximity sensor.
- a pre-determined function of mobile computing device 10 may be in initiated based at least in part on the inferred intent of the user with respect to mobile computing device 10 as described below.
- mobile computing device 10 may be brought out of the power-saving state into a normal operational state (e.g. turn on the display of the mobile device) and input component of mobile computing device 10 may be unlocked automatically based at least in part on inferring the user may be about to use mobile computing device 10.
- FIGURE 3 illustrates an example method for initiating a pre-determined function of a computing device based on an imminent intended use.
- the method may start at step 300, where a computing device receives real-time sensor data from a number of sensor types of computing devices.
- the computing device may calculate a derivative of the sensor data to determine a transition in the sensor data.
- a processor of a mobile computing device may receive the sensor data and perform an operation, such as for example calculating a derivative of the sensor data as a function of time.
- the sensors of one of the computing devices includes different sensor types, such as for example a touch sensor, accelerometer, gyroscope, optical proximity sensor, or any combination thereof.
- Step 302 by the computing device, correlates the real-time sensor data from the sensors of different sensor types.
- a processor may apply a convolution operation to the sensor data to determine whether the data chronologically overlaps.
- An example convolution operation may be illustrated by the following equation:
- M is the result of the convolution of data from multiple types of sensors, and f and g are the derivative of the data from a sensor, such as for example f may be the derivative of the data measured by an accelerometer, g may be the derivative of the data measured by a touch sensor.
- the result of the convolution operation may determine whether a transition in the sensor data from different types of sensors chronologically overlap.
- an a priori function such as for example Heaviside or sigmoid functions, may replace the derivative operator.
- a processor may convolve the data measured a first type of sensor, such as for example touch sensor with data measured by a second type of sensor, such as for example an accelerometer.
- an application processor or sensor hub of a mobile computing device may convolve the data measured a first type of sensor, such as for example a touch sensor, with data measured by a second type of sensor, such as for example an optical-proximity sensor.
- Step 304 by the computing device, may determine an intended imminent use of the computing device based on the correlation. In particular embodiments, based at least in part on a transition in the data of multiple sensor types chronologically overlapping. As an example and not by way of limitation, the computing device may determine the imminent intended use of the computing device based at least in part on a transition in the real-time sensor data from a touch sensor and accelerometer occurring at substantially the same time.
- the computing device may automatically initiate a pre-determined function of the computing device based at least in part on the determination of the intended imminent use of the computing device, at which point the method may end.
- the pre-determined function may be initiated in response to the result of a convolution operation M illustrated by equation (1) being higher than a pre-determined threshold value.
- the pre-determined function may power down the computing device associated with the sensors in response to the result of the convolution operation being higher than a pre-determined threshold value.
- Particular embodiments may repeat one or more steps of the method of FIGURE 3, where appropriate.
- this disclosure describes and illustrates particular components carrying out particular steps of the method of FIGURE 3, this disclosure contemplates any suitable combination of any suitable components, such as for example a processor of a mobile computing device, carrying out any suitable steps of the method of FIGURE 3.
- FIGURES 4A-B illustrate example detection of a transition in example sensor data.
- this disclosure describes pre-processing the sensor data through a particular linear function, such as for example a derivative function, this disclosure contemplates preprocessing the sensor data through any suitable linear function here, such as for example a convolution with a Heaviside or sigmoid function.
- sensor data 52 and 54 from one or more sensors may be measured as a function of time, as illustrated by 44 and 46 in the example of FIGURE 4A, and sensor data 52 and 54 may be analyzed to infer an intention of the user with respect to the computing device associated with the sensors.
- inference of the intention of the user with respect to a particular computing device may be performed sensor data 52 and 54 from multiple sensor types.
- sensor data 52 may be data measured by a touch sensor of a mobile computing device and sensor data 54 may be data measured by an accelerometer.
- this disclosure contemplates any suitable form of sensor data 52 and 54 such as for example current, voltage, charge, or any combination thereof.
- an intended use of the computing device may be determined through a transition in the data from one state to another measured by sensors associated with the computing device, as described above.
- a transition in sensor data may indicate a mobile computing device is being picked up and about to be used, as described above.
- a transition in sensor data 52 and 54 may be detected based at least in part on calculating a derivative 56 and 58 of sensor data 52 and 54, respectively, as illustrated in the example of FIGURE 4B by 48 and 50.
- a change in the derivative 56 and 58 of the sensor data 52 and 54, respectively may be detectable during time period 49 in cases where the change in the sensor data 52 and 54 may be relatively small.
- the derivative 56 and 58 of the sensor data may be provided to a processor to determine an intended immediate use of computing device, as described above.
- FIGURE 5 illustrates an example network environment 100 associated with a social-networking system.
- Network environment 100 includes a user 101, a client system 130, a social-networking system 160, and a third-party system 170 connected to each other by a network 110.
- FIGURE 5 illustrates a particular arrangement of user 101, client system 130, social-networking system 160, third-party system 170, and network 110, this disclosure contemplates any suitable arrangement of user 101, client system 130, social-networking system 160, third-party system 170, and network 110.
- two or more of client system 130, social-networking system 160, and third-party system 170 may be connected to each other directly, bypassing network 110.
- client system 130 may be physically or logically co-located with each other in whole or in part.
- FIGURE 5 illustrates a particular number of users 101, client systems 130, social-networking systems 160, third-party systems 170, and networks 110
- this disclosure contemplates any suitable number of users 101, client systems 130, social-networking systems 160, third-party systems 170, and networks 110.
- network environment 100 may include multiple users 101, client system 130, social-networking systems 160, third-party systems 170, and networks 110.
- social-networking system 160 may include one or more servers. Each server may be a unitary server or a distributed server spanning multiple computers or multiple datacenters. Servers may be of various types, such as, for example and without limitation, web server, news server, mail server, message server, advertising server, file server, application server, exchange server, database server, proxy server, another server suitable for performing functions or processes described herein, or any combination thereof. In particular embodiments, each server may include hardware, software, or embedded logic components or a combination of two or more such components for carrying out the appropriate functionalities implemented or supported by server. In particular embodiments, social-networking system 164 may include one or more data stores. Data stores may be used to store various types of information.
- each data store may be a relational, columnar, correlation, or other suitable database.
- this disclosure describes or illustrates particular types of databases, this disclosure contemplates any suitable types of databases.
- Particular embodiments may provide interfaces that enable a client system 130, a social-networking system 160, or a third-party system 170 to manage, retrieve, modify, add, or delete, the information stored in data store.
- the sensor data received from client system 130 may function as training data for a machine learning algorithm, such as for example SVM, k-means, Bayesian inference, or a neural network, executed on social-networking system 160.
- a machine learning algorithm such as for example SVM, k-means, Bayesian inference, or a neural network
- one or more servers of social- networking system 160 may receive training data from one or more of client systems 130 (e.g. a mobile computing device), and use a machine-learning algorithm to correlate sensor data values from particular activities using client system 130 with one or more particular states of client system 130.
- one or more servers executing the machine-learning algorithm may receive sensor values from sensors of client system 130, such as for example an accelerometer, gyroscope, ambient light sensor, optical proximity sensor, or another sensor of one or more client systems 130.
- data defining a hyperplane determined from the training data may be sent to client system 130 for determining an imminent intended use of client system 130.
- subsequent sensor data may be sent by mobile computing device 10 to re-define the hyperplane.
- updated data re-defining the hyperplane may be received by mobile computing device 10.
- user 101 may be an individual (human user), an entity (e.g., an enterprise, business, or third-party application), or a group (e.g., of individuals or entities) that interacts or communicates with or over social-networking system 160.
- social-networking system 160 may be a network-addressable computing system hosting an online social network. Social-networking system 160 may generate, store, receive, and send social-networking data, such as, for example, user-profile data, concept-profile data, social-graph information, or other suitable data related to the online social network. Social- networking system 160 may be accessed by the other components of network environment 100 either directly or via network 110.
- social-networking system 160 may include an authorization server (or other suitable component(s)) that allows users 101 to opt in to or opt out of having their actions logged by social-networking system 160 or shared with other systems (e.g., third-party systems 170), for example, by setting appropriate privacy settings.
- a privacy setting of a user may determine what information associated with the user may be logged, how information associated with the user may be logged, when information associated with the user may be logged, who may log information associated with the user, whom information associated with the user may be shared with, and for what purposes information associated with the user may be logged or shared.
- Authorization servers may be used to enforce one or more privacy settings of the users of social-networking system 160 through blocking, data hashing, anonymization, or other suitable techniques as appropriate.
- Third-party system 170 may be accessed by the other components of network environment 100 either directly or via network 110.
- one or more users 101 may use one or more client systems 130 to access, send data to, and receive data from social-networking system 160 or third-party system 170.
- Client system 130 may access social-networking system 160 or third- party system 170 directly, via network 110, or via a third-party system.
- client system 130 may access third-party system 170 via social-networking system 160.
- Client system 130 may be any suitable computing device, such as, for example, a personal computer, a laptop computer, a cellular telephone, a smartphone, or a tablet computer.
- network 110 may include any suitable network 110.
- one or more portions of network 110 may include an ad hoc network, an intranet, an extranet, a virtual private network (VPN), a local area network (LAN), a wireless LAN (WLAN), a wide area network (WAN), a wireless WAN (WW AN), a metropolitan area network (MAN), a portion of the Internet, a portion of the Public Switched Telephone Network (PSTN), a cellular telephone network, or a combination of two or more of these.
- Network 110 may include one or more networks 110.
- Links 150 may connect client system 130, social-networking system 160, and third-party system 170 to communication network 110 or to each other.
- This disclosure contemplates any suitable links 150.
- one or more links 150 include one or more wireline (such as for example Digital Subscriber Line (DSL) or Data Over Cable Service Interface Specification (DOCSIS)), wireless (such as for example Wi-Fi or Worldwide Interoperability for Microwave Access (WiMAX)), or optical (such as for example Synchronous Optical Network (SONET) or Synchronous Digital Hierarchy (SDH)) links.
- wireline such as for example Digital Subscriber Line (DSL) or Data Over Cable Service Interface Specification (DOCSIS)
- wireless such as for example Wi-Fi or Worldwide Interoperability for Microwave Access (WiMAX)
- optical such as for example Synchronous Optical Network (SONET) or Synchronous Digital Hierarchy (SDH) links.
- SONET Synchronous Optical Network
- SDH Synchronous Digital Hierarchy
- one or more links 150 each include an ad hoc network, an intranet, an extranet, a VPN, a LAN, a WLAN, a WAN, a WWAN, a MAN, a portion of the Internet, a portion of the PSTN, a cellular technology-based network, a satellite communications technology-based network, another link 150, or a combination of two or more such links 150.
- Links 150 need not necessarily be the same throughout network environment 100.
- One or more first links 150 may differ in one or more respects from one or more second links 150.
- FIGURE 6 illustrates an example classification of sensor data using an example machine learning algorithm.
- training data from one or more sensors of a client system may include sensor data from each sensor captured during the performance of a particular activity and indicator information corresponding to a particular state of the client system associated with the particular activity.
- the sensor data may be raw measurement data from the sensors or sensor data that has been pre-processed, such as for example, to calculate the derivative of the raw sensor data, as described above.
- the sensor data may correspond to a transition in a physical state (e.g. movement) of the client system.
- sensor data may be further processed, such as for example through a filtering or convolution operation.
- the training data from each particular activity may be classified into one of two particular states associated with the client device based at least in part on the indicator information associated with each set of sensor data, as described above.
- one or more sets of sensor data may correspond to activity associated with physical contact with a mobile computing device, e.g. holding the mobile computing device, and one or more sets of sensor data may correspond to activity not associated with physical contact with the mobile computing device, e.g. resting the mobile computing device on a table.
- the training data for each particular action may be represented as a vector 202A-B in a N-dimensional space 200, where N may be equal to the number of sensors of the client system.
- each vector 202A-B may be mapped to N-dimensional space 200 through a kernel function.
- each vector 202A-B may based at least in part on the N-tuple of the derivative of the sensor data.
- vectors 202A-B may be classified with one of two particular states associated with the client system that are separated by a hyperplane 206 or a non-linear surface in N-dimensional space 200.
- hyperplane 206 may have N-l dimensions and be defined by a set of points with a constant dot product with one or more support vectors of each state.
- the support vectors may be defined as the vector for each particular state that has a maximum derivative and the distance between hyperplane 206 and each support vector may be maximized.
- data defining hyperplane 206 may be sent to the client system.
- hyperplane 206 may be modified based on subsequent vectors determined from subsequent sensor data received from the client system.
- the updated data re-defining hyperplane 206 may be sent to the client system.
- an imminent use of the client system may be determined by the client system based at least in part on the classification of a vector corresponding to subsequent sensor data from client system with a particular state of the client system.
- classification of the vector corresponding to subsequent sensor data may be based at least in part on the position of the vector relative to hyperplane 206.
- the user of the client system intends to use the client system based at least in part on the vector corresponding to subsequent sensor data being classified with a state corresponding to physical contact with the client system, such as for example, defined by vectors 202A.
- the imminent use of the client system may be determined to correspond to physical contact with the client system when the vector is on a same side of hyperplane 206 as vectors 202A. Otherwise, if the subsequent vector is located on a same side of hyperplane 206 as vectors 202B, it may be determined the client system is substantially stationary.
- a processor of the client system may initiate a pre-determined function of the client system based at least in part on classifying subsequent vectors with a particular state of a client system.
- FIGURE 7 illustrates an example method of determining whether sensor data corresponds to a pre-determined use of a client system.
- the method may start at step 310, where a computing device receives real-time sensor data from sensors on the computing device.
- the real-time sensor data may correspond to a transition in a physical state of the computing device caused by a user of the computing device.
- Step 312 by the computing device, applies a linear function to the the real-time sensor data from each sensor.
- the linear function may comprise a filtering function, derivative function, convolution of a Heaviside or sigmoid function, or any combination thereof.
- a processor of a mobile computing device may receive the sensor data and perform an operation, such as for example calculating a derivative of the sensor data as a function of time.
- Step 314 by the computing device, determines a vector based on a tuple of the derivatives.
- the tuple may have dimension equal to the number of sensors.
- the computing device may compare the vector with a pre-determined hyperplane. As described above, the hyperplane may have dimensions one fewer than the number of sensors of the computing device.
- the computing device may determine based on the comparison whether the transition is an event that corresponds to any pre-determined imminent use of the computing device, at which point the method may end. In particular embodiments, the determination may be made through determining the position of the vector relative to the predetermined hyperplane.
- this disclosure describes and illustrates particular steps of the method of FIGURE 7 as occurring in a particular order, this disclosure contemplates any suitable steps of the method of FIGURE 7 occurring in any suitable order. Particular embodiments may repeat one or more steps of the method of FIGURE 7, where appropriate.
- this disclosure describes and illustrates particular components carrying out particular steps of the method of FIGURE 7, this disclosure contemplates any suitable combination of any suitable components, such as for example a processor of a mobile computing device, carrying out any suitable steps of the method of FIGURE 7.
- FIGURE 8 illustrates an example isolation of components of sensor data through calculation of an example projection.
- mapping sensor data to N- dimensional space 200 may be used to isolate particular components of the sensor data.
- the linear dependence of one sensor to another sensor with a degree of spatial overlap may be reduced through determination of a projection 84A of the data of one sensor to another, as described below.
- a mobile computing device may include multiple touch sensors in multiple locations of the mobile computing device, as illustrated in the example of FIGURE 1.
- the mobile computing device may include a first touch sensor having a touch-sensitive area with coverage along a side of the mobile computing device and a second touch sensor having a touch- sensitive area that may include at least a portion of two or more surfaces (e.g. a side and bottom).
- the linear dependence of sensor data with a degree of temporal separation may be reduced through determination of a projection of the sensor data.
- a portion of current sensor data may be isolated based at least in part on determining a projection of a vector corresponding to current sensor data on a vector corresponding to a prior steady state condition.
- an imminent use of the client system may be determined by analyzing a projection 84 A or 84B.
- the sensor data from a client system may be temporally separated data from one or more spatially overlapping sensors and correspond to a transition from a steady-state condition of the client system.
- the projection may be calculated using raw measurement data or sensor data that has been pre-processed, such as for example by calculating the derivative of the raw sensor data, as described above.
- sensor data may be further processed, such as for example, through a filtering or convolution operation.
- the sensor data captured at particular times may each be represented as a vector 80 and 82 in a N- dimensional space 200, where N may be equal to the number of sensors of the client system.
- each vector 80 and 82 may be mapped to N- dimensional space 200 through a kernel function.
- each vector 80 and 82 may based at least in part on the N-tuple of the derivative of the sensor data.
- the projection 84A of vector 80 corresponding to the steady-state condition on vector 82 corresponding to real-time sensor data may be determined based at least in part on a dot product of the vectors 80 and 82.
- vectors 80 and 82 may be a derivative of the sensor data of a client system.
- one or more components of vector 82 that differs from vector 80 from temporally separated measurements may be isolated by projection 84A of vector 82 on vector 80.
- An example calculation of projection 84A of vector 82 on vector 80 may be illustrated by the following equation:
- steady-state condition i.e. vector 80 of space 200
- real-time data i.e. vector 82
- projection 84A on vector 80 may be calculated through the dot product, as illustrated by equation (2).
- projection 84A may be translated to an origin of N-dimensional space 200, for inferring an intent of the user with respect to the client system, as described below.
- an intent of the user with respect to a client system may be inferred based at least in part on analysis of projection 84B.
- projection 84B may be classified with a pre-defined imminent use of the client system as described above.
- projection 84B may be compared with a pre-defined projection that corresponds to an imminent use of the client system.
- it may be determined that the user of a particular client system intends to use the client system based at least in part on a projection 84B being classified with a state corresponding to physical contact with the client system.
- a processor of the client system may initiate a pre-determined function of the client system based at least in part on inferring an intent of the user based at least in part on analysis of projection 84B.
- FIGURE 9 illustrates an example method of isolating a component of sensor data.
- the method may start at step 320, where a computing device receives real-time sensor data from sensors on the computing device.
- the sensors may be located on multiple surfaces of the computing device.
- Step 322 by the computing device, detects a transition in the real-time sensor data from a steady state.
- a processor of a mobile computing device may receive the sensor data and perform an operation, such as for example calculating a derivative of the sensor data as a function of time.
- the computing device may determine based on the detection an imminent use of the computing device, at which point the method may end.
- the computing device may include determining a vector based on a tuple of the derivatives and calculating a projection of the vector of the real-time sensor data on a vector of the steady-state of the computing device. In particular embodiments, the determination may be made through comparing the projection with a pre-determined projection corresponding to one or more imminent uses.
- this disclosure describes and illustrates particular steps of the method of FIGURE 9 as occurring in a particular order, this disclosure contemplates any suitable steps of the method of FIGURE 9 occurring in any suitable order. Particular embodiments may repeat one or more steps of the method of FIGURE 9, where appropriate.
- this disclosure describes and illustrates particular components carrying out particular steps of the method of FIGURE 9, this disclosure contemplates any suitable combination of any suitable components, such as for example a processor of a mobile computing device, carrying out any suitable steps of the method of FIGURE 9.
- FIGURE 10 illustrates an example computing system.
- one or more computer systems 60 perform one or more steps of one or more methods described or illustrated herein.
- one or more computer systems 60 provide functionality described or illustrated herein.
- software running on one or more computer systems 60 performs one or more steps of one or more methods described or illustrated herein or provides functionality described or illustrated herein.
- Particular embodiments include one or more portions of one or more computer systems 60.
- reference to a computer system may encompass a computing device, where appropriate.
- reference to a computer system may encompass one or more computer systems, where appropriate.
- This disclosure contemplates any suitable number of computer systems 60.
- This disclosure contemplates computer system 60 taking any suitable physical form.
- computer system 60 may be an embedded computer system, a system- on-chip (SOC), a single-board computer system (SBC) (such as, for example, a computer-on- module (COM) or system-on-module (SOM)), a desktop computer system, a laptop or notebook computer system, an interactive kiosk, a mainframe, a mesh of computer systems, a mobile computing system 10, a personal digital assistant (PDA), a server, a tablet computer system, or a combination of two or more of these.
- SOC system- on-chip
- SBC single-board computer system
- COM computer-on- module
- SOM system-on-module
- desktop computer system such as, for example, a computer-on- module (COM) or system-on-module (SOM)
- mainframe such as, for example, a computer-on- module (COM) or system-on-module
- computer system 60 may include one or more computer systems 60; be unitary or distributed; span multiple locations; span multiple machines; span multiple data centers; or reside in a cloud, which may include one or more cloud components in one or more networks.
- one or more computer systems 60 may perform without substantial spatial or temporal limitation one or more steps of one or more methods described or illustrated herein.
- one or more computer systems 60 may perform in real time or in batch mode one or more steps of one or more methods described or illustrated herein.
- One or more computer systems 60 may perform at different times or at different locations one or more steps of one or more methods described or illustrated herein, where appropriate.
- computer system 60 includes a processor 62, memory 64, storage 66, an input/output (I/O) interface 68, a communication interface 70, and a bus 72.
- I/O input/output
- this disclosure describes and illustrates a particular computer system having a particular number of particular components in a particular arrangement, this disclosure contemplates any suitable computer system having any suitable number of any suitable components in any suitable arrangement.
- processor 62 includes hardware for executing instructions, such as those making up a computer program.
- processor 62 may retrieve (or fetch) the instructions from an internal register, an internal cache, memory 64, or storage 66; decode and execute them; and then write one or more results to an internal register, an internal cache, memory 64, or storage 66.
- processor 62 may include one or more internal caches for data, instructions, or addresses. This disclosure contemplates processor 62 including any suitable number of any suitable internal caches, where appropriate.
- processor 62 may include one or more instruction caches, one or more data caches, and one or more translation lookaside buffers (TLBs).
- TLBs translation lookaside buffers
- Instructions in the instruction caches may be copies of instructions in memory 64 or storage 66, and the instruction caches may speed up retrieval of those instructions by processor 62.
- Data in the data caches may be copies of data in memory 64 or storage 66 for instructions executing at processor 62 to operate on; the results of previous instructions executed at processor 62 for access by subsequent instructions executing at processor 62 or for writing to memory 64 or storage 66; or other suitable data.
- the data caches may speed up read or write operations by processor 62.
- the TLBs may speed up virtual-address translation for processor 62.
- processor 62 may include one or more internal registers for data, instructions, or addresses. This disclosure contemplates processor 62 including any suitable number of any suitable internal registers, where appropriate. Where appropriate, processor 62 may include one or more arithmetic logic units (ALUs); be a multi- core processor; or include one or more processors 62.
- memory 64 includes main memory for storing instructions for processor 62 to execute or data for processor 62 to operate on.
- computer system 60 may load instructions from storage 66 or another source (such as, for example, another computer system 60) to memory 64.
- Processor 62 may then load the instructions from memory 64 to an internal register or internal cache.
- processor 62 may retrieve the instructions from the internal register or internal cache and decode them.
- processor 62 may write one or more results (which may be intermediate or final results) to the internal register or internal cache.
- Processor 62 may then write one or more of those results to memory 64.
- processor 62 executes only instructions in one or more internal registers or internal caches or in memory 64 (as opposed to storage 66 or elsewhere) and operates only on data in one or more internal registers or internal caches or in memory 64 (as opposed to storage 66 or elsewhere).
- One or more memory buses (which may each include an address bus and a data bus) may couple processor 62 to memory 64.
- Bus 72 may include one or more memory buses, as described below.
- one or more memory management units (MMUs) reside between processor 62 and memory 64 and facilitate accesses to memory 64 requested by processor 62.
- memory 64 includes random access memory (RAM). This RAM may be volatile memory, where appropriate Where appropriate, this RAM may be dynamic RAM (DRAM) or static RAM (SRAM). Moreover, where appropriate, this RAM may be single-ported or multi-ported RAM. This disclosure contemplates any suitable RAM.
- Memory 64 may include one or more memories 64, where appropriate.
- storage 66 includes mass storage for data or instructions.
- storage 66 may include a hard disk drive (HDD), a floppy disk drive, flash memory, an optical disc, a magneto-optical disc, magnetic tape, or a Universal Serial Bus (USB) drive or a combination of two or more of these.
- HDD hard disk drive
- floppy disk drive flash memory
- optical disc an optical disc
- magneto-optical disc magnetic tape
- USB Universal Serial Bus
- Storage 66 may include removable or non-removable (or fixed) media, where appropriate.
- Storage 66 may be internal or external to computer system 60, where appropriate.
- storage 66 is non-volatile, solid-state memory.
- storage 66 includes read-only memory (ROM).
- this ROM may be mask-programmed ROM, programmable ROM (PROM), erasable PROM (EPROM), electrically erasable PROM (EEPROM), electrically alterable ROM (EAROM), or flash memory or a combination of two or more of these.
- This disclosure contemplates mass storage 66 taking any suitable physical form.
- Storage 66 may include one or more storage control units facilitating communication between processor 62 and storage 66, where appropriate. Where appropriate, storage 66 may include one or more storages 66. Although this disclosure describes and illustrates particular storage, this disclosure contemplates any suitable storage.
- I/O interface 68 includes hardware, software, or both providing one or more interfaces for communication between computer system 60 and one or more I/O devices.
- Computer system 60 may include one or more of these I/O devices, where appropriate.
- One or more of these I/O devices may enable communication between a person and computer system 60.
- an I/O device may include a keyboard, keypad, microphone, monitor, mouse, printer, scanner, speaker, still camera, stylus, tablet, touch screen, trackball, video camera, another suitable I/O device or a combination of two or more of these.
- An I/O device may include one or more sensors. This disclosure contemplates any suitable I/O devices and any suitable I/O interfaces 68 for them.
- I/O interface 68 may include one or more device or software drivers enabling processor 62 to drive one or more of these I/O devices.
- I/O interface 68 may include one or more I/O interfaces 68, where appropriate. Although this disclosure describes and illustrates a particular I/O interface, this disclosure contemplates any suitable I/O interface.
- communication interface 70 includes hardware, software, or both providing one or more interfaces for communication (such as for example, packet-based communication) between computer system 60 and one or more other computer systems 60 or one or more networks.
- communication interface 70 may include a network interface controller (NIC) or network adapter for communicating with an Ethernet or other wire-based network or a wireless NIC (WNIC) or wireless adapter for communicating with a wireless network, such as a WI-FI network.
- NIC network interface controller
- WNIC wireless NIC
- WI-FI network wireless network
- computer system 60 may communicate with an ad hoc network, a personal area network (PAN), a local area network (LAN), a wide area network (WAN), a metropolitan area network (MAN), or one or more portions of the Internet or a combination of two or more of these.
- PAN personal area network
- LAN local area network
- WAN wide area network
- MAN metropolitan area network
- computer system 60 may communicate with a wireless PAN (WPAN) (such as for example, a BLUETOOTH WPAN), a WI-FI network, a WI-MAX network, a cellular telephone network (such as, for example, a Global System for Mobile Communications (GSM) network), or other suitable wireless network or a combination of two or more of these.
- WPAN wireless PAN
- WI-FI wireless personal area network
- WI-MAX wireless personal area network
- WI-MAX wireless personal area network
- cellular telephone network such as, for example, a Global System for Mobile Communications (GSM) network
- GSM Global System for
- bus 72 includes hardware, software, or both coupling components of computer system 60 to each other.
- bus 72 may include an Accelerated Graphics Port (AGP) or other graphics bus, an Enhanced Industry Standard Architecture (EISA) bus, a front-side bus (FSB), a HYPERTRANSPORT (HT) interconnect, an Industry Standard Architecture (ISA) bus, an INFINIBAND interconnect, a low-pin-count (LPC) bus, a memory bus, a Micro Channel Architecture (MCA) bus, a Peripheral Component Interconnect (PCI) bus, a PCI-Express (PCIe) bus, a serial advanced technology attachment (SATA) bus, a Video Electronics Standards Association local (VLB) bus, or another suitable bus or a combination of two or more of these.
- Bus 72 may include one or more buses 72, where appropriate.
- a computer-readable non-transitory storage medium or media may include one or more semiconductor-based or other integrated circuits (ICs) (such, as for example, field- programmable gate arrays (FPGAs) or application-specific ICs (ASICs)), hard disk drives (HDDs), hybrid hard drives (HHDs), optical discs, optical disc drives (ODDs), magneto-optical discs, magneto-optical drives, floppy diskettes, floppy disk drives (FDDs), magnetic tapes, solid- state drives (SSDs), RAM-drives, SECURE DIGITAL cards or drives, any other suitable computer-readable non-transitory storage media, or any suitable combination of two or more of these, where appropriate.
- ICs semiconductor-based or other integrated circuits
- HDDs hard disk drives
- HHDs hybrid hard drives
- ODDs optical disc drives
- magneto-optical discs magneto-optical drives
- FDDs floppy diskettes
- FDDs floppy disk drives
- an apparatus or system or a component of an apparatus or system being adapted to, arranged to, capable of, configured to, enabled to, operable to, or operative to perform a particular function encompasses that apparatus, system, component, whether or not it or that particular function is activated, turned on, or unlocked, as long as that apparatus, system, or component is so adapted, arranged, capable, configured, enabled, operable, or operative.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Hardware Design (AREA)
- Computing Systems (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Quality & Reliability (AREA)
- Medical Informatics (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Environmental & Geological Engineering (AREA)
- User Interface Of Digital Computer (AREA)
- Telephone Function (AREA)
- Testing Or Calibration Of Command Recording Devices (AREA)
Abstract
Description
Claims
Priority Applications (11)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020167003682A KR101726825B1 (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection |
MX2016000428A MX349635B (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection. |
JP2016525473A JP6117438B2 (en) | 2013-07-12 | 2014-07-10 | Grab detection calibration |
CN201480050432.1A CN105531645B (en) | 2013-07-12 | 2014-07-10 | Grasp the calibration of detection |
EP14823297.8A EP3019934B1 (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection |
AU2014287240A AU2014287240B2 (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection |
CA2917970A CA2917970C (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection |
KR1020177008795A KR102068541B1 (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection |
IL243505A IL243505A (en) | 2013-07-12 | 2016-01-07 | Calibration of grab detection |
IL249974A IL249974B (en) | 2013-07-12 | 2017-01-08 | Calibration of grab detection |
AU2017201926A AU2017201926B2 (en) | 2013-07-12 | 2017-03-22 | Calibration of grab detection |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/941,289 | 2013-07-12 | ||
US13/941,289 US9372103B2 (en) | 2013-07-12 | 2013-07-12 | Calibration of grab detection |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015006523A1 true WO2015006523A1 (en) | 2015-01-15 |
Family
ID=52277778
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2014/046072 WO2015006523A1 (en) | 2013-07-12 | 2014-07-10 | Calibration of grab detection |
Country Status (10)
Country | Link |
---|---|
US (3) | US9372103B2 (en) |
EP (1) | EP3019934B1 (en) |
JP (2) | JP6117438B2 (en) |
KR (2) | KR102068541B1 (en) |
CN (2) | CN105531645B (en) |
AU (2) | AU2014287240B2 (en) |
CA (2) | CA2955848C (en) |
IL (2) | IL243505A (en) |
MX (1) | MX349635B (en) |
WO (1) | WO2015006523A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150019153A1 (en) * | 2013-07-12 | 2015-01-15 | Facebook, Inc. | Calibration of Grab Detection |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160320850A1 (en) * | 2015-04-29 | 2016-11-03 | Samsung Electronics Co., Ltd. | User interface control using impact gestures |
US11301046B2 (en) * | 2017-07-20 | 2022-04-12 | Sony Corporation | Electronic apparatus, information processing device, and information processing method |
JP6963450B2 (en) | 2017-09-22 | 2021-11-10 | 三菱パワー株式会社 | Rotating machine control device, rotating machine equipment, rotating machine control method, and rotating machine control program |
CN109491525B (en) * | 2018-11-13 | 2022-04-01 | 宁波视睿迪光电有限公司 | Method and device for realizing low power consumption of interactive pen |
CN113487674B (en) * | 2021-07-12 | 2024-03-08 | 未来元宇数字科技(北京)有限公司 | Human body pose estimation system and method |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080318626A1 (en) * | 2007-06-22 | 2008-12-25 | Broadcom Corporation | Multi-mode mobile communication device with motion sensor and methods for use therewith |
US20090088204A1 (en) * | 2007-10-01 | 2009-04-02 | Apple Inc. | Movement-based interfaces for personal media device |
US20090265671A1 (en) | 2008-04-21 | 2009-10-22 | Invensense | Mobile devices with motion gesture recognition |
US20090296991A1 (en) * | 2008-05-29 | 2009-12-03 | Anzola Carlos A | Human interface electronic device |
US20090303204A1 (en) | 2007-01-05 | 2009-12-10 | Invensense Inc. | Controlling and accessing content using motion processing on mobile devices |
US20120036433A1 (en) * | 2010-08-04 | 2012-02-09 | Apple Inc. | Three Dimensional User Interface Effects on a Display by Using Properties of Motion |
US20120280917A1 (en) | 2011-05-03 | 2012-11-08 | Toksvig Michael John Mckenzie | Adjusting Mobile Device State Based on User Intentions and/or Identity |
US20130162525A1 (en) * | 2009-07-14 | 2013-06-27 | Cywee Group Limited | Method and apparatus for performing motion recognition using motion sensor fusion, and associated computer program product |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5911011A (en) * | 1997-02-28 | 1999-06-08 | Canon Kabushiki Kaisha | Multidimensional close neighbor search |
JP2003141498A (en) * | 2001-10-30 | 2003-05-16 | Nippon Telegr & Teleph Corp <Ntt> | Answer extracting method, device, program and recording medium recorded the same therein |
KR100668341B1 (en) * | 2005-06-29 | 2007-01-12 | 삼성전자주식회사 | Method and apparatus for function selection by user's hand grip shape |
US8836502B2 (en) * | 2007-12-28 | 2014-09-16 | Apple Inc. | Personal media device input and output control based on associated conditions |
FI20095570L (en) * | 2009-05-22 | 2009-09-11 | Valtion Teknillinen | Context recognition in mobile devices |
US9398536B2 (en) * | 2009-05-29 | 2016-07-19 | Qualcomm Incorporated | Method and apparatus for movement detection by evaluating elementary movement patterns |
JP5356923B2 (en) | 2009-06-11 | 2013-12-04 | Kddi株式会社 | Method and system for estimating movement state of portable terminal device |
WO2010151183A1 (en) * | 2009-06-23 | 2010-12-29 | Telefonaktiebolaget L M Ericsson (Publ) | Method and an arrangement for a mobile telecommunications network |
US8847880B2 (en) * | 2009-07-14 | 2014-09-30 | Cywee Group Ltd. | Method and apparatus for providing motion library |
US8432368B2 (en) * | 2010-01-06 | 2013-04-30 | Qualcomm Incorporated | User interface methods and systems for providing force-sensitive input |
US8335935B2 (en) * | 2010-03-29 | 2012-12-18 | Intel Corporation | Power management based on automatic workload detection |
US8874129B2 (en) * | 2010-06-10 | 2014-10-28 | Qualcomm Incorporated | Pre-fetching information based on gesture and/or location |
CN102288398B (en) * | 2011-05-16 | 2013-04-10 | 南京航空航天大学 | Momentum wheel fault detection device and method based on support vector machine |
US9372103B2 (en) * | 2013-07-12 | 2016-06-21 | Facebook, Inc. | Calibration of grab detection |
US9393693B1 (en) * | 2014-07-10 | 2016-07-19 | Google Inc. | Methods and systems for determining and modeling admissible gripper forces for robotic devices |
-
2013
- 2013-07-12 US US13/941,289 patent/US9372103B2/en active Active
-
2014
- 2014-07-10 CN CN201480050432.1A patent/CN105531645B/en active Active
- 2014-07-10 KR KR1020177008795A patent/KR102068541B1/en active IP Right Grant
- 2014-07-10 MX MX2016000428A patent/MX349635B/en active IP Right Grant
- 2014-07-10 JP JP2016525473A patent/JP6117438B2/en active Active
- 2014-07-10 KR KR1020167003682A patent/KR101726825B1/en active IP Right Grant
- 2014-07-10 AU AU2014287240A patent/AU2014287240B2/en not_active Ceased
- 2014-07-10 WO PCT/US2014/046072 patent/WO2015006523A1/en active Application Filing
- 2014-07-10 EP EP14823297.8A patent/EP3019934B1/en active Active
- 2014-07-10 CA CA2955848A patent/CA2955848C/en not_active Expired - Fee Related
- 2014-07-10 CN CN201710463551.5A patent/CN107422790B/en active Active
- 2014-07-10 CA CA2917970A patent/CA2917970C/en not_active Expired - Fee Related
-
2016
- 2016-01-07 IL IL243505A patent/IL243505A/en not_active IP Right Cessation
- 2016-05-12 US US15/153,547 patent/US10582038B2/en not_active Expired - Fee Related
-
2017
- 2017-01-08 IL IL249974A patent/IL249974B/en not_active IP Right Cessation
- 2017-03-22 AU AU2017201926A patent/AU2017201926B2/en not_active Ceased
- 2017-03-22 JP JP2017055735A patent/JP6416962B2/en not_active Expired - Fee Related
-
2019
- 2019-12-19 US US16/721,877 patent/US10742798B1/en not_active Expired - Fee Related
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090303204A1 (en) | 2007-01-05 | 2009-12-10 | Invensense Inc. | Controlling and accessing content using motion processing on mobile devices |
US20080318626A1 (en) * | 2007-06-22 | 2008-12-25 | Broadcom Corporation | Multi-mode mobile communication device with motion sensor and methods for use therewith |
US20090088204A1 (en) * | 2007-10-01 | 2009-04-02 | Apple Inc. | Movement-based interfaces for personal media device |
US20090265671A1 (en) | 2008-04-21 | 2009-10-22 | Invensense | Mobile devices with motion gesture recognition |
US20090296991A1 (en) * | 2008-05-29 | 2009-12-03 | Anzola Carlos A | Human interface electronic device |
US20130162525A1 (en) * | 2009-07-14 | 2013-06-27 | Cywee Group Limited | Method and apparatus for performing motion recognition using motion sensor fusion, and associated computer program product |
US20120036433A1 (en) * | 2010-08-04 | 2012-02-09 | Apple Inc. | Three Dimensional User Interface Effects on a Display by Using Properties of Motion |
US20120280917A1 (en) | 2011-05-03 | 2012-11-08 | Toksvig Michael John Mckenzie | Adjusting Mobile Device State Based on User Intentions and/or Identity |
Non-Patent Citations (1)
Title |
---|
See also references of EP3019934A4 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150019153A1 (en) * | 2013-07-12 | 2015-01-15 | Facebook, Inc. | Calibration of Grab Detection |
US9372103B2 (en) * | 2013-07-12 | 2016-06-21 | Facebook, Inc. | Calibration of grab detection |
Also Published As
Publication number | Publication date |
---|---|
JP2016532186A (en) | 2016-10-13 |
JP6117438B2 (en) | 2017-04-19 |
US20160261734A1 (en) | 2016-09-08 |
IL249974A0 (en) | 2017-03-30 |
KR20160032173A (en) | 2016-03-23 |
CA2955848C (en) | 2018-10-09 |
IL249974B (en) | 2018-02-28 |
US20150019153A1 (en) | 2015-01-15 |
MX2016000428A (en) | 2016-10-26 |
AU2017201926A1 (en) | 2017-04-13 |
CA2955848A1 (en) | 2015-01-15 |
CN105531645B (en) | 2017-07-14 |
CA2917970A1 (en) | 2015-01-15 |
US10582038B2 (en) | 2020-03-03 |
EP3019934A1 (en) | 2016-05-18 |
AU2014287240B2 (en) | 2017-04-13 |
JP6416962B2 (en) | 2018-10-31 |
US10742798B1 (en) | 2020-08-11 |
AU2017201926B2 (en) | 2019-05-16 |
JP2017182800A (en) | 2017-10-05 |
IL243505A (en) | 2017-02-28 |
EP3019934A4 (en) | 2017-03-15 |
EP3019934B1 (en) | 2019-09-04 |
MX349635B (en) | 2017-08-07 |
US9372103B2 (en) | 2016-06-21 |
KR102068541B1 (en) | 2020-01-22 |
CN107422790B (en) | 2020-07-07 |
CA2917970C (en) | 2017-03-07 |
AU2014287240A1 (en) | 2016-03-03 |
CN107422790A (en) | 2017-12-01 |
CN105531645A (en) | 2016-04-27 |
KR20170039765A (en) | 2017-04-11 |
KR101726825B1 (en) | 2017-04-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10742798B1 (en) | Calibration of grab detection | |
CA2917974C (en) | Multi-sensor hand detection | |
AU2014287167B2 (en) | Isolating mobile device electrode |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480050432.1 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14823297 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 243505 Country of ref document: IL |
|
ENP | Entry into the national phase |
Ref document number: 2917970 Country of ref document: CA Ref document number: 2016525473 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2016/000428 Country of ref document: MX |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014823297 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112016000648 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 20167003682 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2014287240 Country of ref document: AU Date of ref document: 20140710 Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 249974 Country of ref document: IL |
|
ENP | Entry into the national phase |
Ref document number: 112016000648 Country of ref document: BR Kind code of ref document: A2 Effective date: 20160112 |