CN105074626B - Detection and response to extras touch event - Google Patents
Detection and response to extras touch event Download PDFInfo
- Publication number
- CN105074626B CN105074626B CN201480019431.0A CN201480019431A CN105074626B CN 105074626 B CN105074626 B CN 105074626B CN 201480019431 A CN201480019431 A CN 201480019431A CN 105074626 B CN105074626 B CN 105074626B
- Authority
- CN
- China
- Prior art keywords
- touch event
- extras
- mobile device
- movement
- touch
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1694—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a single or a set of motion sensors for pointer control or gesture input obtained by sensing movements of the portable computer
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The technology for allowing that extras touch event (such as on the surface around mobile device hand strike or finger or pencil tapping) is detected and positioned is described herein.The application executed in the operating system of mobile device or the equipment can take various movements based on the position of touch event.Detection device is mobile and in response to touch event, and mobile device sensor (such as accelerometer) detection has occurred and that touch event.In some embodiments, touch event feature vector is generated according to sensor reading, and the touch event feature vector is passed to the classifier of determining touch event position.May further determine that other touch event information, such as timing, intensity and can cause touch event object profile.Detection and positioning to touch event allow the interaction area of mobile device to extend beyond the physical boundary of the equipment, and multiple users is allowed to use mobile device in a manner of more convenient simultaneously.
Description
Background technique
The size of the interactive space of most of current mobile devices is restricted.For current tablet computer and intelligence
Energy phone, interactive space are generally limited by its touch screen and control button.
Detailed description of the invention
Fig. 1 shows detectable extras touch event and can execute the example of different movements according to the position of touch event
Property mobile device.
Fig. 2 shows detectable extras touch event and different movements can be executed according to the position of touch event and timing
EXEMPLARY MOBILE DEVICE.
Fig. 3 is the flow chart for the illustrative methods for detecting extras touch event and determining its position.
Fig. 4 is the block diagram that can determine the exemplary computer device of position of extras touch event.
Fig. 5 is the position for determining the extras touch event at mobile device and the illustrative methods of execution movement
Flow chart.
Fig. 6 is the block diagram that the exemplary computer device of technology described herein can be achieved.
Fig. 7 is the example of the executable computer executable instructions as a part for realizing technology described herein
The block diagram of property processor core.
Specific embodiment
Technology disclosed herein makes the interactive space of mobile computing device extend beyond the physical boundary of the equipment.It rings
Should be in extras touch event, the touch event that can occur near the equipment, the sensor being built into the equipment detects
Movement of the equipment on the surface that the equipment is just stopping.Movement detected can be used at least roughly determining touch event
Position.For example, mobile phone can detect finger tapping on the surface around phone of user using inbuilt accelerometers, determine
The position of tapping, and where foundation occurs tapping and executes movement.
The extension of the interactive space of equipment allows user using its whole hand or both hands rather than is only used only one or several
A finger (finger is usually only interacted with touch screen or control button) interacts with the equipment.For example, on a cellular telephone
The music application of execution may be in response to user and strike his hand on the surface around the equipment and play idiophonic sound.
Where the application can beat his or her hand according to user and play different sound.Interactive space through extending also allows more
Multi-user interacts with individual equipment simultaneously, or multiple users is at least allowed conveniently to interact with individual equipment simultaneously.For example,
Consider the question-and-answer game executed on tablet computer, the surface which can be neighbouring in plate based on which user
His hand of upper tapping and which of detect several users and be given the chance answered a question.If having to crowd around with user
Around the plate by with the touch screen interaction of plate come compared with participating in, user may more facilitate with this game interaction.
Referring now to the drawings, wherein similar appended drawing reference is used to quote similar element from beginning to end.It is described below
In, for illustrative purposes, a large amount of details are elaborated to provide the thorough understanding to it, however, can not have these
Novel embodiment is practiced in the case where detail can be obviously.In other instances, show in block diagram form
Well-known structure and equipment are out in order to description of them.Be intended that covering scope of the claims in all modifications,
Equivalent and replacement.
As used herein, term " touch event " and " exceptional space touch event " refer to that object is set with movement
The interaction on the surface where standby neighbouring equipment.Object can be the finger of such as user or whole hand, pencil, drumstick or refer to
Show pen, and surface can be such as table top, desk or floor.Touch event can be various interactions as a result, such as
Finger or the pencil palm of tapping or opening on desk tap on table top.If the detectable touching of the sensor of mobile device
Event is touched, then the touch event occurs near the equipment.That is, if touch event causes the detectable movement of mobile device,
Then the touch event is near the equipment.Correspondingly, mobile device nearby can (such as the equipment just stopping according to various factors
The material of platform, the object for creating touch event, object tap the intensity on surface and the sensitivity of movable sensor)
And change.
Fig. 1 shows detectable extras touch event and can execute the example of different movements according to the position of touch event
Property mobile device 100.Mobile computing device (mobile device) 100 can be any kind of movement described herein and set
It is standby, such as mobile phone (as shown in the figure), plate or laptop computer.Mobile device 100 is just resting on the surface 110 of platform
And it is carrying out electronic drum application, which applies and beat the surface near the equipment in response to user and play various beat
The sound of percussion instrument.When the hand 120 or 130 of user beats surface 110, creation can cause 140 He of vibration of the movement of equipment 100
150, which is detected by the built-in sensors of the equipment.
Sensor generates the output of the movement of instruction mobile device.Typically, sensor includes accelerometer, the acceleration family planning
At the output for reflecting acceleration of the equipment on one or more planes (for example, X, Y and Z plane).In some embodiments, it moves
Dynamic equipment may include detecting the more than one accelerometer of touch event.In the other embodiments still having, gyroscope can be with one
A or multiple accelerometers are used in combination to detect touch event, and in some embodiments, gyroscope provides indicating equipment and surrounds one
Or the output of the angular orientation of multiple axis (for example, X, Y and Z axis), to realize the shake to equipment, the measurement fallen and rolled of inclining.
Mobile device 100 can execute different movements according to the position of touch event detected.For example, mobile device
Where 100 can tap neighbouring circumferential surface according to user and play different sound.For example, if hand is on surface
It is tapped on 110 and causes the small drum sound to be played, then peripheral region can be divided into multiple regions, and foundation by mobile device 100
Where hand taps and plays different small drum sound.In another embodiment, it can be played according to touch event position different
Idiophonic sound.For example, can be programmed to equipment 100, so that the left hand in response to user beats the equipment left side
Surface, can play army's drum sound, and if the right hand of user beats the surface on the right of the equipment, can play ground flop flop make an uproar
Sound.Region around mobile device 100 can be divided into any number of subregion, and any shape can be used in each subregion,
As long as being distinguished between the touch event that user equipment can occur in individual subregion.For example, tool there are two with
The splitting scheme of upper subregion allows user to pretend him or she and beats a whole set of frame drum, wherein according to user on the surface 110 where
It beats his or her hand and plays and step on small cymbals, fried small cymbals, army's drum, various flop bulging and pucking sound.
The position of identified touch event may include for example relative to the direction of mobile device and/or away from mobile device
Distance.For example, the direction can indicate touch event from mobile device particular edge (for example, top, the following, left side,
The right) most nearby occur, or occur at specific angular coordinate.For example, with reference to Fig. 1, the direction of hand 120 can be confirmed as
In left the half of peripheral region or left lower quadrant, at 200 degree at 7 positions or in angular coordinate system.The side of hand 130
To can be determined to be in right half, right lower quadrant, at 4 positions or at 315 degree.As a result, in some embodiments, should
Direction can be based at least partially on the equipment how its peripheral region of subregion.
Fig. 2 shows detectable extras touch event and different movements can be executed according to the position of touch event and timing
EXEMPLARY MOBILE DEVICE 200.Calculating equipment 200 can be tablet computer or any other movement described herein
Equipment is carrying out question-and-answer game application, beats surface 210 with hand (for example, hand 220,221,222 or 223) wherein giving
The chance that first user answers a question.By making interaction surface extend beyond the touch screen 230 of mobile device 200, it is possible to increase
Add the quantity for the player that can play multi-player gaming on the mobile apparatus or at least makes multiple users' object for appreciation game more convenient.Pass through letter
Singly tapping the surface 210 near the equipment can interact with equipment 200, and multiple users do not need to crowd around at touch screen 230 weeks
It encloses to play game.
Fig. 3 is the flow chart for the illustrative methods 300 for detecting extras touch event and determining its position.Method 300
It can be operated (such as in software application grade or operating system grade) in any rank of mobile device.Typically, method 300
It operates in operating system grade, and is communicated with software application and/or operating system component.
310 are acted in processing, sensor values is read in the buffer (or other memory constructions or component) of the equipment.
Although method 300 shows the sensor values as continued to read, as finally returned that indicated by the path in 300 flow chart
, as long as the reading of sensor values is realized by application or operating system component, sensor values can synchronously, asynchronously or continuously
It reads on ground.One or more nearest sensor values can be stored in buffer, and sensor values may include indicating that this sets
The standby acceleration in one or more planes or the equipment surround the value of the acceleration of one or more axis.
320 are acted in processing, determines whether sensor values indicates the generation of touch event.Various methods can be used for determining biography
Whether sensor value indicates touch event.For example, in some embodiments, sensor values and touch event threshold value can be compared
Compared with wherein touch event is by being more than that one or more accelerometer readings of the threshold value indicate.It include multiple acceleration in sensor values
In the embodiment of plane or the value of multiple sensors, from less than all planes and/or less than the sensing of all sensors
Device reads the presence that can be used for determining touch event.
If sensor values indicates touch event, 330 are acted in processing, determines that touch event is extras touching
Touch event.If NO, then this method returns to 310 and reads new sensor values.Determine whether touch event additionally sets
Standby touch event may include determining user whether via with the equipment direct physical contact (such as via with touch screen
Interaction presses control button) input is supplied to the equipment.If touch event is confirmed as extras touch event,
340 then are acted in processing, method 300 calculates the feature vector of touch event and this feature vector is passed to classifier.If
Touch event is determined not to be extras touch event, then method 300 returns to 310 and reads new sensor values.?
In some embodiments of method 300, if the presence of sensor values instruction touch event, save processing movement 300 and square
Method 300 from 320 proceeds to 340.Processing movement 330 can be supplied directly to the situation of mobile device not expecting user's input for example
In save.
340 are acted in processing, the vector including one or more touch event features is determined according to sensor values.Touch thing
Part feature vector may include for example being obtained and carrying out Fourier transform (such as Fast Fourier Transform) to sensor values
Frequency domain information, minimum value and maximum value etc..Touch event feature vector may include time domain, frequency domain, other kinds of information or
Any combination thereof.
Once being determined, touch event feature vector is just delivered to classifier.In some embodiments, using engineering
Principle is practised to train classifier, wherein one group of training example (or vector) is supplied to classifier to train classifier can determine
The position of touch event.In some embodiments, classifier can be closest (kNN) classifier of k-.In kNN classifier,
The given input touch event feature vector for indicating query point, found in training data on closest to query point
K example.Then, it is voted based on the majority in k example and distributes class to input feature value.Such is identified touch thing
Part position, in some embodiments, if training set appropriate is already provided to classifier, to touch event feature vector
Classification can determine multiple while tapping position provided by multiple users.
Various methods can be used to create classifier.Shown by one group of training for providing known touch event to mobile device
Example creates classifier, this is referred to as the training classifier.Training set can be created in various ways.In one embodiment, will
Mobile device places on the surface and starts trained application.The application provides instructions to user, this apply upper, Yong Hu
Touch event is created on the surface of phone.Training example may include the touching occurred at the different location around mobile device
Touch event, the touch event with intensity, using a part of different objects (for example, pencil, stylus) or the hand of user
The touch event of (for example, hand of single, multiple fingers, fist, opening) creation is supplied directly to mobile device (for example, touching
Screen or control button) touch event, be supplied to the touch event of different platform (for example, wood, granite) and send out simultaneously
Raw multiple touch events.
In some embodiments, when mobile device is moved to different surfaces, can re -training classifier to allow this to set
It is standby more reliably to detect touch event.For example, and user his or her palmistry ratio is tapped on carpet, user is on wood desktop
It taps his or her hand and usually generates different accelerometer readings at mobile device.Equipment can prompt user to provide one group of training
Example so as to for example the equipment detect it be moved into new surface when, starting utilize touch event position application
When or user request when re -training classifier.
Processing act 350, by the position of the one or more touch event such as classifier determined by pass to apply or
Operating system component.
In some embodiments, equipment can determine the intensity of touch event, and reflection such as user uses object or body
Position taps the dynamics on the surface near the equipment.In these embodiments, executable at least partly using (or operating system)
Movement of the ground according to touch event intensity.Alternatively, the characteristic for the movement that the application is taken may depend on touch event intensity.Example
Such as, in music application example, the volume for playing musical instrument sound may be based partly on touch event intensity.In another example, exist
In game application, the intensity that user taps surface may make game role to take different movements (for example, if intensity is a certain
Hereinafter, then the role can jump, and if the intensity is more than threshold value, which can take off threshold value).
In some embodiments, which can determine the profile of the object of creation touch event.The profile can be for creating
It is distinguished between the various objects for the touch event built in one group of training example.In these embodiments, using (or operation be
System) can movement at least partly be executed according to contour of object.Alternatively, the characteristic for the movement that the application is taken can be according to object wheel
It is wide.For example, music application example is returned to, if user uses single compared with user taps surface using the palm that he opens
Finger tapping surface may make the different tones or sound of the application plays percussion instrument (such as small drum).Similarly, if divided
Class device can distinguish between the hand or drumstick (or pencil, pen or stylus) of user, then music application can make difference
Musical instrument played, such as if touch event is caused by drumstick, pencil, pen or stylus the army of broadcasting drum or flop drum,
And small drum is played if touch event is caused by the hand or finger touch-surface of user.
In some embodiments, touch event position is determined with rough precision.For example, being swum in previously described question and answer
In Application Example of playing, if four players are playing game, mobile device, which is only required to determine that touch event is located at, encloses
Which quadrant around the surface of the equipment.It, can be by less training example as a result, in the embodiment for only needing rough precision
Mobile device is supplied to train classifier.It, can be by biggish one in the embodiment of relatively precision accuracy for requiring touch location
Group training example is supplied to mobile device for classifier training.
In some embodiments, technology herein can determine the clocking information of touch event.Clocking information may include appointing
The clocking information of what type, such as absolute time (for example, the date, when, minute, second), relative to another touch event or equipment
Or the time (for example, in question-and-answer game, time) from there is a upper problem of application affairs.
Technology described herein may further determine that the clocking information of multiple overlapping touch events.That is, touch event when
Between upper enough near-earths occur so that accelerometer can be still presented when it starts to respond subsequent touch event in preceding touching
Touch the response of event.In some embodiments, the first touch event occur in a series of multiple overlapping touch events when
Between can be confirmed as sensor reading for the first time and be more than time of touch event threshold value.Since the sensor reading of touch event is usual
Rapidly drop to touch event threshold value hereinafter, the time that therefore second touch event Chong Die with the first touch event occurs can
To be time that sensor reading is more than touch event threshold value next time.
Fig. 4 is the block diagram that can determine the exemplary computer device 400 of the position of extras touch event.Calculate equipment
400 include one or more sensors 410, buffer 420, one or more application 430 and operating system 440.One or
Multiple applications 430 and operating system 440 operate on calculating equipment 400.One or more sensors 410 may include one or more
The mobile other sensors of a accelerometer, gyroscope or detectable equipment.Buffer 420 stores one or more sensors
410 values generated.In other embodiments, sensor values can be stored in mobile device 400 or by mobile device 400
In other accessible memory constructions (such as cache memory).It may include utilizing touch event position using 430
One or more application.
Operating system 440 includes touch event determining module 450, vector generation module 460, classifier 470 and report
Module 480.It is attached in calculating equipment to determine that touch event determining module 450 assesses the sensor values being stored in buffer 420
Closely whether there is touch event.Vector generation module 460 generates the touch event feature vector of one group of sensor values.The vector
It may include the feature of sensor values, minimum, maximum, centre and average sensor values and frequency domain in such as certain window
Information.Classifier 470 can be used for determining the position of touch event according to touch event feature vector.For example, classifier can be
Based on position associated with the top n example being most closely matched with supplied touch event feature vector in training class
Determine the kNN classifier of touch event position.Classifier 470 can be provided by the component and/or application 430 of operating system 440
One group of training example training.Reporting modules 480 can be reported to the other assemblies of application 430 or operating system 440 about touching
Touch the information of event, such as touch event position, clocking information, intensity, object for creating touch event profile.
It should be understood that Fig. 4 shows an example of the one group of module and component that can be included in mobile computing device.?
In other embodiments, mobile device can have module or component more than module or component shown in Fig. 4 or few.For example,
Mobile device may include multiple classifiers corresponding with different types of surface (for example, wood, granite).In some implementations
In example, the mobile device including multiple classifiers corresponding with different surfaces type may include the mould for automatically selecting classifier
Block.For example, the equipment can determine the type on surface by image captured by analysis built-in camera come selection sort device.Another
In one example, equipment 600 can will be led to by instructing user to carry out one or more specific tappings on desk come selection sort device
The training touch event feature vector for crossing tapping generation is compared from the training vector for corresponding to different stored classifiers.
In these embodiments, mobile device 600 may not need to collect one group of new instruction when it is moved to different types of surface
Practice example.
In addition, individual module or individual module shown in Fig. 4 can be combined into shown in Fig. 4 for isolated module
It may be logically divided into multiple modules.In addition, being illustrated as any of the module in the operating system for calculating equipment 400 in Fig. 4
It can be operated in another software layer (such as application layer).In addition, module shown in Fig. 4 can with software, hardware, firmware or
A combination thereof is realized.Refer to for be programmed to perform the calculating equipment of method can be programmed to via software, hardware, firmware or
A combination thereof executes this method.
Fig. 5 is the position for determining the extras touch event at mobile device and the illustrative methods of execution movement
500 flow chart.Method 500 can be executed by the tablet computer for for example executing the question-and-answer game that four players are playing.Player
The region that surface near the plate for distributing to them must be beaten, is answered a question with " coming in ".Plate is detected using accelerometer
The movement of plate.
Act 510 in processing, the movement based on such as one or more mobile device sensor mobile device detected come
Determine the position of the extras touch event occurred near mobile device.The position includes the side relative to mobile device
To.In this example, problem is proposed to only No. 3 players to know the answer.No. 3 players tap after reading problem distributes to him
Desk surface.Plate accelerometer taps desk in response to the hand of player and detects the movement of plate, and determination is distributing
To the position of the touch event occurred in the equipment near zone of No. 3 players.
520 are acted in processing, the position is based at least partially on and determines the movement executed at mobile device.In the example
In, No. 3 players chance responded to problem is given in plate determination.
530 are acted in processing, the movement is executed at mobile device.In this example, tablet computer inquires No. 3 players
It answers a question.
Technology described herein at least has the advantages that following exemplary.Pass through the touch thing near detection mobile device
Part allows user to interact with mobile device, and the interactive space of the equipment extends beyond the physical boundary of the equipment.In addition, institute is public
The technology opened extends the type of supported interaction.Typically, user is via single finger or the interaction of more fingers and mobile device
Interaction (for example, typewrite on hard or soft keyboard, supply single finger or multi-finger gesture to touch screen), but it is described herein
Technology can detect the interaction of whole hand, such as user taps the hand opened or fist of fiercelying attack on the surface.In addition, disclosed
Technology allows multiple users to interact simultaneously with mobile device, or multiple users is at least allowed conveniently to hand over mobile device
Mutually.As described above, multiple users, which can be gathered in around mobile device, plays game, without crowd around around the equipment with intelligence
The interaction of the limited touch area of energy phone or tablet computer.In addition, disclosed technology provide low cost solution with
The interactive space of extension device.It is modified to current mobile device to realize that it is soft that technology described herein may only need
Part modification, because many current mobile devices have possessed the mobile inbuilt accelerometers and gyroscope of detection device.In addition, passing through
It can be fallen on surface in the hand for making them or other objects with the surface interaction near the equipment, user there is no concern that making a mess of this
Equipment.
Technology, skill and embodiment described herein can be executed by any one of various calculating equipment, these meters
Calculating equipment includes mobile device (such as smart phone, handheld computer, tablet computer, laptop computer, media play
Device, portable game controller, camera and video recorder), non-mobile device (such as desktop computer, server, solid
Determine game console, smart television) and embedded device (being such as integrated to the equipment in vehicle).As used herein
, term " calculating equipment " includes computing system, and including the equipment with multiple discrete physical assemblies.
Fig. 6 is the block diagram that the exemplary computer device of technology described herein can be achieved.In general, shown in fig. 6 group
Part can be communicated with component shown in other, but for the ease of explaining, all connections are not shown.Equipment 600 be include first
The multicomputer system of processor 602 and second processor 604, and be explained as including point-to-point (P-P) interconnection.For example,
Point-to-point (P-P) interface 606 of processor 602 is coupled to the point-to-point interface 607 of processor 604 via point-to-point interconnection 605.
It should be appreciated that any or all point-to-point interconnection illustrated in fig. 6 is alternatively embodied as multi-point bus, and shown in Fig. 6
Any or all bus out can be substituted by point-to-point interconnection.
As shown in fig. 6, processor 602 and 604 is multi-core processor.Processor 602 includes processor core 608 and 609, and
Processor 604 includes processor core 610 and 611.Processor core 608-611 can be similar to the side discussed below in conjunction with Fig. 7
Formula otherwise executes computer executable instructions.
Processor 602 and 604 further comprises at least one shared cache memory 612 and 614 respectively.It is shared
Cache 612 and 614 can store one or more components (such as processor core 608-609 and 610- of the processor
611) data (for example, instruction) utilized.Shared cache 612 and 614 can be the storage levels of equipment 600
(hierarchy) a part.For example, shared cache 612 can be also stored in memory 616 being locally stored
Data are to allow the component of processor 602 quickly to access these data.In some embodiments, shared cache 612
It may include multiple cache layers with 614, such as grade 1 (L1), grade 2 (L2), grade 3 (L3), grade 4 (L4), and/or other high speeds are slow
It deposits or cache layer (such as last level cache (LLC)).
Although equipment 600 is shown to have two processors, equipment 600 may include any number of processor.This
Outside, processor may include any number of processor core.Various forms, such as central processing unit, control can be used in processor
Device, graphics processor, accelerometer (such as figure accelerometer or digital signal processor (DSP)) or field programmable gate array
(FPGA).Processor in equipment can be identical or different with other processors in the equipment.In some embodiments, equipment
600 may include and first processor, accelerometer, FPGA or any other processor isomery or asymmetric one or more
Processor.In terms of including the range of advantage measurement of framework, micro-architecture, heat, power consumption characteristics etc., processing element in systems
Between each species diversity may be present.These differences can effectively make to manifest itself by asymmetric and different between the processor in system
Structure.In some embodiments, processor 602 and 604 resides in identical die package.
Processor 602 and 604 further comprises Memory Controller logic (MC) 620 and 622.As shown in fig. 6, MC 620
Control the memory 616 and 618 for being coupled to processor 602 and 604 respectively with 640.Memory 616 and 618 may include various types of
The memory of type, such as volatile memory are (for example, dynamic random access memory (DRAM), static random access memory
Or nonvolatile memory (for example, flash memory) (SRAM)).Although MC 620 and 622 is shown as being integrated into 602 He of processor
In 604, but in an alternate embodiment, these MC can be the logic outside processor, and may include memory hierarchy
One or more layers.
Processor 602 and 604 is coupled to input/output (I/O) subsystem 630 via P-P interconnection 632 and 634.It is point-to-point
The point-to-point interface 636 of processor 602 is connected by interconnection 632 with the point-to-point interface 638 of I/O subsystem 630, and point-to-point
The point-to-point interface 640 of processor 604 is connected by interconnection 634 with the point-to-point interface 642 of I/O subsystem 630.Input/defeated
Subsystem 630 further comprises the interface 650 for making I/O subsystem 630 be coupled to graphics engine 652 out, which can be with
It is high performance graphics engine.I/O subsystem 630 and graphics engine 652 are coupled via bus 654.Alternatively, bus 654 can be with
It is point-to-point interconnection.
Input/output subsystem 630 is further coupled to the first bus 660 via interface 662.First bus 660 can be with
It is peripheral component interconnection (PCI) bus, PCI high-speed bus, another third generation I/O interconnection bus or any other type
Bus.
Various I/O equipment 664 can be coupled to the first bus 660.First bus 660 can be coupled to second by bus bridge 670
Bus 680.In some embodiments, the second bus 680 can be low pin count (LPC) bus.Various equipment can be coupled to
Two lines bus 680, these equipment include such as keyboard/mouse 682, audio I/O equipment 688 and store equipment 690, such as firmly
Dish driving, solid-state driving or other storage equipment for storing computer executable instructions (code) 692.Code 692 can
Including the computer executable instructions for executing technology described herein.It can be coupled to additional group of the second bus 680
Part includes one or more communication equipments 684, the communication equipment can via use one or more communication standards (for example, IEEE
802.11 standards and its supplement) one or more wired or wireless communication links (for example, electric wire, cable, Ethernet connection,
Radio frequency (RF) channel, infrared channel, Wi-Fi channel) provide equipment 600 and one or more wired or wireless networks 686 (example
Such as, Wi-Fi, honeycomb or satellite network) between communication.In addition, equipment 600 may include the one of the movement of detectable equipment 600
A or multiple sensors 689, such as accelerometer or gyroscope.
Equipment 600 may include removable memory, such as flash card (for example, SD (secure digital) blocks), memory stick, subscriber
Identity module (SIM) card.Memory (including cache 612 and 614, memory 616 and 618, Yi Jicun in equipment 600
Storage equipment 690) data and/or computer executable instructions for executing operating system 694 Yu application program 696 can be stored.
Sample data includes that equipment 600 is sent to and/or received from multiple network services via one or more wired or wireless networks
One of device or other equipment or the webpage used for equipment 600, text message, image, audio files, video data, classification
Device training data or other data sets.Equipment 600 may also access external memory (not shown), such as external hard-drive or
Person's storage based on cloud.
Operating system 694 can control the distribution and use of component illustrated in fig. 6, and support one or more application journey
Sequence 696.Operating system 694 may include collecting reading from one or more sensors, determining whether sensor reading indicates
The touch event in region around the equipment, the feature vector for determining touch event, the component for determining touch event position, with
And one or more classifiers 697 for these positions to be supplied to application program 696.Various classifiers can be included in and set
In standby 600, in terms of and for example various types of surfaces and the configuration of various sensors.For example, if mobile device 600 includes
Two accelerometers, then the first classifier can be used for providing touch event position from single accelerometer for the data collected, and second
Classifier can provide position for the data collected at two accelerometers.
Application program 696 may include that common moving calculates equipment application (for example, e-mail applications, calendar, contact person's pipe
Manage device, web browser, information receiving and transmitting application) and other calculating applications, it such as utilizes in the region around equipment 600
The detection of the touch event of generation and its application of position.
Equipment 600 can support various input equipments (such as touch screen, microphone, camera, physical keyboard and trace ball) with
And one or more output equipments (such as loudspeaker and display).Other possible input and output devices include piezoelectricity and its
His tactile I/O equipment.Input or any of output equipment can inside equipment 600, it is external or removable with equipment 600
Except ground attaches.External input and output equipment can be communicated via wired or wireless connection with equipment 600.
In addition, calculating equipment 600 can provide one or more natural user interfaces (UI).For example, operating system 694 or answering
Using 696 may include the speech recognition logic as a part of Voice User Interface, the Voice User Interface allow user via
Speech command operation equipment 600.In addition, equipment 600 may include allowing user via the posture and equipment 600 of body, hand or face
Interactive input equipment and logic.For example, detectable and interpreting user gesture is to provide input to game application.
Equipment 600 can further comprise one or more communication components 684.Component 684 may include being coupled to one or more
The wireless communication components of a antenna are to support the communication between system 60 and external equipment.Wireless communication components can support various nothings
Line communication protocol and technology, such as near-field communication (NFC), Wi-Fi, bluetooth, 4G long term evolution (LTE), CDMA (CDMA),
Universal Mobile Telecommunications System (UMTS) and global system for mobile communications (GSM).In addition, radio modem can support with
For within single cellular network, between multiple cellular networks or mobile computing device and public switched telephone network (PSTN)
Between data and voice communication one or more cellular networks communication.
Equipment 600 can further comprise that (input/output end port can be for example at least one input/output end port
USB, IEEE 1394 (firewire), Ethernet, and/or the port RS-232), including physical connector, power supply, Proximity Sensor,
Compass and receiver of satellite navigation system (such as GPS receiver).GPS receiver can be coupled to GPS antenna.Equipment 600 can
Further comprise one or more additional antennas, the additional antenna be coupled to one or more accessory receivers, transmitter and/
Or transceiver is to realize additional function.
It should be appreciated that Fig. 6 shows only one exemplary computer device framework.Calculating equipment based on replacement framework can be used for
Realize technology described herein.For example, being located at discrete integrated electricity instead of processor 602 and 604 and graphics engine 652
On the road, calculating equipment may include that in conjunction with multiple processors, graphics engine and add-on assemble SoC (system on chip) is integrated
Circuit.It can be via different from bus shown in fig. 6 or point-to-point configuration connecting element in addition, calculating equipment.Moreover, in Fig. 6
Shown component is not required or is included entirely, because component and can add shown in can be removed in an alternate embodiment
Add other assemblies.
Fig. 7 is performed as realizing the exemplary of the computer executable instructions of a part of technology described herein
The block diagram of processor core 700.Processor core 700 can be any kind of processor (such as microprocessor, embedded processing
Device, digital signal processor (DSP) or network processing unit) core.Processor core 700 can be single thread core or multithreading
Core because it can each core include more than one hardware thread contexts (or " logic processor ").
The memory 710 for being coupled to processor 700 is also shown in Fig. 7.Memory 710 can be described herein any
Memory or any other memory well known by persons skilled in the art.Memory 710, which can store, to be held by processor core 700
Capable computer executable instructions 715 (code).
The processor core includes the front end logic 720 that instruction is received from memory 710.Instruction can be decoded by one or more
The processing of device 730.Decoder 730 produces microoperation (such as using the fixed width microoperation of predefined format) and exports as it,
Or generate other instructions, microcommand or the control signal of reflection original code instruction.Front end logic 720 further comprises posting
Storage renames logic 735 and scheduling logic 740, which usually distributes resource and to the finger with conversion for execution
Corresponding operation is enabled to be lined up.
Processor core 700 further comprises executing logic 750, which includes one or more execution units
(EU) 765-1 to 765-N.Some processor core embodiments may include specific function or multiple function is dedicated multiple executes list
Member.Other embodiments may only include an execution unit or may include an execution unit of executable specific function.It holds
Row logic 750 executes operation specified by code command.After executing operation specified by code command and completing, back-end logic
770 make Retirement using resignation (retirement) logic 775.In some embodiments, processor core 700 allows unordered
It executes, but requires that instruction is made to retire from office in an orderly manner.Various forms well known by persons skilled in the art can be used in retirement logic 770
(for example, resequencing buffer etc.).
Processor core 700 is during executing instruction at least in the output generated of decoder 730, register renaming logic
735 using hardware register and table and become in terms of executing any register (not shown) modified of logic 750
It changes.
Although being not shown in Fig. 7, processor may include its being located at together in integrated chip with processor core 700
His element.For example, processor may include add ons, such as store control logic, one or more graphics engines, I/O control
Logic processed, and/or one or more caches.
Referring back to Fig. 6, network 686 can provide the various clothes based on cloud that can be used for realizing technology described herein
Business.For example, for training the training example of classifier that can be provided by service based on cloud for common surface.In some embodiments
In, touch feature vector generation can locally executed at mobile device, and be subsequently transmitted to it is based on cloud service for
Position determines.That is, classifier can be realized remotely.
Any one of disclosed method can be implemented as computer executable instructions or computer program product.This finger
Order can make computer execute any one of disclosed method.In general, as used herein, term " computer " refers to herein
Mentioned in any calculating equipment or system or any other calculate equipment.Term " computer executable instructions " as a result,
Refer to the instruction that can be executed by any calculating equipment described herein and mentioned or any other calculating equipment.
It computer executable instructions or computer program product and during realizing disclosed technology creates and uses
Any data can be stored in one or more tangible computer readable storage mediums (such as optical medium disk (for example, DVD, CD),
Volatile memory component (for example, DRAM, SRAM) or non-volatile storage component (for example, flash memory, dish driving)) on.
Computer readable storage medium can be comprised in computer readable storage medium (such as solid-state driving, USB flash memory driving and
Memory module) on.Alternatively, computer executable instructions can be by comprising for executing all or part of disclosed method
Hard wire logic specific hardware components or by any combination of computer readable storage medium or hardware component execute.
Computer-readable instruction can be such as special-purpose software application or via web browser or other software application
A part of the software application of (such as remote computation application) access.For example, can be on a single computing device or in network rings
This software is executed in border using one or more network computers.It is further understood that disclosed technology is not limited to appoint
What certain computer language or program.For example, can by with C++, Java, Perl, JavaScript, Adobe Flash,
Or any other suitable programming language software for writing realizes disclosed technology.Equally, disclosed technology is unlimited
In any certain computer or any kind of hardware.Certain details of suitable computer and hardware are known, and
It does not need to be set forth in the disclosure.
Moreover, software-based embodiment can be updated by suitable means of communication, downloaded or remotely be accessed
Any of (including for example for making computer execute the computer executable instructions of any one of disclosed method).This
A little suitable means of communication include that such as internet, WWW, Intranet, cable (including fibre-optic cable), magnetic communication, electromagnetism are logical
Believe (including RF, microwave and infrared communication), electronic communication or other such means of communication.
As in this application and used in the claims, the list of items that term "and/or" is added can refer to institute
Any combination of list of items.For example, phrase " A, B, and/or C " can refer to A, B, C, A and B, A and C, B and C or A, B and C.
As used in this application and in the claims, be added term " at least one of " a list of items can refer to it is listed
Any combination of project.For example, phrase " at least one of A, B or C " can refer to A, B, C, A and B, A and C, B and C, or
A, B and C.
Disclosed method, equipment and system must not be explained to be limited in any way.On the contrary, this disclosure relates to each
A disclosed embodiment independent and with all novel and non-obvious features of various combinations each other and sub-portfolio and
Aspect.Disclosed method, equipment and system is not limited to any particular aspects or feature or a combination thereof, disclosed implementation
Example does not require any one or more specific advantages there are problems that yet or as solution.
Theory of operation, the principles of science or other theories that equipment or method herein with reference to the disclosure are presented are retouched
It states to better understand and provides, and do not expect to be limited in range.Equipment and method in appended claims
Those of be not limited to work in a manner of described in these theory of operation equipment and method.
It is described although the operation of some disclosed methods shows for convenience with specific, sequence order,
It should be understood that the describing mode is covered and is rearranged, unless language-specific set forth herein requires particular sorted.For example,
It can rearrange or be executed concurrently in some cases the operation of sequence description.Moreover, for simplicity, appended attached drawing
The various modes that disclosed method can be used in combination with other methods may be not shown.
Following example is related to the additional embodiment of technology disclosed herein.
A kind of mobile device of example 1., comprising: one or more sensors;One or more classifiers, this or more
A classifier is based at least partially on the reading of one or more sensors to determine that is occurred near mobile device additionally sets
The position of standby touch event;And reporting modules, this report module is to the application report extras executed on the mobile apparatus
The position of touch event.
The mobile device of 2. example 1 of example, wherein the position of extras touch event includes relative to mobile device
Direction.
The mobile device of 3. example 1 of example, wherein the position of extras touch event includes at a distance from mobile device.
The mobile device of 4. example 1 of example, wherein one or more sensors include at least one accelerometer.
The mobile device of 5. example 1 of example, wherein one or more sensors include at least one gyroscope.
The mobile device of 6. example 1 of example, wherein the application, which can be based at least partially on reporting modules and be supplied to this, answers
Extras touch event position executes movement.
The mobile device of 7. example 6 of example, if wherein the application is further able to extras touch event position and exists
Then executed in first area near mobile device the first movement and if extras touch event position in mobile device
The second movement is then executed in neighbouring second area, the first movement is different from the second movement.
The mobile device of 8. example 6 of example further comprises extras touch event determining module, extras touching
The clocking information that event determination module determines extras touch event is touched, reporting modules are further additionally set to the application report
The clocking information of standby touch event, the meter of extras touch event is further based in part on by the movement of the first application execution
When information.
The mobile device of 9. example 6 of example, at least one of these classifiers further determine that extras touch thing
The intensity of part, for reporting modules further to the intensity of the application report extras touch event, which can be at least partly
Ground is supplied to the extras touch event position of the application based on reporting modules and the intensity of extras touch events is held
Action is made.
The mobile device of 10. example 1 of example, at least one of one or more classifiers are further determined that for creating
The profile of the object of extras touch event, for reporting modules further to the profile of the application report object, which can
Be based at least partially on reporting modules be supplied to the application extras touch event position and object profile it is dynamic to execute
Make.
The mobile device of 11. example 1 of example further comprises: vector generation module, which is based on one
Or the reading of multiple sensors generates extras touch event feature vector so that one or more classifiers are for determining
The position of extras touch event;And extras touch event determining module, the extras touch event determine mould
Block determines whether the reading of one or more sensors indicates the generation of extras touch event.
The mobile device of 12. example 1 of example further comprises: vector generation module, which is based on one
Or the reading of multiple sensors generates extras touch event feature vector so that one or more classifiers are for determining
The position of extras touch event and the intensity of extras touch event;And extras touch event determining module,
The extras touch event determining module determines whether the reading of one or more sensors indicates extras touch event
Generation.
The mobile device of 13. example 1 of example further comprises: vector generation module, which is based on one
Or the reading of multiple sensors generates extras touch event feature vector so that one or more classifiers are for determining
The profile of the position of extras touch event and the object for creating extras touch event;And extras touch
Event determination module, the extras touch event determining module determine whether the reading of one or more sensors indicates additionally
The generation of equipment touch event.
A kind of method of example 14. detection and response extras touch event, this method comprises: based on such as one or more
The movement of a mobile device sensor mobile device detected determines that the extras that occur near mobile device touch
The position of event;It is based at least partially on the position determination movement to be executed;And the movement is executed at mobile device.
The method of 15. example 14 of example, wherein the position includes the direction relative to mobile device.
The method of 16. example 14 of example, wherein the position includes at a distance from mobile device.
The method of 17. example 14 of example, further comprises: determining the clocking information of extras touch event, wherein really
Surely the movement to be executed is based further on clocking information.
The method of 18. example 14 of example further comprises: determining the clocking information of extras touch event and is moving
The one or more that dynamic equipment nearby occurs adds the clocking information of extras touch event, wherein the determination movement to be executed
It is based further on the clocking information based in part on extras touch event and one or more additional extras
The clocking information of touch event, which of extras touch event and one or more additional extras touch events
Occur earliest.
The method of 19. example 14 of example, wherein the movement to be executed is one of multiple movements, if multiple movements include should
The first movement and if extras touch event is determined that position then executes in the first area near mobile device
For the second movement then executed in the second area near mobile device, first event is different from second event.
The method of 20. example 14 of example, further comprises: determining the intensity of extras touch event, wherein to execute
Movement be based further on the intensity of extras touch event.
The method of 21. example 14 of example further comprises: determining the wheel for creating the object of extras touch event
Exterior feature, wherein the movement to be executed is based further on the profile of object.
The method of 22. example 14 of example, wherein determining that position includes: to read one from one or more mobile device sensors
A or multiple sensor values;Determine the generation of one or more sensors value instruction extras touch event;And according to one
A or multiple sensor values generate extras touch event feature vector;
Wherein determine that position includes that extras touch event feature vector is supplied to classifier.
The method of 23. example 22 of example, wherein determining one or more sensors value instruction extras touch event
Comprise determining that at least one of one or more sensors value is more than extras touch event threshold value.
The method of 24. example 22 of example, wherein the classifier is one of multiple classifiers, and this method further comprises: base
Surface type near mobile device selects the classifier from multiple classifiers.
The method of 25. example 24 of example, is based in part on the image on the surface near mobile device wherein at least to determine
Surface type.
The method of 26. example 24 of example is wherein at least based in part on and carries out one on desk in response to instruction user
Or multiple specific tappings and generate training extras touch event feature vector, will training extras touch event feature
Vector is compared to determine surface type from the training vector for corresponding to different stored classifiers.
The one or more computers for executing any one of method described in 4-26 according to claim 1 of example 27. can
Read storage medium.
A kind of equipment of example 28., the device including executing any one of method described in 4-26 according to claim 1.
The one or more computer-readable mediums with the instruction stored thereon of example 29., these instructions are for making to move
Dynamic equipment executes method, this method comprises: the shifting based on such as one or more mobile device sensors mobile device detected
Move the position to determine the extras touch event occurred near mobile device;Position determination is based at least partially on to want
The movement of execution;And the movement is executed at mobile device.
One or more computer-readable mediums of 30. example 29 of example, wherein the position includes relative to mobile device
Direction.
One or more computer-readable mediums of 31. example 29 of example, wherein the position include with mobile device away from
From.
One or more computer-readable mediums of 32. example 29 of example further comprise: determining that extras touch thing
The clocking information of part, wherein the determination movement to be executed are based further on clocking information.
One or more computer-readable mediums of 33. example 29 of example further comprise: determining that extras touch thing
The clocking information of part and the one or more occurred near mobile device add the clocking information of extras touch event,
The wherein determination movement to be executed is based further on the clocking information and one based in part on extras touch event
Outside the clocking information of a or multiple additional extras touch events, extras touch event and one or more plus
Which of equipment touch event occurs earliest.
One or more computer-readable mediums of 34. example 29 of example, wherein the movement to be executed is multiple movements
One, it is multiple movement include executed if the position is in the first area near mobile device first movement and if
Extras touch event is determined to be in the second movement then executed in the second area near mobile device, and first event is not
It is same as second event.
One or more computer-readable mediums of 35. example 29 of example further comprise: determining that extras touch thing
The intensity of part, wherein the movement to be executed is based further on the intensity of extras touch event.
One or more computer-readable mediums of 36. example 29 of example further comprise: determination is additionally set for creating
The profile of the object of standby touch event, wherein the movement to be executed is based further on the profile of object.
One or more computer-readable mediums of 37. example 29 of example, wherein determining that position includes: from one or more
Mobile device sensor reads one or more sensors value;Determine that one or more sensors value instruction extras touch thing
The generation of part;And extras touch event feature vector is generated according to one or more sensors value;Wherein determine position
Including extras touch event feature vector is supplied to classifier.
One or more computer-readable mediums of 38. example 37 of example, wherein determining that one or more sensors value refers to
Show that extras touch event comprises determining that at least one of one or more sensors value is touched more than extras
Touch event threshold.
One or more computer-readable mediums of 39. example 37 of example, wherein the classifier is one of multiple classifiers,
This method further comprises: the classifier is selected from multiple classifiers based on the surface type near mobile device.
One or more computer-readable mediums of 40. example 39 of example, it is attached to be wherein at least based in part on mobile device
The image on close surface determines surface type.
A kind of equipment of example 41. includes: for being set based on such as one or more mobile device sensor movements detected
Standby movement determines the device of the position of the extras touch event occurred near mobile device;For at least partly
The device of the determination movement to be executed based on the position;And the device for executing the movement at mobile device.
The equipment of 42. example 41 of example further comprises: for determine extras touch event clocking information and
The one or more occurred near mobile device adds the device of the clocking information of extras touch event, wherein for true
Surely the device for the movement to be executed is based further on clocking information and one based in part on extras touch event
Outside the clocking information of a or multiple additional extras touch events, extras touch event and one or more plus
Which of equipment touch event occurs earliest.
The equipment of 43. example 41 of example, wherein the device for determining extras touch event position include: for from
One or more mobile device sensors read the device of one or more sensors value;For determining one or more sensors
The device of the generation of value instruction extras touch event;And for generating extras according to one or more sensors value
The device of touch event feature vector;Wherein determine that position includes that extras touch event feature vector is supplied to classification
Device.
Claims (25)
1. a kind of mobile device, comprising:
One or more sensors, it is described for detecting the movement of the mobile device as caused by extras touch event
Extras touch event is on the surface that the mobile device is placed and in the mobile device near zone;
One or more classifiers, one or more of classifiers are for being based at least partially on one or more of sensings
The position of device read to determine the extras touch event;And
Reporting modules, the reporting modules are used to touch to extras described in the application report executed on the mobile device
The position of event,
Wherein, the near zone can be divided into multiple subregions by the mobile device, and each subregion is in multiple users
One.
2. mobile device according to claim 1, which is characterized in that the position of the extras touch event includes phase
For the direction of the mobile device.
3. mobile device according to claim 1, which is characterized in that the position of the extras touch event include with
The distance of the mobile device.
4. mobile device according to claim 1, which is characterized in that the application can be based at least partially on the report
It accuses module and is supplied to the extras touch event position of the application to execute movement.
5. mobile device according to claim 4, which is characterized in that if the application is further able to described additionally set
The first movement is then executed in the first area near the mobile device for standby touch event position and if described additionally set
Standby touch event position then executes the second movement in the second area near the mobile device, and first movement is different from
Second movement.
6. mobile device according to claim 4, which is characterized in that further comprise that extras touch event determines mould
Block, the extras touch event determining module are used to determine the clocking information of the extras touch event, the report
Module is accused to be used to be held further to the clocking information of extras touch event described in the application report by the application
Capable movement is further based in part on the clocking information of the extras touch event.
7. mobile device according to claim 4, which is characterized in that at least one of described classifier is for further
Determine the intensity of the extras touch event, the reporting modules to described in the application report for further additionally setting
The intensity of standby touch event, the application can be based at least partially on the institute that the reporting modules are supplied to the application
The intensity of extras touch event position and the extras touch event is stated to execute movement.
8. mobile device according to claim 1, which is characterized in that at least one of one or more of classifiers
For further determining that the profile of the object for creating the extras touch event, the reporting modules are for further
To the profile of object described in the application report, the application can be based at least partially on the reporting modules and be supplied to
The profile of the extras touch event position of the application and the object executes movement.
9. mobile device according to claim 1, which is characterized in that further comprise:
Vector generation module, the vector generation module are generated for the reading based on one or more of sensors
Extras touch event feature vector is so that one or more of classifiers are for determining the extras touch event
The position;And
Extras touch event determining module, the extras touch event determining module are one or more for determining
Whether the reading of a sensor indicates the generation of extras touch event.
10. mobile device according to claim 1, which is characterized in that further comprise:
Vector generation module, the vector generation module are generated for the reading based on one or more of sensors
Extras touch event feature vector is so that one or more of classifiers are for determining the extras touch event
The position and the extras touch event intensity;And
Extras touch event determining module, the extras touch event determining module are one or more for determining
Whether the reading of a sensor indicates the generation of extras touch event.
11. mobile device according to claim 1, which is characterized in that further comprise:
Vector generation module, the vector generation module are generated for the reading based on one or more of sensors
Extras touch event feature vector is so that one or more of classifiers are for determining the extras touch event
The position and the object for creating the extras touch event profile;And
Extras touch event determining module, the extras touch event determining module are one or more for determining
Whether the reading of a sensor indicates the generation of extras touch event.
12. a kind of method of detection and response extras touch event, which comprises
Based on as caused by the extras touch event and by one or more mobile device sensor shiftings detected
The movement of dynamic equipment, the volume for determining on the surface that the mobile device is placed and occurring in the mobile device near zone
The position of external equipment touch event;
The near zone is divided into multiple subregions, each subregion is for one in multiple users;
It is based at least partially on the position, the determination movement to be executed;And
The movement is executed at the mobile device.
13. according to the method for claim 12, which is characterized in that the position includes the side relative to the mobile device
To and/or distance.
14. according to the method for claim 12, which is characterized in that further comprise: determining that the extras touch thing
The clocking information of part, wherein the determination movement to be executed is based further on the clocking information.
15. according to the method for claim 12, which is characterized in that further comprise: determining that the extras touch thing
The clocking information of part and the one or more occurred near the mobile device add the timing of extras touch event
Information, wherein the determination movement to be executed is based further on the institute based in part on the extras touch event
State the clocking information of clocking information and one or more of additional extras touch events, the extras touching
Which of event and one or more of additional extras touch events is touched to occur earliest.
16. according to the method for claim 12, which is characterized in that the movement to be executed is one of multiple movements, institute
State it is multiple movement include to be executed if the position is in the first area near the mobile device first movement, with
And to be executed if the extras touch event is determined to be in the second area near the mobile device
Two movements, first movement are different from second movement.
17. according to the method for claim 12, which is characterized in that further comprise: determining that the extras touch thing
The intensity of part, wherein the movement to be executed is based further on the intensity of the extras touch event.
18. according to the method for claim 12, which is characterized in that further comprise: determining and described additionally set for creating
The profile of the object of standby touch event, wherein the movement to be executed is based further on the profile of the object.
19. according to the method for claim 12, which is characterized in that the determination position includes:
One or more sensors value is read from one or more of mobile device sensors;
Determine the generation of one or more of sensor values instruction extras touch events;And
Extras touch event feature vector is generated according to one or more of sensor values;
Wherein the determination position includes that the extras touch event feature vector is supplied to classifier.
20. according to the method for claim 19, which is characterized in that the classifier is one of multiple classifiers, the side
Method further comprises: the classifier is selected from the multiple classifier based on the surface type near the mobile device.
21. according to the method for claim 20, which is characterized in that the table being based at least partially near the mobile device
The image in face determines the surface type.
22. according to the method for claim 20, which is characterized in that be based at least partially in response to instruction user in desk
The upper training extras touch event feature vector for carrying out one or more specific tappings and generating additionally sets the training
Standby touch event feature vector is compared to determine the surface from the training vector for corresponding to different stored classifiers
Type.
23. one or more computer-readable storage mediums for realizing any one of method described in 2-22 according to claim 1
Matter.
24. the equipment of a kind of detection and response extras touch event, including for executing described in 2-22 according to claim 1
Any one of method device.
25. the equipment of a kind of detection and response extras touch event, comprising:
For based on as caused by extras touch event and by such as one or more mobile device sensors shifting detected
The movement of equipment is moved to determine on the surface that the mobile device is placed and occur in the mobile device near zone
The device of the position of extras touch event;
For the near zone to be divided into the device of multiple subregions, each subregion is for one in multiple users;
The device for the movement to be executed is determined for being based at least partially on the position;And
For executing the device of the movement at the mobile device.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/874,905 US9448637B2 (en) | 2013-05-01 | 2013-05-01 | Detection of and response to extra-device touch events |
US13/874,905 | 2013-05-01 | ||
PCT/US2014/034809 WO2014179096A1 (en) | 2013-05-01 | 2014-04-21 | Detection of and response to extra-device touch events |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105074626A CN105074626A (en) | 2015-11-18 |
CN105074626B true CN105074626B (en) | 2019-03-05 |
Family
ID=51841213
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201480019431.0A Active CN105074626B (en) | 2013-05-01 | 2014-04-21 | Detection and response to extras touch event |
Country Status (3)
Country | Link |
---|---|
US (1) | US9448637B2 (en) |
CN (1) | CN105074626B (en) |
WO (1) | WO2014179096A1 (en) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9619036B2 (en) * | 2012-05-11 | 2017-04-11 | Comcast Cable Communications, Llc | System and methods for controlling a user experience |
US9355418B2 (en) | 2013-12-19 | 2016-05-31 | Twin Harbor Labs, LLC | Alerting servers using vibrational signals |
JP2018508865A (en) * | 2015-02-27 | 2018-03-29 | エントイット ソフトウェア エルエルシーEntit Software Llc | Application event tracking |
EP3093767A1 (en) * | 2015-05-12 | 2016-11-16 | Optofidelity OY | Method, apparatus and computer program product for testing a device |
TWI584239B (en) * | 2015-10-02 | 2017-05-21 | Jon Chao Hong | Electronic test system and answer device |
JP2018012188A (en) * | 2016-06-03 | 2018-01-25 | ファナック アメリカ コーポレイション | Dynamic laser touch sensing by multiple robots, and dynamic user coordinate system |
US10261685B2 (en) * | 2016-12-29 | 2019-04-16 | Google Llc | Multi-task machine learning for predicted touch interpretations |
WO2018119996A1 (en) * | 2016-12-30 | 2018-07-05 | Intel Corporation | Unification of classifier models across device platforms |
GB201718007D0 (en) * | 2017-10-31 | 2017-12-13 | Mogees Ltd | Trigger for game events |
EP3857337A1 (en) * | 2018-09-28 | 2021-08-04 | Snap Inc. | Neural network system for gesture, wear, activity, or carry detection on a wearable or mobile device |
US11580187B2 (en) * | 2020-12-11 | 2023-02-14 | Statepoint Media, Inc. | Method and system for localized content distribution for internet media |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1926496A (en) * | 2004-03-01 | 2007-03-07 | 苹果电脑有限公司 | Methods and apparatuses for operating a portable device based on an accelerometer |
CN102640086A (en) * | 2009-12-04 | 2012-08-15 | 微软公司 | Sensing mechanical energy to appropriate the body for data input |
CN102763057A (en) * | 2010-03-15 | 2012-10-31 | 日本电气株式会社 | Input device, input method and program |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100595925B1 (en) | 1998-01-26 | 2006-07-05 | 웨인 웨스터만 | Method and apparatus for integrating manual input |
US20070177804A1 (en) | 2006-01-30 | 2007-08-02 | Apple Computer, Inc. | Multi-touch gesture dictionary |
US20070254271A1 (en) * | 2006-04-28 | 2007-11-01 | Volodimir Burlik | Method, apparatus and software for play list selection in digital music players |
KR101615461B1 (en) * | 2007-09-24 | 2016-04-25 | 애플 인크. | Embedded authentication systems in an electronic device |
US8624836B1 (en) * | 2008-10-24 | 2014-01-07 | Google Inc. | Gesture-based small device input |
KR101179466B1 (en) | 2009-09-22 | 2012-09-07 | 에스케이플래닛 주식회사 | Mobile terminal and method for displaying object using approach sensing of touch tool thereof |
US8549418B2 (en) * | 2009-12-23 | 2013-10-01 | Intel Corporation | Projected display to enhance computer device use |
KR101743948B1 (en) | 2010-04-07 | 2017-06-21 | 삼성전자주식회사 | Method for hover sensing in the interactive display and method for processing hover sensing image |
US8954099B2 (en) | 2010-06-16 | 2015-02-10 | Qualcomm Incorporated | Layout design of proximity sensors to enable shortcuts |
US20120052944A1 (en) * | 2010-08-26 | 2012-03-01 | Mattel, Inc. | Time reaction game with vibration sensors |
US9479698B2 (en) * | 2012-12-17 | 2016-10-25 | Egos Ventures Inc. | Vibration-induced rotation of electronic devices |
-
2013
- 2013-05-01 US US13/874,905 patent/US9448637B2/en active Active
-
2014
- 2014-04-21 CN CN201480019431.0A patent/CN105074626B/en active Active
- 2014-04-21 WO PCT/US2014/034809 patent/WO2014179096A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1926496A (en) * | 2004-03-01 | 2007-03-07 | 苹果电脑有限公司 | Methods and apparatuses for operating a portable device based on an accelerometer |
CN102640086A (en) * | 2009-12-04 | 2012-08-15 | 微软公司 | Sensing mechanical energy to appropriate the body for data input |
CN102763057A (en) * | 2010-03-15 | 2012-10-31 | 日本电气株式会社 | Input device, input method and program |
Also Published As
Publication number | Publication date |
---|---|
US9448637B2 (en) | 2016-09-20 |
WO2014179096A1 (en) | 2014-11-06 |
CN105074626A (en) | 2015-11-18 |
US20140327655A1 (en) | 2014-11-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105074626B (en) | Detection and response to extras touch event | |
US11009958B2 (en) | Method and apparatus for providing sight independent activity reports responsive to a touch gesture | |
US9666173B2 (en) | Method for playing virtual musical instrument and electronic device for supporting the same | |
CN105283840B (en) | For synchronizing the equipment, method and graphic user interface of two or more displays | |
Gummeson et al. | An energy harvesting wearable ring platform for gestureinput on surfaces | |
EP2945045B1 (en) | Electronic device and method of playing music in electronic device | |
US9430106B1 (en) | Coordinated stylus haptic action | |
CN108700940A (en) | Scale of construction virtual reality keyboard method, user interface and interaction | |
CN109154859A (en) | For providing the equipment, method and graphic user interface of touch feedback | |
CN108139778A (en) | The screen display method of portable device and portable device | |
US20130257807A1 (en) | System and method for enhancing touch input | |
CN110308853A (en) | Equipment, method and graphic user interface for mobile user interface object | |
JP2017534132A (en) | Device for sharing user interaction | |
KR20160057407A (en) | Simultaneous hover and touch interface | |
CN104823198A (en) | Secure identification of computing device and secure identification methods | |
US8525780B2 (en) | Method and apparatus for inputting three-dimensional location | |
CN101384317B (en) | Trace information processing device, trace information processing method | |
US9395911B2 (en) | Computer input using hand drawn symbols | |
KR20140141089A (en) | Electronic device for executing application in response to pen input | |
CN109982129A (en) | Control method for playing back, device and the storage medium of short-sighted frequency | |
CN105183217A (en) | Touch display device and touch display method | |
CN107111441A (en) | Multi-stage user interface | |
CN104137026A (en) | Interactive drawing recognition | |
CN113515209A (en) | Music screening method, device, equipment and medium | |
CN108829329B (en) | Operation object display method and device and readable medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |