CN109254662A - Mobile device operation method, apparatus, computer equipment and storage medium - Google Patents
Mobile device operation method, apparatus, computer equipment and storage medium Download PDFInfo
- Publication number
- CN109254662A CN109254662A CN201811026907.XA CN201811026907A CN109254662A CN 109254662 A CN109254662 A CN 109254662A CN 201811026907 A CN201811026907 A CN 201811026907A CN 109254662 A CN109254662 A CN 109254662A
- Authority
- CN
- China
- Prior art keywords
- track
- eyeball
- user
- real time
- time kinematics
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
Abstract
The embodiment of the invention discloses a kind of mobile device operation method, apparatus, computer equipment and storage mediums, wherein the method includes obtaining the real time kinematics track of the eyeball of user according to Vision Tracking;Whether the acquired real time kinematics track of judgement matches with the standard movement track in preset track database, and the preset track database includes multiple and different standard movement track, and each standard movement track is associated with a corresponding screen operator instruction;If acquired real time kinematics track matches with the standard movement track in preset track database, screen operator instruction associated with the standard movement track is executed.The present invention can rapidly and accurately operating terminal screen, improve user to the experience in terms of the intelligence and safety of terminal.
Description
Technical field
The present invention relates to field of computer technology more particularly to a kind of mobile device operation method, apparatus, computer equipment
And storage medium.
Background technique
With making constant progress for mobile internet, various intelligent mobile terminals come into being, such as smart phone,
Tablet computer, smart television, smartwatch etc. have not only driven the various interconnections such as interaction, exchange, amusement and chat of people
Net online activity, and the mobile terminal based on touch operation mode greatly facilitates manipulation of the people to mobile terminal.Mesh
Before, people are to trigger key, triggering touch screen or the triggering control modes such as key and touch screen by finger to realize to intelligence
The operation of terminal however, the above-mentioned control mode operation to intelligent terminal is comparatively laborious, and reduces user terminal operations intelligence
Experience in terms of energy property and safety.
Summary of the invention
The embodiment of the present invention provides a kind of mobile device operation method, apparatus, computer equipment and storage medium, can be fast
Speed accurately operating terminal screen improves user to the experience in terms of the intelligence and safety of terminal.
In a first aspect, the embodiment of the invention provides a kind of mobile device operation methods, this method comprises:
The real time kinematics track of the eyeball of user is obtained according to Vision Tracking;
Whether the acquired real time kinematics track of judgement matches with the standard movement track in preset track database,
The preset track database includes multiple and different standard movement track, and it is corresponding that each standard movement track is associated with one
Screen operator instruction;
If acquired real time kinematics track matches with the standard movement track in preset track database, execute with
The associated screen operator instruction in standard movement track.
Second aspect, the embodiment of the invention also provides a kind of mobile device operation device, which includes for executing
The unit of the above method.
The third aspect, the embodiment of the invention also provides a kind of computer equipments comprising memory and processor, it is described
Computer program is stored on memory, the processor realizes the above method when executing the computer program.
Fourth aspect, the embodiment of the invention also provides a kind of computer readable storage medium, the storage medium storage
There is computer program, the computer program includes program instruction, and described program instruction can be realized when being executed by a processor
State method.
The embodiment of the invention provides a kind of mobile device operation method, apparatus, computer equipment and storage mediums.Its
In, the method includes the real time kinematics track of the eyeball of user is obtained according to Vision Tracking;It is real-time acquired in judgement
Whether motion profile matches with the standard movement track in preset track database, and the preset track database includes
Multiple and different standard movement tracks, each standard movement track are associated with a corresponding screen operator instruction;If acquired
Real time kinematics track matches with the standard movement track in preset track database, executes and standard movement track phase
Associated screen operator instruction.The embodiment of the present invention can directly by the accurate rapidly operating terminal screen of the movement of eyeball,
It realizes the automatic page turning etc. of screen, and can be improved user to the experience in terms of the intelligence and safety of terminal.
Detailed description of the invention
Technical solution in order to illustrate the embodiments of the present invention more clearly, below will be to needed in embodiment description
Attached drawing is briefly described, it should be apparent that, drawings in the following description are some embodiments of the invention, general for this field
For logical technical staff, without creative efforts, it is also possible to obtain other drawings based on these drawings.
Fig. 1 is a kind of flow diagram of mobile device operation method provided in an embodiment of the present invention;
Fig. 1 a is a kind of application scenarios schematic diagram of mobile device operation method provided in an embodiment of the present invention;
Fig. 2 is a kind of sub-process schematic diagram of mobile device operation method provided in an embodiment of the present invention;
Fig. 3 is a kind of sub-process schematic diagram of mobile device operation method provided in an embodiment of the present invention;
Fig. 4 be another embodiment of the present invention provides a kind of mobile device operation method flow diagram;
Fig. 5 is a kind of schematic block diagram of mobile device operation device provided in an embodiment of the present invention;
Fig. 6 is a kind of schematic frame of the track acquiring unit of mobile device operation device provided in an embodiment of the present invention
Figure;
Fig. 7 is a kind of schematic frame of the track judging unit of mobile device operation device provided in an embodiment of the present invention
Figure;
Fig. 8 be another embodiment of the present invention provides a kind of mobile device operation device schematic block diagram;
Fig. 9 is a kind of computer equipment structure composition schematic diagram provided in an embodiment of the present invention.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete
Site preparation description, it is clear that described embodiments are some of the embodiments of the present invention, instead of all the embodiments.Based on this hair
Embodiment in bright, every other implementation obtained by those of ordinary skill in the art without making creative efforts
Example, shall fall within the protection scope of the present invention.
It should be appreciated that ought use in this specification and in the appended claims, term " includes " and "comprising" instruction
Described feature, entirety, step, operation, the presence of element and/or component, but one or more of the other feature, whole is not precluded
Body, step, operation, the presence or addition of element, component and/or its set.
It is also understood that mesh of the term used in this description of the invention merely for the sake of description specific embodiment
And be not intended to limit the present invention.As description of the invention and it is used in the attached claims, unless on
Other situations are hereafter clearly indicated, otherwise " one " of singular, "one" and "the" are intended to include plural form.
Fig. 1 and Fig. 1 a is please referred to, Fig. 1 is a kind of exemplary flow of mobile device operation method provided by the embodiments of the present application
Figure, Fig. 1 a is the schematic diagram of a scenario of mobile device operation method in the embodiment of the present application.The mobile device operation method is applied to
In terminal, so that user controls by watching terminal screen attentively and operating terminal screen, that is, the control to terminal is realized.
Referring to Fig. 1, Fig. 1 is a kind of schematic flow diagram of mobile device operation method provided in an embodiment of the present invention.It should
Method may operate in smart phone (such as Android phone, IOS mobile phone), tablet computer, laptop and intelligence and set
In the terminals such as standby.As shown in Figure 1, the step of this method includes step S101~S103.
Step S101 obtains the real time kinematics track of the eyeball of user according to Vision Tracking.
In the present embodiment, under normal circumstances, the real-time fortune of the eyeball of user can be obtained according to Vision Tracking
Dynamic rail mark.The real time kinematics track of the eyeball of the user refers to the track of the eyeball of user rotated in real time.
Wherein vision tracking, which refers to, is detected, is extracted, identified and is tracked to the moving target in image sequence, is transported
The kinematic parameter of moving-target, such as position, speed, acceleration and motion profile, so that the processing and analysis of next step are carried out, and
The behavior understanding to moving target is realized, to complete more higher leveled Detection task.
In one embodiment, as shown in Fig. 2, the step S101 may include step S201~S203.
Step S201 obtains the video information within the scope of default recording time by terminal camera in real time.
In the present embodiment, terminal camera can need to carry out corresponding video capture according to user, but due to mirror
The video of the limitation of head, shooting has certain recording range, and not can be carried out 360 ° of full angle shooting.In addition, being
The motion profile for judging the eyeball of user needs to obtain the video information within the scope of default recording time in real time.Such as it can be with
Obtain the video information within first five millisecond of current point in time, first three millisecond of available current point in time or two milliseconds
Or the video information within one millisecond;At the same time it can also obtain current point in time toward be pushed forward time point that one millisecond obtains it
First two milliseconds or one millisecond of video information.Under normal circumstances, it is smaller to preset recording time range, then shows to obtain at this time
Video information is more accurate, and certainly, the setting of the size of default recording time range can be carried out according to the actual demand of user
Corresponding setting.So when can obtain video information within the scope of default recording time in real time by terminal camera.
Step S202, judge the video information content whether include user eyeball.
In the present embodiment, due to it needs to be determined that the eyeball of user real time kinematics track, need to judge the view at this time
In the content of frequency information whether include user eyeball, mainly judge in the video frame of video information whether to include user's
The pictorial information of eyeball.If the content of the video information does not include the eyeball of user, showing can not be according to video information
The corresponding movement for determining the eyeball of user is analyzed, to not can be carried out corresponding screen operator, therefore needs to reacquire default
Video information within the scope of recording time.
Step S203, if the content of the video information includes the eyeball of user, analyzed according to Vision Tracking described in
Video information, to determine the real time kinematics track of the eyeball of user.
In the present embodiment, if the content of the video information includes the eyeball of user, showing can be according to the video
Information analysis determines the corresponding movement of the eyeball of user, is able to carry out corresponding screen operator.Specifically, when the video is believed
The content of breath includes the eyeball of user, then can analyze the video information according to Vision Tracking.
Heat subject one of of the Vision Tracking as computer vision field, be to continuous image sequence into
Row moving object detection extracts feature, Classification and Identification, tracking filter, Activity recognition, to obtain the accurate motion information ginseng of target
Number (such as position, speed), and analysis is performed corresponding processing to it, realize the method to the behavior understanding of target.
More commonly used target tracking algorism has following several: the track algorithm based on Target Motion Character, such as: frame difference
Method, tracking based on light stream etc.;Based on the target tracking algorism of correlation before and after video sequence, such as: the phase based on template
Close track algorithm, correlation tracking algorithm based on characteristic point etc.;Based on the track algorithm of target signature parameter, as based on profile
Track algorithm, track algorithm based on characteristic point etc..In addition, Many researchers are by correlations such as small echo, artificial intelligence, neural networks
Knowledge is applied to target tracking domain, and achieves good effect.The above algorithm respectively has its advantage and disadvantage, it should according to answering
It is selected with occasion.
Further, the step S203 can specifically include: the video information be analyzed according to optical flow algorithm, with determination
The real time kinematics track of the eyeball of user.
In the present embodiment, the optical flow algorithm is the important method analyzed movement sequence image, light stream
Motion information not only comprising target in image, but also the abundant information of three dimensional physical structure is contained, therefore may be used to determine
The information such as the motion conditions of target and reflection image are other.
The basic principle of optical flow method detection moving object is: assigning a speed arrow to each of image pixel
Amount, which forms an image motion fields, the point one on point and three-dimension object in a particular moment of movement, on image
One is corresponding, and this corresponding relationship can be obtained by projection relation, according to the velocity vector feature of each pixel, can to image into
Mobile state analysis.If not having moving object in image, light stream vector is consecutive variations in whole image region.Work as image
In when having moving object, there are relative motion, moving object is formed by velocity vector certainty and neighborhood for target and image background
Background velocity vector is different, to detect moving object and position.
The optical flow algorithm can specifically include a pyramid optical flow algorithm (L-K), and specific working principle includes
Firstly, light stream and affine transformation matrix calculate on the image of maximum layer;Using upper one layer of calculated result as initial value
Next tomographic image is passed to, the image of this layer calculates the light stream and affine variation of this layer on the basis of this initial value
Matrix;Next tomographic image is passed to using the light stream of this layer and affine matrix as initial value again, until passing to the last layer,
That is original image layer, light stream that this layer calculates and affine transformation matrix are as last light stream and affine transformation matrix
As a result.The real time kinematics track that the video information can determine the eyeball of user is analyzed by pyramid optical flow algorithm.
In addition, further, the step S203 can specifically include: analyzing the video according to signature tracking algorithm and believe
Breath, to determine the real time kinematics track of the eyeball of user.
In the present embodiment, the signature tracking algorithm generally comprises two processes of feature extraction and characteristic matching, wherein
The feature extraction refers to extracts suitable illustrative feature from target in the image area.These features not only should be preferable
Regional partial objectives for and background, and situations such as target scale is flexible, target shape changes, target occlusion is coped with robustness.
Common target signature includes color characteristic, gray feature, textural characteristics, profile, Optical-flow Feature, corner feature etc..The spy
Sign matching refers to the similitude for exactly calculating by the way of certain and measuring candidate region and target area, and true according to similitude
Set the goal position, realization target following, so that it is determined that the real time kinematics track of the eyeball of user.In computer vision field,
Common similarity measurement criterion includes Weighted distance, Bhattacharyya coefficient, Euclidean distance, Hausdorff distance etc..
Wherein, Bhattacharyya coefficient and Euclidean distance are the most commonly used.
Step S102, judge acquired real time kinematics track whether with the standard movement rail in preset track database
Mark matches, and the preset track database includes multiple and different standard movement tracks, each standard movement track Jun Guan
Join a corresponding screen operator instruction.
In the present embodiment, under normal circumstances, the preset track database may include multiple and different standard fortune
Dynamic rail mark, wherein each standard movement track is associated with a matched screen operator instruction, by executing phase
The screen operator instruction answered can be realized the corresponding operation to terminal screen.
For example, associated screen operator instruction is at this time when one of standard movement track is to move from left to right
Realize the page turning of screen;When one of standard movement track is to move from right to left, associated screen operator is instructed at this time
To realize exiting or returning for application program;It is associated at this time when one of standard movement track is from the top down to move
Screen operator instruction is control screen slide downward;When one of standard movement track is to move from bottom to top, it is closed at this time
The screen operator instruction of connection is control screen upward sliding.
Usually require whether the acquired real time kinematics track of judgement can transport with the standard in preset track database
Dynamic rail mark matches, so that it is determined that screen operator corresponding to the acquired real time kinematics track instructs.
Further, as shown in figure 3, the step S102 may include step S301~S303.
Step S301 determines each standard movement rail in acquired real time kinematics track and preset track database
The matching degree of mark.
In the present embodiment, in the preset track database include multiple and different standard movement tracks, therefore need
Acquired real time kinematics track is matched with each standard movement track, to obtain multiple corresponding matching degrees.It is logical
The comparative analysis of multiple matching degrees is crossed, can more accurately determine the standard movement track to match with real time kinematics track.
Step S302 judges whether the maximum matching degree of numerical value is greater than preset matching degree in all matching degrees.
In the present embodiment, it can determine the mark corresponding to it according to the maximum matching degree of numerical value in all matching degrees
Quasi-moving track.In order to further prevent the erroneous judgement to the real time kinematics track of the eyeball of user, therefore need to judge all
Whether it is greater than preset matching degree with the maximum matching degree of numerical value in degree.If the maximum matching degree of numerical value is not in all matching degrees
Greater than preset matching degree, then show the real time kinematics track and undesirable, it cannot be according to the real time kinematics rail of user
Mark determines the operation of terminal screen.
Step S303 determines acquired if the maximum matching degree of numerical value is greater than preset matching degree in all matching degrees
Real time kinematics track matches with the standard movement track in preset track database.
In the present embodiment, if the maximum matching degree of numerical value is greater than preset matching degree in all matching degrees, can sentence
Fixed acquired real time kinematics track matches with the standard movement track in preset track database, at this time can according to
The standard movement track matched determines associated screen operator instruction, to realize corresponding screen operator.
Step S103, if acquired real time kinematics track and the standard movement track phase in preset track database
Match, executes screen operator instruction associated with the standard movement track.
In the present embodiment, if acquired real time kinematics track and the standard movement track in preset track database
Match, then show to can determine corresponding screen operator according to the real time kinematics track, that is, executes and the standard movement
The associated screen operator instruction in track, to realize the behaviour of the real time kinematics TRAJECTORY CONTROL terminal screen of the eyeball by user
Make, screen operator can not only be directly controlled by eye movement, moreover it is possible to effectively reduce the error to screen control, improve use
The using experience degree at family.
To sum up, the embodiment of the present invention can directly by the movement of eyeball, precisely rapidly operating terminal screen, realization be shielded
The automatic page turning etc. of curtain, and can be improved user to the experience in terms of the intelligence and safety of terminal.
Referring to Fig. 4, Fig. 4 be another embodiment of the present invention provides a kind of mobile device operation method exemplary flow
Figure.This method may operate in smart phone (such as Android phone, IOS mobile phone), tablet computer, laptop and
In the terminals such as smart machine.As shown in figure 4, the step of this method includes step S401a~S403.Wherein and in above-described embodiment
The similar step of step S101-S103 relevant explanation and be described in detail that details are not described herein, the following detailed description of being this
The step of increased in embodiment.
Whether step S401a watches terminal screen attentively by the eyeball of Algorithm of Iris Recognition real-time judge user.
In the present embodiment, iris belongs to eyeball middle layer, positioned at the forefront of tunica vasculose, in front of ciliary body, has automatic
The size of pupil is adjusted, adjusts and enters the how many effect of intraocular light.Its working principle is that utilizing the reflectivity of pupil infrared ray
It is low, and iris reflectivity is higher, keeps the luminance difference of pupil and iris in image big, and the luminance difference between iris and sclera
It is small.Eyeball image, which is obtained, by fixed eye camera extracts the pupil in eyeball image using the principle in bright pupil hole and dark pupil
Hole, using corneal reflection method correction eye camera and eyeball relative position, using corneal reflection point data as eye camera with
The basic point of the relative position of eyeball, pupil center location coordinate mean that the position of sight.It is identified by Algorithm of Iris Recognition
Whether the eyeball of user watches terminal screen attentively, can come according further to the motion profile of the eyeball of user to terminal screen into
Row more accurately operates, and prevents the maloperation generated by error.
Wherein, it if passing through the eyeball fixes terminal screen of Algorithm of Iris Recognition real-time judgment user, executes described according to view
Feel the step of track algorithm obtains the real time kinematics track of the eyeball of user, passes through user eyeball so as to further improve
Motion control screen operation precision.
Step S401b determines user if the eyeball by Algorithm of Iris Recognition real-time judgment user does not watch terminal screen attentively
Eyeball do not watch duration of terminal screen attentively.
In the present embodiment, in order to further be defined, the accuracy of control screen operator is improved, if passing through iris
The eyeball of recognizer real-time judgment user does not watch terminal screen attentively, then can then determine that the eyeball of user does not watch end panel attentively
The duration of curtain.
Step S401c judges whether the duration is more than preset duration.
Wherein, it is less than preset duration when the duration, then shows that user eyeball watches always terminal screen attentively at this time, it should
The duration that the eyeball of user does not watch screen attentively can be ignored.The even described duration is less than preset duration,
The step of then executing the real time kinematics track of the eyeball that user is obtained according to Vision Tracking.And if the duration
More than preset duration, then whether terminal screen can be watched attentively again by the eyeball of Algorithm of Iris Recognition real-time judge user, if
Determine that the eyeball of user does not still watch terminal screen attentively, and the eyeball of user does not watch the duration of terminal screen attentively more than default
Duration, then terminal screen rests on current state or enters standby mode at this time.
Step S401 obtains the eyeball of user according to Vision Tracking if the duration is less than preset duration
Real time kinematics track.
Step S402, judge acquired real time kinematics track whether with the standard movement rail in preset track database
Mark matches, and the preset track database includes multiple and different standard movement tracks, each standard movement track Jun Guan
Join a corresponding screen operator instruction.
Step S403, if acquired real time kinematics track and the standard movement track phase in preset track database
Match, executes screen operator instruction associated with the standard movement track.
Those having ordinary skill in the art is understood that realize all or part of the process in above-described embodiment method, is that can lead to
Computer program is crossed to instruct relevant hardware and complete, the program can be stored in a computer-readable storage medium
In, the program is when being executed, it may include such as the process of the embodiment of above-mentioned each method.Wherein, the storage medium can be magnetic
Dish, CD, read-only memory (Read-Only Memory, ROM) etc..
Referring to Fig. 5, corresponding a kind of above-mentioned mobile device operation method, the embodiment of the present invention also propose a kind of mobile device
Operating device, the device 100 include: track acquiring unit 101, track judging unit 102 and execution unit 103.
The track acquiring unit 101, the real time kinematics track of the eyeball for obtaining user according to Vision Tracking.
In the present embodiment, under normal circumstances, the real-time fortune of the eyeball of user can be obtained according to Vision Tracking
Dynamic rail mark.The real time kinematics track of the eyeball of the user refers to the track of the eyeball of user rotated in real time.
Wherein vision tracking, which refers to, is detected, is extracted, identified and is tracked to the moving target in image sequence, is transported
The kinematic parameter of moving-target, such as position, speed, acceleration and motion profile, so that the processing and analysis of next step are carried out, and
The behavior understanding to moving target is realized, to complete more higher leveled Detection task.
In one embodiment, as shown in fig. 6, the track acquiring unit 101 may include information acquisition unit 201, it is interior
Hold judging unit 202 and storage unit 203.
The information acquisition unit 201, for obtaining the view within the scope of default recording time in real time by terminal camera
Frequency information.
In the present embodiment, terminal camera can need to carry out corresponding video capture according to user, but due to mirror
The video of the limitation of head, shooting has certain recording range, and not can be carried out 360 ° of full angle shooting.In addition, being
The motion profile for judging the eyeball of user needs to obtain the video information within the scope of default recording time in real time.
The content judging unit 202, for judge the video information content whether include user eyeball.
In the present embodiment, due to it needs to be determined that the eyeball of user real time kinematics track, need to judge the view at this time
In the content of frequency information whether include user eyeball, mainly judge in the video frame of video information whether to include user's
The pictorial information of eyeball.
The storage unit 203, if the content for the video information includes the eyeball of user, according to vision with
Track algorithm analyzes the video information, to determine the real time kinematics track of the eyeball of user.
In the present embodiment, if the content of the video information includes the eyeball of user, showing can be according to the video
Information analysis determines the corresponding movement of the eyeball of user, is able to carry out corresponding screen operator.Specifically, when the video is believed
The content of breath includes the eyeball of user, then can analyze the video information according to Vision Tracking.
Further, the storage unit 203 specifically can be used for analyzing the video information according to optical flow algorithm,
To determine the real time kinematics track of the eyeball of user.
In the present embodiment, the optical flow algorithm is the important method analyzed movement sequence image, light stream
Motion information not only comprising target in image, but also the abundant information of three dimensional physical structure is contained, therefore may be used to determine
The information such as the motion conditions of target and reflection image are other.
The basic principle of optical flow method detection moving object is: assigning a speed arrow to each of image pixel
Amount, which forms an image motion fields, the point one on point and three-dimension object in a particular moment of movement, on image
One is corresponding, and this corresponding relationship can be obtained by projection relation, according to the velocity vector feature of each pixel, can to image into
Mobile state analysis.If not having moving object in image, light stream vector is consecutive variations in whole image region.Work as image
In when having moving object, there are relative motion, moving object is formed by velocity vector certainty and neighborhood for target and image background
Background velocity vector is different, to detect moving object and position.
The optical flow algorithm can specifically include a pyramid optical flow algorithm (L-K), and specific working principle includes
Firstly, light stream and affine transformation matrix calculate on the image of maximum layer;Using upper one layer of calculated result as initial value
Next tomographic image is passed to, the image of this layer calculates the light stream and affine variation of this layer on the basis of this initial value
Matrix;Next tomographic image is passed to using the light stream of this layer and affine matrix as initial value again, until passing to the last layer,
That is original image layer, light stream that this layer calculates and affine transformation matrix are as last light stream and affine transformation matrix
As a result.The real time kinematics track that the video information can determine the eyeball of user is analyzed by pyramid optical flow algorithm.
In addition, further, the storage unit 203 specifically can be used for according to the analysis of signature tracking algorithm
Video information, to determine the real time kinematics track of the eyeball of user.
In the present embodiment, the signature tracking algorithm generally comprises two processes of feature extraction and characteristic matching, wherein
The feature extraction refers to extracts suitable illustrative feature from target in the image area.These features not only should be preferable
Regional partial objectives for and background, and situations such as target scale is flexible, target shape changes, target occlusion is coped with robustness.
Common target signature includes color characteristic, gray feature, textural characteristics, profile, Optical-flow Feature, corner feature etc..The spy
Sign matching refers to the similitude for exactly calculating by the way of certain and measuring candidate region and target area, and true according to similitude
Set the goal position, realization target following, so that it is determined that the real time kinematics track of the eyeball of user.In computer vision field,
Common similarity measurement criterion includes Weighted distance, Bhattacharyya coefficient, Euclidean distance, Hausdorff distance etc..
Wherein, Bhattacharyya coefficient and Euclidean distance are the most commonly used.
The track judging unit 102, for judge acquired real time kinematics track whether with preset track data
Standard movement track in library matches, and the preset track database includes multiple and different standard movement track, each
Standard movement track is associated with a corresponding screen operator instruction.
Further, as shown in fig. 7, the track judging unit 102 may include that matching determination unit 301, numerical value are sentenced
Disconnected unit 302 and the first judging unit 303.
The matching determination unit 301, for determining in acquired real time kinematics track and preset track database
Each standard movement track matching degree.
In the present embodiment, in the preset track database include multiple and different standard movement tracks, therefore need
Acquired real time kinematics track is matched with each standard movement track, to obtain multiple corresponding matching degrees.It is logical
The comparative analysis of multiple matching degrees is crossed, can more accurately determine the standard movement track to match with real time kinematics track.
The numerical value judging unit 302, for judging whether the maximum matching degree of numerical value is greater than default in all matching degrees
Matching degree.
In the present embodiment, it can determine the mark corresponding to it according to the maximum matching degree of numerical value in all matching degrees
Quasi-moving track.In order to further prevent the erroneous judgement to the real time kinematics track of the eyeball of user, therefore need to judge all
Whether it is greater than preset matching degree with the maximum matching degree of numerical value in degree.If the maximum matching degree of numerical value is not in all matching degrees
Greater than preset matching degree, then show the real time kinematics track and undesirable, it cannot be according to the real time kinematics rail of user
Mark determines the operation of terminal screen.
First judging unit 303, if being greater than preset matching for the maximum matching degree of numerical value in all matching degrees
Degree determines that acquired real time kinematics track matches with the standard movement track in preset track database.
In the present embodiment, if the maximum matching degree of numerical value is greater than preset matching degree in all matching degrees, can sentence
Fixed acquired real time kinematics track matches with the standard movement track in preset track database, at this time can according to
The standard movement track matched determines associated screen operator instruction, to realize corresponding screen operator.
The execution unit 103, if for the standard in acquired real time kinematics track and preset track database
Motion profile matches, and executes screen operator instruction associated with the standard movement track.
In the present embodiment, if acquired real time kinematics track and the standard movement track in preset track database
Match, then show to can determine corresponding screen operator according to the real time kinematics track, that is, executes and the standard movement
The associated screen operator instruction in track, to realize the behaviour of the real time kinematics TRAJECTORY CONTROL terminal screen of the eyeball by user
Make, screen operator can not only be directly controlled by eye movement, moreover it is possible to effectively reduce the error to screen control, improve use
The using experience degree at family.
Referring to Fig. 8, corresponding a kind of above-mentioned mobile device operation method, another embodiment of the present invention also propose a kind of movement
Apparatus for operating equipment, the device 400 include: to watch judging unit 401a, time determination unit 401b, time judging unit attentively
401c, track acquiring unit 401, track judging unit 402 and execution unit 403.
It is described to watch judging unit 401a attentively, for whether watching end attentively by the eyeball of Algorithm of Iris Recognition real-time judge user
Hold screen.
In the present embodiment, iris belongs to eyeball middle layer, positioned at the forefront of tunica vasculose, in front of ciliary body, has automatic
The size of pupil is adjusted, adjusts and enters the how many effect of intraocular light.Its working principle is that utilizing the reflectivity of pupil infrared ray
It is low, and iris reflectivity is higher, keeps the luminance difference of pupil and iris in image big, and the luminance difference between iris and sclera
It is small.Eyeball image, which is obtained, by fixed eye camera extracts the pupil in eyeball image using the principle in bright pupil hole and dark pupil
Hole, using corneal reflection method correction eye camera and eyeball relative position, using corneal reflection point data as eye camera with
The basic point of the relative position of eyeball, pupil center location coordinate mean that the position of sight.
If the track acquiring unit 401 is also used to the eyeball fixes terminal by Algorithm of Iris Recognition real-time judgment user
Screen obtains the real time kinematics track of the eyeball of user according to Vision Tracking, passes through use so as to further improve
The precision of the operation of the motion control screen of family eyeball.
The time determination unit 401b, if for not watching end attentively by the eyeball of Algorithm of Iris Recognition real-time judgment user
Screen is held, determines that the eyeball of user does not watch the duration of terminal screen attentively.
In the present embodiment, in order to further be defined, the accuracy of control screen operator is improved, if passing through iris
The eyeball of recognizer real-time judgment user does not watch terminal screen attentively, then can then determine that the eyeball of user does not watch end panel attentively
The duration of curtain.
The time judging unit 401c, for judging whether the duration is more than preset duration.
It is less than preset duration when the duration, then shows that user eyeball watches always terminal screen attentively at this time, the user
Eyeball do not watch duration of screen attentively and can be ignored.The even described duration is less than preset duration, then holds
The step of real time kinematics track of the row eyeball that user is obtained according to Vision Tracking.And if the duration is more than
Whether preset duration then can watch terminal screen attentively again by the eyeball of Algorithm of Iris Recognition real-time judge user, if it is determined that
The eyeball of user does not still watch terminal screen attentively, and the eyeball of user does not watch the duration of terminal screen attentively when being more than default
Long, then terminal screen rests on current state or enters standby mode at this time.
If the track acquiring unit 401 is less than preset duration for the duration, according to Vision Tracking
Obtain the real time kinematics track of the eyeball of user.
The track judging unit 402, for judge acquired real time kinematics track whether with preset track data
Standard movement track in library matches, and the preset track database includes multiple and different standard movement track, each
Standard movement track is associated with a corresponding screen operator instruction.
The execution unit 403, if for the standard in acquired real time kinematics track and preset track database
Motion profile matches, and executes screen operator instruction associated with the standard movement track.
It should be noted that it is apparent to those skilled in the art that, above-mentioned mobile device operation device
100 and each unit specific implementation process, can with reference to the corresponding description in preceding method embodiment, for convenience of description and
Succinctly, details are not described herein.
As seen from the above, in hardware realization, the above track acquiring unit 101, track judging unit 102 and list is executed
Member 103 etc. can be embedded in the form of hardware or independently of in the device of mobile device operation, can also store in a software form
In the memory of mobile device operation device, the corresponding operation of above each unit is executed so that processor calls.The processing
Device can be central processing unit (CPU), microprocessor, single-chip microcontroller etc..
Above-mentioned mobile device operation device can be implemented as a kind of form of computer program, and computer program can be such as
It is run in computer equipment shown in Fig. 9.
Fig. 9 is a kind of structure composition schematic diagram of computer equipment of the present invention.The equipment can be terminal, wherein terminal
Smart phone, tablet computer, laptop, desktop computer, personal digital assistant and wearable device etc. be can be with logical
The electronic equipment of telecommunication function.
Referring to Fig. 9, which includes processor 502, memory, the memory connected by system bus 501
Reservoir 504 and network interface 505, wherein memory may include non-volatile memory medium 503 and built-in storage 504.
The non-volatile memory medium 503 can storage program area 5031 and computer program 5032, the computer program
5032 include program instruction, which is performed, and processor 502 may make to execute a kind of mobile device operation method.
The processor 502 supports the operation of entire computer equipment 500 for providing calculating and control ability.
The built-in storage 504 provides environment for the operation of the computer program 5032 in non-volatile memory medium 503, should
When computer program 5032 is executed by processor 502, processor 502 may make to execute a kind of mobile device operation method.
The network interface 505 is used to carry out network communication with other equipment.It will be understood by those skilled in the art that in Fig. 9
The structure shown, only the block diagram of part-structure relevant to application scheme, does not constitute and is applied to application scheme
The restriction of computer equipment 500 thereon, specific computer equipment 500 may include more more or fewer than as shown in the figure
Component perhaps combines certain components or with different component layouts.
Wherein, the processor 502 is for running computer program 5032 stored in memory, to realize following step
It is rapid: the real time kinematics track of the eyeball of user is obtained according to Vision Tracking;Whether the acquired real time kinematics track of judgement
Match with the standard movement track in preset track database, the preset track database includes multiple and different mark
Quasi-moving track, each standard movement track are associated with a corresponding screen operator instruction;If acquired real time kinematics track
Match with the standard movement track in preset track database, executes screen behaviour associated with the standard movement track
It instructs.
In one embodiment, processor 502 is realizing the real-time of the eyeball that user is obtained according to Vision Tracking
When motion profile step, it is implemented as follows step: obtaining the view within the scope of default recording time in real time by terminal camera
Frequency information;Judge the video information content whether include user eyeball;If the content of the video information includes user
Eyeball, the video information is analyzed according to Vision Tracking, to determine the real time kinematics track of the eyeball of user.
In one embodiment, processor 502 is described according to the Vision Tracking analysis video information in realization, with true
When determining the real time kinematics Trace step of the eyeball of user, it is implemented as follows step: the video being analyzed according to optical flow algorithm and is believed
Breath, to determine the real time kinematics track of the eyeball of user.
In one embodiment, processor 502 is described according to the Vision Tracking analysis video information in realization, with true
When determining the real time kinematics Trace step of the eyeball of user, it is implemented as follows step: the view is analyzed according to signature tracking algorithm
Frequency information, to determine the real time kinematics track of the eyeball of user.
In one embodiment, processor 502 is realizing the real-time of the eyeball that user is obtained according to Vision Tracking
Before motion profile step, it is implemented as follows step: whether being watched attentively by the eyeball of Algorithm of Iris Recognition real-time judge user
Terminal screen;If passing through the eyeball fixes terminal screen of Algorithm of Iris Recognition real-time judgment user, execute the acquisition user's
The step of real time kinematics track of eyeball.
In one embodiment, processor 502, which is also realized, realizes following steps: if being used by Algorithm of Iris Recognition real-time judgment
The eyeball at family does not watch terminal screen attentively, determines that the eyeball of user does not watch the duration of terminal screen attentively;If the duration
The step of being less than preset duration, then executing the real time kinematics track of the eyeball for obtaining user.
It should be appreciated that in the embodiment of the present application, processor 502 can be central processing unit (Central
Processing Unit, CPU), which can also be other general processors, digital signal processor (Digital
Signal Processor, DSP), specific integrated circuit (Application Specific Integrated Circuit,
ASIC), ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic
Device, discrete gate or transistor logic, discrete hardware components etc..Wherein, general processor can be microprocessor or
Person's processor is also possible to any conventional processor etc..
Those of ordinary skill in the art will appreciate that be realize above-described embodiment method in all or part of the process,
It is that relevant hardware can be instructed to complete by computer program.The computer program includes program instruction, computer journey
Sequence can be stored in a storage medium, which is computer readable storage medium.The program instruction is by the department of computer science
At least one processor in system executes, to realize the process step of the embodiment of the above method.
Therefore, the present invention also provides a kind of storage mediums.The storage medium can be computer readable storage medium.This is deposited
Storage media is stored with computer program, and wherein computer program includes program instruction.The program instruction makes when being executed by processor
Processor executes following steps: the real time kinematics track of the eyeball of user is obtained according to Vision Tracking;Acquired in judgement
Whether real time kinematics track matches with the standard movement track in preset track database, the preset track database
Including multiple and different standard movement tracks, each standard movement track is associated with a corresponding screen operator instruction;If being obtained
The real time kinematics track taken matches with the standard movement track in preset track database, executes and the standard movement rail
The associated screen operator instruction of mark.
In one embodiment, the processor is realized and described is obtained according to Vision Tracking executing described program instruction
When taking the real time kinematics Trace step of the eyeball at family, it is implemented as follows step: being obtained in real time by terminal camera default
Video information within the scope of recording time;Judge the video information content whether include user eyeball;If the video
The content of information includes the eyeball of user, analyze the video information according to Vision Tracking, to determine the eyeball of user
Real time kinematics track.
In one embodiment, the processor is realized described according to Vision Tracking point in the instruction of execution described program
The video information is analysed, when real time kinematics Trace step to determine the eyeball of user, is implemented as follows step: according to light stream
Algorithm analyzes the video information, to determine the real time kinematics track of the eyeball of user.
In one embodiment, the processor is realized described according to Vision Tracking point in the instruction of execution described program
The video information is analysed, when real time kinematics Trace step to determine the eyeball of user, is implemented as follows step: according to feature
Track algorithm analyzes the video information, to determine the real time kinematics track of the eyeball of user.
In one embodiment, the processor realizes real-time fortune acquired in the judgement executing described program instruction
Dynamic rail mark whether with the standard movement track in preset track database match step when, be implemented as follows step: really
The matching degree of each standard movement track in fixed acquired real time kinematics track and preset track database;Judgement is all
Whether the maximum matching degree of numerical value is greater than preset matching degree in matching degree;If the maximum matching degree of numerical value is big in all matching degrees
In preset matching degree, acquired real time kinematics track and the standard movement track phase in preset track database are determined
Match.
In one embodiment, the processor is realized and described is obtained according to Vision Tracking executing described program instruction
Before taking the real time kinematics Trace step of the eyeball at family, also realizes following steps: being used by Algorithm of Iris Recognition real-time judge
Whether the eyeball at family watches terminal screen attentively;If passing through the eyeball fixes terminal screen of Algorithm of Iris Recognition real-time judgment user, hold
The step of real time kinematics track of the row eyeball for obtaining user.
In one embodiment, if the processor is also executed the following steps: through Algorithm of Iris Recognition real-time judgment user
Eyeball do not watch terminal screen attentively, determine that the eyeball of user does not watch the duration of terminal screen attentively;If the duration is not
More than preset duration, then execute it is described obtain user eyeball real time kinematics track the step of.
The storage medium can be USB flash disk, mobile hard disk, read-only memory (Read-Only Memory, ROM), magnetic disk
Or the various computer readable storage mediums that can store program code such as CD.
Those of ordinary skill in the art may be aware that list described in conjunction with the examples disclosed in the embodiments of the present disclosure
Member and algorithm steps, can be realized with electronic hardware, computer software, or a combination of the two, in order to clearly demonstrate hardware
With the interchangeability of software, each exemplary composition and step are generally described according to function in the above description.This
A little functions are implemented in hardware or software actually, the specific application and design constraint depending on technical solution.Specially
Industry technical staff can use different methods to achieve the described function each specific application, but this realization is not
It is considered as beyond the scope of this invention.
In several embodiments provided by the present invention, it should be understood that disclosed device and method can pass through it
Its mode is realized.For example, the apparatus embodiments described above are merely exemplary.For example, the division of each unit, only
Only a kind of logical function partition, there may be another division manner in actual implementation.Such as multiple units or components can be tied
Another system is closed or is desirably integrated into, or some features can be ignored or not executed.
The steps in the embodiment of the present invention can be sequentially adjusted, merged and deleted according to actual needs.This hair
Unit in bright embodiment device can be combined, divided and deleted according to actual needs.In addition, in each implementation of the present invention
Each functional unit in example can integrate in one processing unit, is also possible to each unit and physically exists alone, can also be with
It is that two or more units are integrated in one unit.
If the integrated unit is realized in the form of SFU software functional unit and when sold or used as an independent product,
It can store in one storage medium.Based on this understanding, technical solution of the present invention is substantially in other words to existing skill
The all or part of part or the technical solution that art contributes can be embodied in the form of software products, the meter
Calculation machine software product is stored in a storage medium, including some instructions are used so that a computer equipment (can be a
People's computer, terminal or network equipment etc.) it performs all or part of the steps of the method described in the various embodiments of the present invention.
The above description is merely a specific embodiment, but scope of protection of the present invention is not limited thereto, any
Those familiar with the art in the technical scope disclosed by the present invention, can readily occur in various equivalent modifications or replace
It changes, these modifications or substitutions should be covered by the protection scope of the present invention.Therefore, protection scope of the present invention should be with right
It is required that protection scope subject to.
Claims (10)
1. a kind of mobile device operation method, which is characterized in that the described method includes:
The real time kinematics track of the eyeball of user is obtained according to Vision Tracking;
Whether the acquired real time kinematics track of judgement matches with the standard movement track in preset track database, described
Preset track database includes multiple and different standard movement track, and each standard movement track is associated with a corresponding screen
Operational order;
If acquired real time kinematics track matches with the standard movement track in preset track database, execute with it is described
The associated screen operator instruction in standard movement track.
2. the method as described in claim 1, which is characterized in that the reality of the eyeball for obtaining user according to Vision Tracking
When motion profile, comprising:
Obtain the video information within the scope of default recording time in real time by terminal camera;
Judge the video information content whether include user eyeball;
If the content of the video information includes the eyeball of user, the video information is analyzed according to Vision Tracking, with true
Determine the real time kinematics track of the eyeball of user.
3. method according to claim 2, which is characterized in that it is described that the video information is analyzed according to Vision Tracking,
To determine the real time kinematics track of the eyeball of user, comprising:
The video information is analyzed according to optical flow algorithm, to determine the real time kinematics track of the eyeball of user.
4. method according to claim 2, which is characterized in that it is described that the video information is analyzed according to Vision Tracking,
To determine the real time kinematics track of the eyeball of user, comprising:
The video information is analyzed according to signature tracking algorithm, to determine the real time kinematics track of the eyeball of user.
5. the method as described in claim 1, which is characterized in that real time kinematics track acquired in the judgement whether with it is default
Track database in standard movement track match, comprising:
Determine the matching degree of each standard movement track in acquired real time kinematics track and preset track database;
Judge whether the maximum matching degree of numerical value is greater than preset matching degree in all matching degrees;
If the maximum matching degree of numerical value is greater than preset matching degree in all matching degrees, determine acquired real time kinematics track with
Standard movement track in preset track database matches.
6. the method as described in claim 1, which is characterized in that the reality of the eyeball for obtaining user according to Vision Tracking
When motion profile before, further includes:
Whether terminal screen is watched attentively by the eyeball of Algorithm of Iris Recognition real-time judge user;
If the eyeball fixes terminal screen of user, the real time kinematics of the eyeball that user is obtained according to Vision Tracking are executed
The step of track.
7. method as claimed in claim 6, which is characterized in that the method also includes:
If the eyeball of user does not watch terminal screen attentively, determine that the eyeball of user does not watch the duration of terminal screen attentively;
Judge whether the duration is more than preset duration;
If the duration is less than preset duration, the reality of the eyeball that user is obtained according to Vision Tracking is executed
When motion profile the step of.
8. a kind of mobile device operation device, which is characterized in that including for executing such as any one of claim 1-7 the method
Unit.
9. a kind of computer equipment, which is characterized in that the computer equipment includes memory and processor, on the memory
It is stored with computer program, the processor is realized as described in any one of claim 1-7 when executing the computer program
Method.
10. a kind of computer readable storage medium, which is characterized in that the storage medium is stored with computer program, the meter
Calculation machine program includes program instruction, and described program instruction can be realized when being executed by a processor such as any one of claim 1-7
The method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811026907.XA CN109254662A (en) | 2018-09-04 | 2018-09-04 | Mobile device operation method, apparatus, computer equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811026907.XA CN109254662A (en) | 2018-09-04 | 2018-09-04 | Mobile device operation method, apparatus, computer equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109254662A true CN109254662A (en) | 2019-01-22 |
Family
ID=65046896
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811026907.XA Pending CN109254662A (en) | 2018-09-04 | 2018-09-04 | Mobile device operation method, apparatus, computer equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109254662A (en) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110166529A (en) * | 2019-04-16 | 2019-08-23 | 平安普惠企业管理有限公司 | It keeps logging in state method, apparatus, equipment and storage medium |
CN110958736A (en) * | 2019-12-17 | 2020-04-03 | 东华大学 | Black light lamp operation method and device, black light lamp and storage medium |
CN111159678A (en) * | 2019-12-26 | 2020-05-15 | 联想(北京)有限公司 | Identity recognition method, device and storage medium |
CN111223251A (en) * | 2019-12-31 | 2020-06-02 | 航天信息股份有限公司 | Method and system for processing information based on eyeball tracking technology |
CN111753613A (en) * | 2019-09-18 | 2020-10-09 | 杭州海康威视数字技术股份有限公司 | Image analysis method, device and equipment based on experimental operation and storage medium |
CN112270210A (en) * | 2020-10-09 | 2021-01-26 | 珠海格力电器股份有限公司 | Data processing method, data processing device, operation instruction identification method, operation instruction identification device, equipment and medium |
CN112732080A (en) * | 2020-12-30 | 2021-04-30 | 宇龙计算机通信科技(深圳)有限公司 | Operation instruction generation method and device, storage medium and electronic equipment |
CN114690909A (en) * | 2022-06-01 | 2022-07-01 | 润芯微科技(江苏)有限公司 | AI visual self-adaption method, device, system and computer readable medium |
CN116185192A (en) * | 2023-02-09 | 2023-05-30 | 北京航空航天大学 | Eye movement identification VR interaction method based on denoising variation encoder |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105324794A (en) * | 2013-04-10 | 2016-02-10 | 奥克兰联合服务有限公司 | Head and eye tracking |
CN105425968A (en) * | 2015-12-17 | 2016-03-23 | 无锡天脉聚源传媒科技有限公司 | Screen unlocking method and device |
CN106104418A (en) * | 2014-03-20 | 2016-11-09 | 索尼公司 | Generate the track data for video data |
CN107239144A (en) * | 2017-06-09 | 2017-10-10 | 歌尔股份有限公司 | The input method and device of a kind of equipment |
-
2018
- 2018-09-04 CN CN201811026907.XA patent/CN109254662A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105324794A (en) * | 2013-04-10 | 2016-02-10 | 奥克兰联合服务有限公司 | Head and eye tracking |
CN106104418A (en) * | 2014-03-20 | 2016-11-09 | 索尼公司 | Generate the track data for video data |
CN105425968A (en) * | 2015-12-17 | 2016-03-23 | 无锡天脉聚源传媒科技有限公司 | Screen unlocking method and device |
CN107239144A (en) * | 2017-06-09 | 2017-10-10 | 歌尔股份有限公司 | The input method and device of a kind of equipment |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110166529A (en) * | 2019-04-16 | 2019-08-23 | 平安普惠企业管理有限公司 | It keeps logging in state method, apparatus, equipment and storage medium |
CN111753613B (en) * | 2019-09-18 | 2023-10-10 | 杭州海康威视数字技术股份有限公司 | Image analysis method, device, equipment and storage medium based on experimental operation |
CN111753613A (en) * | 2019-09-18 | 2020-10-09 | 杭州海康威视数字技术股份有限公司 | Image analysis method, device and equipment based on experimental operation and storage medium |
CN110958736B (en) * | 2019-12-17 | 2021-12-31 | 东华大学 | Black light lamp operation method and device, black light lamp and storage medium |
CN110958736A (en) * | 2019-12-17 | 2020-04-03 | 东华大学 | Black light lamp operation method and device, black light lamp and storage medium |
CN111159678B (en) * | 2019-12-26 | 2023-08-18 | 联想(北京)有限公司 | Identity recognition method, device and storage medium |
CN111159678A (en) * | 2019-12-26 | 2020-05-15 | 联想(北京)有限公司 | Identity recognition method, device and storage medium |
CN111223251B (en) * | 2019-12-31 | 2023-01-10 | 航天信息股份有限公司 | Method and system for processing information based on eyeball tracking technology |
CN111223251A (en) * | 2019-12-31 | 2020-06-02 | 航天信息股份有限公司 | Method and system for processing information based on eyeball tracking technology |
CN112270210A (en) * | 2020-10-09 | 2021-01-26 | 珠海格力电器股份有限公司 | Data processing method, data processing device, operation instruction identification method, operation instruction identification device, equipment and medium |
CN112270210B (en) * | 2020-10-09 | 2024-03-01 | 珠海格力电器股份有限公司 | Data processing and operation instruction identification method, device, equipment and medium |
CN112732080A (en) * | 2020-12-30 | 2021-04-30 | 宇龙计算机通信科技(深圳)有限公司 | Operation instruction generation method and device, storage medium and electronic equipment |
CN114690909A (en) * | 2022-06-01 | 2022-07-01 | 润芯微科技(江苏)有限公司 | AI visual self-adaption method, device, system and computer readable medium |
CN114690909B (en) * | 2022-06-01 | 2022-08-16 | 润芯微科技(江苏)有限公司 | AI visual self-adaption method, device, system and computer readable medium |
CN116185192A (en) * | 2023-02-09 | 2023-05-30 | 北京航空航天大学 | Eye movement identification VR interaction method based on denoising variation encoder |
CN116185192B (en) * | 2023-02-09 | 2023-10-20 | 北京航空航天大学 | Eye movement identification VR interaction method based on denoising variation encoder |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109254662A (en) | Mobile device operation method, apparatus, computer equipment and storage medium | |
US11650659B2 (en) | User input processing with eye tracking | |
US20180211104A1 (en) | Method and device for target tracking | |
US9940507B2 (en) | Image processing device and method for moving gesture recognition using difference images | |
US8824802B2 (en) | Method and system for gesture recognition | |
US8509484B2 (en) | Information processing device and information processing method | |
US20170192500A1 (en) | Method and electronic device for controlling terminal according to eye action | |
US10528131B2 (en) | Method to reliably detect correlations between gaze and stimuli | |
CN109375765B (en) | Eyeball tracking interaction method and device | |
CN109343700B (en) | Eye movement control calibration data acquisition method and device | |
US10990170B2 (en) | Eye tracking method, electronic device, and non-transitory computer readable storage medium | |
US20120163661A1 (en) | Apparatus and method for recognizing multi-user interactions | |
CN109977906B (en) | Gesture recognition method and system, computer device and storage medium | |
CN111259751A (en) | Video-based human behavior recognition method, device, equipment and storage medium | |
CN107273869B (en) | Gesture recognition control method and electronic equipment | |
CN110298309A (en) | Motion characteristic processing method, device, terminal and storage medium based on image | |
CN114138121B (en) | User gesture recognition method, device and system, storage medium and computing equipment | |
CN111580665B (en) | Method and device for predicting fixation point, mobile terminal and storage medium | |
KR20150106823A (en) | Gesture recognition apparatus and control method of gesture recognition apparatus | |
JP2001016606A (en) | Operation recognition system and recording medium recording operation recognition program | |
KR101374316B1 (en) | Apparatus for recognizing gesture by using see-through display and Method thereof | |
KR20200081529A (en) | HMD based User Interface Method and Device for Social Acceptability | |
Malakhova et al. | Temporal data processing from webcam eye tracking using artificial neural networks | |
CN112527103A (en) | Remote control method and device for display equipment, equipment and computer readable storage medium | |
CN113491502A (en) | Eyeball tracking calibration inspection method, device, equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20190122 |
|
WD01 | Invention patent application deemed withdrawn after publication |