CN106325517A - Target object trigger method and system and wearable equipment based on virtual reality - Google Patents
Target object trigger method and system and wearable equipment based on virtual reality Download PDFInfo
- Publication number
- CN106325517A CN106325517A CN201610757729.2A CN201610757729A CN106325517A CN 106325517 A CN106325517 A CN 106325517A CN 201610757729 A CN201610757729 A CN 201610757729A CN 106325517 A CN106325517 A CN 106325517A
- Authority
- CN
- China
- Prior art keywords
- user
- head
- virtual reality
- destination object
- presetting action
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
Abstract
The invention relates to a target object trigger method and system and wearable equipment based on virtual reality. The method includes the following steps: tracing the user's point of regard in the human-computer interaction of virtual reality, and judging whether the user's head has the preset action when the point of regard rests on the interactive target objects. If so, the interactive operation set beforehand to the target object will be triggered; otherwise, the interactive operation will not be triggered. Action of the head is preset to trigger the target object, which makes human-computer interaction become more natural and independent, happen faster in real time without delay, and have strong operability.
Description
Technical field
The present invention relates to field of virtual reality, particularly relate to a kind of destination object triggering method based on virtual reality, be
System and wearable device.
Background technology
The goal in research of virtual reality (Virtual Reality) technology is to eliminate the environment residing for people and computer system
Between boundary, i.e. in the Virtual Space that computer system provides, people can use eyes, gesture, action and language etc. each
Plant sensory organ to occur the most therewith alternately.Man-machine interaction in virtual reality, emphasizes with " the principle of artificial " center ".Such as,
People are applied to man-machine interaction to by eye tracking, and in the three dimensions of virtual reality, the thing that you are hoped is exactly that you want
Thing, be also your thing of wanting to interact.
All of man-machine interaction is required for destination object is performed " confirmation " action, such as when using PC software, by moving
Dynamic cursor of mouse rests on presses mouse button execution confirmation action on destination object, in current virtual reality applications, use
After family is by sight line Scan orientation to destination object, generally uses two ways and perform the confirmation to destination object, the first side
Formula is to perform confirmation by external equipment (button as on remote controller or VR equipment), and the second way is by judging that sight line is stopped
The time stayed on destination object triggers confirmation.First kind of way needs by external device, does not really realize with artificially
Center;After the second way specifically user is by sight line Scan orientation to destination object, then by judging that sight line rests on mesh
Time on mark object reaches certain length to trigger confirmation, and its handling process is: sight line enters destination object-beginning timing-regard
Line rests on and reaches the stipulated time (such as 3 seconds) on destination object and then trigger " confirmation " destination object and perform corresponding operating, if
Sight line removal destination object before arriving the stipulated time, then cancellation processes;Make in this way under the scene that some is concrete
There is certain drawback, such as, after user scans object by sight line, work as user think that clicking on this immediately presses if object is button
Button, still needs to wait for the time of regulation, could really perform click after such as 3 seconds, and user cannot click on immediately, and man-machine interaction is real
Time property is poor;If object is picture, when user just hopes and checks that picture is not desired to click on this picture, be necessary for the stipulated time it
Front removal sight line, otherwise can perform the operation of picture association, the handling difference of man-machine interaction automatically;If can it addition, user is whole
When viewed area is all destination object, because the sight line of user cannot leave destination object, sentence now by timing mode
Disconnected whether trigger confirmation event and be not suitable for, the most existing judge that the man-machine sight line performing confirmation action is handed over by timing mode
Mode mutually, as a mouse not having button, it is impossible to providing the user with natural and tripping manipulation sense, real-time is with handling
Difference.
Summary of the invention
The technical problem to be solved is for the deficiencies in the prior art, it is provided that a kind of mesh based on virtual reality
Mark objects trigger method, system and wearable device.
The first technical scheme that the present invention solves above-mentioned technical problem is as follows: a kind of destination object based on virtual reality
Triggering method, comprises the steps:
In the man-machine interaction of virtual reality, follow the tracks of user's point of fixation;
When user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head occurs presetting dynamic
Make, the most then trigger being set in advance in the interactive operation on described destination object, do not trigger.
The invention has the beneficial effects as follows: the present invention triggers destination object by the presetting action of head, makes man-machine interaction
The most independently, real-time without waiting for and handling by force.
On the basis of technique scheme, the present invention can also do following improvement.
Further, described presetting action includes nodding or shaking the head.
Further, it is judged that user's head whether occur presetting action be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By real-time translational speed and/or the presetting action judgment models of acceleration input prebuild of described user's head
In, whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Further, the presetting action judgment models of described prebuild includes accelerating, slowing down and reversely accelerate three rank
Section.
The second technical scheme that the present invention solves above-mentioned technical problem is as follows: a kind of destination object based on virtual reality
Triggering system, including:
Gaze tracking module, in the man-machine interaction of virtual reality, follows the tracks of user's point of fixation;
Judge module, for when user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head
There is presetting action;
Trigger module, for when user's head occurs presetting action, triggering and be set in advance on described destination object
Interactive operation.
The invention has the beneficial effects as follows: the present invention triggers destination object by the presetting action of head, makes man-machine interaction
The most independently, real-time without waiting for and handling by force.
On the basis of technique scheme, the present invention can also do following improvement.
Further, described presetting action includes nodding or shaking the head.
Further, described judge module be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By real-time translational speed and/or the presetting action judgment models of acceleration input prebuild of described user's head
In, whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Further, the presetting action judgment models of described prebuild includes accelerating, slowing down and reversely accelerate three rank
Section.
Further, the head position change of gyroscope and/or accelerometer real-time tracking user is used.
The third technical scheme that the present invention solves above-mentioned technical problem is as follows: a kind of wearing based on virtual reality sets
Standby, trigger system including above-mentioned destination object based on virtual reality.
The invention has the beneficial effects as follows: the present invention triggers destination object by the presetting action of head, makes man-machine interaction
The most independently, real-time without waiting for and handling by force.
The advantage of the aspect that the present invention adds will part be given in the following description, and part will become from the following description
Obtain substantially, or recognized by present invention practice.
Accompanying drawing explanation
Fig. 1 is destination object triggering method flow chart based on virtual reality of the present invention;
Fig. 2 is that destination object based on virtual reality of the present invention triggers system construction drawing.
Detailed description of the invention
Being described principle and the feature of the present invention below in conjunction with accompanying drawing, example is served only for explaining the present invention, and
Non-for limiting the scope of the present invention.
Fig. 1 is destination object triggering method flow chart based on virtual reality of the present invention.
As it is shown in figure 1, a kind of destination object triggering method based on virtual reality, comprise the steps:
In the man-machine interaction of virtual reality, follow the tracks of user's point of fixation;
When user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head occurs presetting dynamic
Make, the most then trigger being set in advance in the interactive operation on described destination object, the most do not trigger described destination object.
In a particular embodiment of the present invention, user's point of fixation is to represent that on the display interface of man-machine interaction sight line is watched attentively
The focus of position, in a particular application, such as when using wear-type VR glasses, user opens VR application, after putting on VR glasses
Display interface center is user's " point of fixation " (is usually shown as dot or do not shows), is similar to Mus in man-machine interaction
Target cursor.
In a particular embodiment of the present invention, of the present invention can be mutual destination object can be expressed as man-machine interaction
In a concrete object, can be such as a button, a pictures etc., it is also possible to be not specific to a concrete object,
And represent a piece of region that can be mutual.
In a particular embodiment of the present invention, it is judged that whether user's head occurs the original position of presetting action for watching attentively
Point starts to rest on the position in certain destination object.
In a particular embodiment of the present invention, when user's point of fixation rest on can be mutual destination object on time, use account
If portion does not occur presetting action, the most do not trigger destination object, can return and re-execute the step following the tracks of user's point of fixation.
In a particular embodiment of the present invention, it is judged that user's head whether occur presetting action be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to user and/or acceleration;
By in the presetting action judgment models of the real-time translational speed of user's head and/or acceleration input prebuild,
Whether presetting action judgment models output user's head of prebuild there is the judged result of presetting action.
In a particular embodiment of the present invention, can by relevant sensor determine the head space position of user with
And the head position following the tracks of user changes, such as, can use gyroscope and/or accelerometer etc., in a particular application, if originally
The method of invention uses in VR glasses, and user opens intelligent mobile terminal (including mobile phone, panel computer, palm PC etc.) and takes
The APP (the most existing VR panorama player) carried, then puts into intelligent mobile terminal VR glasses, VR glasses is applied to head
On, it is possible to 360 degree of viewing contents, user moves up and down head, then the content seen changes along with the change of head,
The real scene of the current periphery of perceptual image user, can be by being built in the sensor (top on intelligent mobile terminal or VR glasses
Spiral shell instrument and/or accelerometer) determine the head space position of user.
In a particular embodiment of the present invention, be may determine that the head space position of user by relevant sensor, its
In can use Eulerian angles representation space position, can become with the head position of real-time tracking user according to the head space position of user
Change, the real-time translational speed of head and/or acceleration can be calculated by change in location and sensing data.
In a particular embodiment of the present invention, presetting action includes nodding or shaking the head or other head movements, nods
It is relatively common confirmation action, but owing to the people of different regions is accustomed to difference, expression of shaking the head may be used to confirm, it is also possible to
Confirmation is represented with other specific actions, accordingly, it is emphasized that, presetting action of the present invention not only comprises nods
With shake the head, some other represent confirm trigger head movement also within protection scope of the present invention.
In a particular embodiment of the present invention, for nodding, the presetting action judgment models of prebuild can be the thickest
It is divided into acceleration, slows down and reversely accelerate three phases, it is also possible in the middle of accelerating and slowing down, add the downward high-speed mobile stage.Point
Head moves up and down with normal header and compares, and when nodding, head translational speed is very fast, and acceleration acutely increases and acceleration maintains
Time is shorter, and when normal header moves up and down, speed is slow, and mobile steadily degree of hence speeding up is less, by configuration pin to pre-structure of nodding
Relevant parameter in the presetting action judgment models built, can make a distinction nodding to move up and down with normal head.With
Reason, for shaking the head, the presetting action judgment models of prebuild substantially rough segmentation can also be acceleration, slows down and reversely accelerate three
Stage, it is also possible to add the high-speed mobile stage in the middle of accelerating and slowing down.Shake the head to move left and right with normal header and compare, when shaking the head
Head translational speed is very fast, and it is shorter that acceleration acutely increases and acceleration is held time, speed when normal header moves left and right
Slowly, mobile steadily degree of hence speeding up is less, by being correlated with in the configuration pin presetting action judgment models to prebuild of shaking the head
Parameter, can make a distinction shaking the head to move left and right with normal head.Therefore, by by the real-time translational speed of user's head
And/or in the presetting action judgment models of acceleration input prebuild, can obtain whether user's head occurs presetting dynamic
The judged result made.
The present invention is that user is in the sight line interaction scenarios of man-machine interaction, it is provided that a kind of new judgement confirms to trigger target pair
The mode of elephant, i.e. after user is by sight line Scan orientation to destination object, in conjunction with the head tracking technology of user, by pre-structure
The judgment models built judges that user " nods " or " shaking the head " or other actions are confirmed whether to trigger the behaviour of destination object association
Making or event, judge whether by the way of confirmation triggers different with current by the sight line time of staying, therefore the present invention is by using
Family " nodding " or " shaking the head " naturally or other actions complete the confirmation to destination object, trigger destination object, just look like
Increase by one " physical button " to " the sight line mouse " in the VR world, make man-machine interaction the most independently, real-time without
Wait and handling by force.
Fig. 2 is that destination object based on virtual reality of the present invention triggers system construction drawing.
Present invention also offers a kind of based on virtual reality destination object as shown in Figure 2 and trigger system, including:
Gaze tracking module, in the man-machine interaction of virtual reality, follows the tracks of user's point of fixation;
Judge module, for when user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head
There is presetting action;
Trigger module, for when user's head occurs presetting action, triggering and be set in advance on described destination object
Interactive operation.
In a particular embodiment of the present invention, presetting action includes nodding or shaking the head.
In a particular embodiment of the present invention, it is judged that being embodied as of module includes:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By real-time translational speed and/or the presetting action judgment models of acceleration input prebuild of described user's head
In, whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
In a particular embodiment of the present invention, the presetting action judgment models of prebuild includes accelerating, slowing down and reverse
Accelerate three phases.
In a particular embodiment of the present invention, gyroscope and/or the head position of accelerometer real-time tracking user are used
Change.
Present invention also offers a kind of wearable device based on virtual reality, including above-mentioned target pair based on virtual reality
As triggering system, wearable device based on virtual reality described here can be VR glasses, the VR helmet etc..
In the description of this specification, reference term " embodiment one ", " embodiment two ", " example ", " concrete example " or
The description of " some examples " etc. means that the concrete grammar, device or the feature that combine this embodiment or example description are contained in this
In at least one bright embodiment or example.In this manual, the schematic representation of above-mentioned term is necessarily directed to
Identical embodiment or example.And, the specific features of description, method, device or feature can be with one or more realities in office
Execute in example or example and combine in an appropriate manner.Additionally, in the case of the most conflicting, those skilled in the art can be by
Different embodiments or the feature of example and different embodiment or example described in this specification are combined and combine.
The foregoing is only presently preferred embodiments of the present invention, not in order to limit the present invention, all spirit in the present invention and
Within principle, any modification, equivalent substitution and improvement etc. made, should be included within the scope of the present invention.
Claims (10)
1. a destination object triggering method based on virtual reality, it is characterised in that comprise the steps:
In the man-machine interaction of virtual reality, follow the tracks of user's point of fixation;
When user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head there is presetting action, if
It is then to trigger being set in advance in the interactive operation on described destination object, do not trigger.
Destination object triggering method based on virtual reality the most according to claim 1, it is characterised in that described presetting
Action includes nodding or shaking the head.
Destination object triggering method based on virtual reality the most according to claim 1 and 2, it is characterised in that judge to use
Account portion whether occur presetting action be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By in the presetting action judgment models of the real-time translational speed of described user's head and/or acceleration input prebuild,
Whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Destination object triggering method based on virtual reality the most according to claim 3, it is characterised in that described prebuild
Presetting action judgment models include accelerate, slow down and reversely accelerate three phases.
5. a destination object based on virtual reality triggers system, it is characterised in that including:
Gaze tracking module, in the man-machine interaction of virtual reality, follows the tracks of user's point of fixation;
Judge module, for when user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head occurs
Presetting action;
Trigger module, for when user's head occurs presetting action, triggering and be set in advance in the friendship on described destination object
Interoperability.
Destination object based on virtual reality the most according to claim 5 triggers system, it is characterised in that described presetting
Action includes nodding or shaking the head.
7. according to described in claim 5 or 6 destination object based on virtual reality trigger system, it is characterised in that described in sentence
Being embodied as of disconnected module includes:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By in the presetting action judgment models of the real-time translational speed of described user's head and/or acceleration input prebuild,
Whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Destination object based on virtual reality the most according to claim 7 triggers system, it is characterised in that described prebuild
Presetting action judgment models include accelerate, slow down and reversely accelerate three phases.
Destination object based on virtual reality the most according to claim 7 triggers system, it is characterised in that use gyroscope
And/or the head position change of accelerometer real-time tracking user.
10. a wearable device based on virtual reality, it is characterised in that include described in any one of claim 5-9 based on
The destination object of virtual reality triggers system.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610757729.2A CN106325517A (en) | 2016-08-29 | 2016-08-29 | Target object trigger method and system and wearable equipment based on virtual reality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610757729.2A CN106325517A (en) | 2016-08-29 | 2016-08-29 | Target object trigger method and system and wearable equipment based on virtual reality |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106325517A true CN106325517A (en) | 2017-01-11 |
Family
ID=57788414
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610757729.2A Pending CN106325517A (en) | 2016-08-29 | 2016-08-29 | Target object trigger method and system and wearable equipment based on virtual reality |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106325517A (en) |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107025784A (en) * | 2017-03-30 | 2017-08-08 | 北京奇艺世纪科技有限公司 | A kind of remote control, helmet and system |
CN107158709A (en) * | 2017-05-16 | 2017-09-15 | 杭州乐见科技有限公司 | A kind of method and apparatus based on game guided-moving |
CN107340950A (en) * | 2017-07-18 | 2017-11-10 | 深圳创维-Rgb电子有限公司 | A kind of method of adjust automatically menu interface position, VR equipment and storage medium |
CN107358074A (en) * | 2017-06-29 | 2017-11-17 | 维沃移动通信有限公司 | A kind of unlocking method and virtual reality device |
CN108470205A (en) * | 2018-02-11 | 2018-08-31 | 北京光年无限科技有限公司 | Head exchange method based on visual human and system |
CN108470206A (en) * | 2018-02-11 | 2018-08-31 | 北京光年无限科技有限公司 | Head exchange method based on visual human and system |
CN109242976A (en) * | 2018-08-02 | 2019-01-18 | 实野信息科技(上海)有限公司 | A method of based on the automatic rotary display of WebGL virtual reality |
CN109696191A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of virtual reality wears the mobile delay measuring method of display equipment |
CN109696190A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of rotation delay measuring method of the VR helmet based on gradual change coding |
CN109696189A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of rotation delay measuring method of the VR helmet based on encoder |
CN109696188A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of rotation delay measuring method of the VR helmet based on laser tube |
CN109737997A (en) * | 2018-12-29 | 2019-05-10 | 北京理工大学 | A kind of virtual reality wears the rotation delay measuring method of display equipment |
WO2021036624A1 (en) * | 2019-08-28 | 2021-03-04 | 北京市商汤科技开发有限公司 | Interaction method, apparatus and device, and storage medium |
WO2023245316A1 (en) * | 2022-06-20 | 2023-12-28 | 北京小米移动软件有限公司 | Human-computer interaction method and device, computer device and storage medium |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101308400A (en) * | 2007-05-18 | 2008-11-19 | 肖斌 | Novel human-machine interaction device based on eye-motion and head motion detection |
CN103279253A (en) * | 2013-05-23 | 2013-09-04 | 广东欧珀移动通信有限公司 | Method and terminal device for theme setting |
CN103294180A (en) * | 2012-03-01 | 2013-09-11 | 联想(北京)有限公司 | Man-machine interaction control method and electronic terminal |
CN103432677A (en) * | 2013-09-08 | 2013-12-11 | 陈火 | Lucidity maintaining device |
CN103677232A (en) * | 2012-09-17 | 2014-03-26 | 联想(北京)有限公司 | Information processing method, action recognition device and electronic equipment |
CN104106262A (en) * | 2012-02-08 | 2014-10-15 | 微软公司 | Head pose tracking using a depth camera |
CN104615231A (en) * | 2013-11-01 | 2015-05-13 | 中国移动通信集团公司 | Determination method for input information, and equipment |
CN204814759U (en) * | 2013-09-08 | 2015-12-02 | 陈火 | Wear -type lucidity holdout device |
CN105528084A (en) * | 2016-01-21 | 2016-04-27 | 京东方科技集团股份有限公司 | Display control device, display control method thereof and display control system |
CN105590015A (en) * | 2014-10-24 | 2016-05-18 | 中国电信股份有限公司 | Information graph hotspot collection method and method, information graph hotspot processing method and device, and information graph hotspot system |
-
2016
- 2016-08-29 CN CN201610757729.2A patent/CN106325517A/en active Pending
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101308400A (en) * | 2007-05-18 | 2008-11-19 | 肖斌 | Novel human-machine interaction device based on eye-motion and head motion detection |
CN104106262A (en) * | 2012-02-08 | 2014-10-15 | 微软公司 | Head pose tracking using a depth camera |
CN103294180A (en) * | 2012-03-01 | 2013-09-11 | 联想(北京)有限公司 | Man-machine interaction control method and electronic terminal |
CN103677232A (en) * | 2012-09-17 | 2014-03-26 | 联想(北京)有限公司 | Information processing method, action recognition device and electronic equipment |
CN103279253A (en) * | 2013-05-23 | 2013-09-04 | 广东欧珀移动通信有限公司 | Method and terminal device for theme setting |
CN103432677A (en) * | 2013-09-08 | 2013-12-11 | 陈火 | Lucidity maintaining device |
CN204814759U (en) * | 2013-09-08 | 2015-12-02 | 陈火 | Wear -type lucidity holdout device |
CN104615231A (en) * | 2013-11-01 | 2015-05-13 | 中国移动通信集团公司 | Determination method for input information, and equipment |
CN105590015A (en) * | 2014-10-24 | 2016-05-18 | 中国电信股份有限公司 | Information graph hotspot collection method and method, information graph hotspot processing method and device, and information graph hotspot system |
CN105528084A (en) * | 2016-01-21 | 2016-04-27 | 京东方科技集团股份有限公司 | Display control device, display control method thereof and display control system |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107025784A (en) * | 2017-03-30 | 2017-08-08 | 北京奇艺世纪科技有限公司 | A kind of remote control, helmet and system |
CN107158709A (en) * | 2017-05-16 | 2017-09-15 | 杭州乐见科技有限公司 | A kind of method and apparatus based on game guided-moving |
CN107358074A (en) * | 2017-06-29 | 2017-11-17 | 维沃移动通信有限公司 | A kind of unlocking method and virtual reality device |
CN107340950A (en) * | 2017-07-18 | 2017-11-10 | 深圳创维-Rgb电子有限公司 | A kind of method of adjust automatically menu interface position, VR equipment and storage medium |
CN108470205A (en) * | 2018-02-11 | 2018-08-31 | 北京光年无限科技有限公司 | Head exchange method based on visual human and system |
CN108470206A (en) * | 2018-02-11 | 2018-08-31 | 北京光年无限科技有限公司 | Head exchange method based on visual human and system |
CN109242976A (en) * | 2018-08-02 | 2019-01-18 | 实野信息科技(上海)有限公司 | A method of based on the automatic rotary display of WebGL virtual reality |
CN109696191A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of virtual reality wears the mobile delay measuring method of display equipment |
CN109696190A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of rotation delay measuring method of the VR helmet based on gradual change coding |
CN109696189A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of rotation delay measuring method of the VR helmet based on encoder |
CN109696188A (en) * | 2018-12-29 | 2019-04-30 | 北京理工大学 | A kind of rotation delay measuring method of the VR helmet based on laser tube |
CN109737997A (en) * | 2018-12-29 | 2019-05-10 | 北京理工大学 | A kind of virtual reality wears the rotation delay measuring method of display equipment |
WO2021036624A1 (en) * | 2019-08-28 | 2021-03-04 | 北京市商汤科技开发有限公司 | Interaction method, apparatus and device, and storage medium |
WO2023245316A1 (en) * | 2022-06-20 | 2023-12-28 | 北京小米移动软件有限公司 | Human-computer interaction method and device, computer device and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106325517A (en) | Target object trigger method and system and wearable equipment based on virtual reality | |
US11410392B2 (en) | Information display in augmented reality systems | |
US20090153468A1 (en) | Virtual Interface System | |
WO2015188614A1 (en) | Method and device for operating computer and mobile phone in virtual world, and glasses using same | |
AU2010366331B2 (en) | User interface, apparatus and method for gesture recognition | |
EP3117290B1 (en) | Interactive information display | |
US20140068526A1 (en) | Method and apparatus for user interaction | |
CN107479691B (en) | Interaction method, intelligent glasses and storage device thereof | |
CN107562186B (en) | 3D campus navigation method for emotion operation based on attention identification | |
US10296359B2 (en) | Interactive system control apparatus and method | |
CN110633008A (en) | User interaction interpreter | |
WO2018139203A1 (en) | Information processing device, information processing method, and program | |
US11308694B2 (en) | Image processing apparatus and image processing method | |
US20220262080A1 (en) | Interfaces for presenting avatars in three-dimensional environments | |
CN108369451B (en) | Information processing apparatus, information processing method, and computer-readable storage medium | |
CN110941337A (en) | Control method of avatar, terminal device and computer readable storage medium | |
JPS61199178A (en) | Information input system | |
US20230259265A1 (en) | Devices, methods, and graphical user interfaces for navigating and inputting or revising content | |
GB2535730A (en) | Interactive system control apparatus and method | |
Lee et al. | Mouse operation on monitor by interactive analysis of intuitive hand motions | |
Piumsomboon | Natural hand interaction for augmented reality. | |
US20140300683A1 (en) | Data processing method, data processing apparatus and electronic device | |
EP4275108A1 (en) | Interfaces for presenting avatars in three-dimensional environments | |
CN116670627A (en) | Method for grouping user interfaces in an environment | |
KR20230068752A (en) | Method and apparatus for setting reference point of mouse using hand recognition in head-mounted display |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170111 |
|
RJ01 | Rejection of invention patent application after publication |