CN106325517A - Target object trigger method and system and wearable equipment based on virtual reality - Google Patents

Target object trigger method and system and wearable equipment based on virtual reality Download PDF

Info

Publication number
CN106325517A
CN106325517A CN201610757729.2A CN201610757729A CN106325517A CN 106325517 A CN106325517 A CN 106325517A CN 201610757729 A CN201610757729 A CN 201610757729A CN 106325517 A CN106325517 A CN 106325517A
Authority
CN
China
Prior art keywords
user
head
virtual reality
destination object
presetting action
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201610757729.2A
Other languages
Chinese (zh)
Inventor
袁超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201610757729.2A priority Critical patent/CN106325517A/en
Publication of CN106325517A publication Critical patent/CN106325517A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/012Head tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/012Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment

Abstract

The invention relates to a target object trigger method and system and wearable equipment based on virtual reality. The method includes the following steps: tracing the user's point of regard in the human-computer interaction of virtual reality, and judging whether the user's head has the preset action when the point of regard rests on the interactive target objects. If so, the interactive operation set beforehand to the target object will be triggered; otherwise, the interactive operation will not be triggered. Action of the head is preset to trigger the target object, which makes human-computer interaction become more natural and independent, happen faster in real time without delay, and have strong operability.

Description

A kind of destination object triggering method based on virtual reality, system and wearable device
Technical field
The present invention relates to field of virtual reality, particularly relate to a kind of destination object triggering method based on virtual reality, be System and wearable device.
Background technology
The goal in research of virtual reality (Virtual Reality) technology is to eliminate the environment residing for people and computer system Between boundary, i.e. in the Virtual Space that computer system provides, people can use eyes, gesture, action and language etc. each Plant sensory organ to occur the most therewith alternately.Man-machine interaction in virtual reality, emphasizes with " the principle of artificial " center ".Such as, People are applied to man-machine interaction to by eye tracking, and in the three dimensions of virtual reality, the thing that you are hoped is exactly that you want Thing, be also your thing of wanting to interact.
All of man-machine interaction is required for destination object is performed " confirmation " action, such as when using PC software, by moving Dynamic cursor of mouse rests on presses mouse button execution confirmation action on destination object, in current virtual reality applications, use After family is by sight line Scan orientation to destination object, generally uses two ways and perform the confirmation to destination object, the first side Formula is to perform confirmation by external equipment (button as on remote controller or VR equipment), and the second way is by judging that sight line is stopped The time stayed on destination object triggers confirmation.First kind of way needs by external device, does not really realize with artificially Center;After the second way specifically user is by sight line Scan orientation to destination object, then by judging that sight line rests on mesh Time on mark object reaches certain length to trigger confirmation, and its handling process is: sight line enters destination object-beginning timing-regard Line rests on and reaches the stipulated time (such as 3 seconds) on destination object and then trigger " confirmation " destination object and perform corresponding operating, if Sight line removal destination object before arriving the stipulated time, then cancellation processes;Make in this way under the scene that some is concrete There is certain drawback, such as, after user scans object by sight line, work as user think that clicking on this immediately presses if object is button Button, still needs to wait for the time of regulation, could really perform click after such as 3 seconds, and user cannot click on immediately, and man-machine interaction is real Time property is poor;If object is picture, when user just hopes and checks that picture is not desired to click on this picture, be necessary for the stipulated time it Front removal sight line, otherwise can perform the operation of picture association, the handling difference of man-machine interaction automatically;If can it addition, user is whole When viewed area is all destination object, because the sight line of user cannot leave destination object, sentence now by timing mode Disconnected whether trigger confirmation event and be not suitable for, the most existing judge that the man-machine sight line performing confirmation action is handed over by timing mode Mode mutually, as a mouse not having button, it is impossible to providing the user with natural and tripping manipulation sense, real-time is with handling Difference.
Summary of the invention
The technical problem to be solved is for the deficiencies in the prior art, it is provided that a kind of mesh based on virtual reality Mark objects trigger method, system and wearable device.
The first technical scheme that the present invention solves above-mentioned technical problem is as follows: a kind of destination object based on virtual reality Triggering method, comprises the steps:
In the man-machine interaction of virtual reality, follow the tracks of user's point of fixation;
When user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head occurs presetting dynamic Make, the most then trigger being set in advance in the interactive operation on described destination object, do not trigger.
The invention has the beneficial effects as follows: the present invention triggers destination object by the presetting action of head, makes man-machine interaction The most independently, real-time without waiting for and handling by force.
On the basis of technique scheme, the present invention can also do following improvement.
Further, described presetting action includes nodding or shaking the head.
Further, it is judged that user's head whether occur presetting action be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By real-time translational speed and/or the presetting action judgment models of acceleration input prebuild of described user's head In, whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Further, the presetting action judgment models of described prebuild includes accelerating, slowing down and reversely accelerate three rank Section.
The second technical scheme that the present invention solves above-mentioned technical problem is as follows: a kind of destination object based on virtual reality Triggering system, including:
Gaze tracking module, in the man-machine interaction of virtual reality, follows the tracks of user's point of fixation;
Judge module, for when user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head There is presetting action;
Trigger module, for when user's head occurs presetting action, triggering and be set in advance on described destination object Interactive operation.
The invention has the beneficial effects as follows: the present invention triggers destination object by the presetting action of head, makes man-machine interaction The most independently, real-time without waiting for and handling by force.
On the basis of technique scheme, the present invention can also do following improvement.
Further, described presetting action includes nodding or shaking the head.
Further, described judge module be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By real-time translational speed and/or the presetting action judgment models of acceleration input prebuild of described user's head In, whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Further, the presetting action judgment models of described prebuild includes accelerating, slowing down and reversely accelerate three rank Section.
Further, the head position change of gyroscope and/or accelerometer real-time tracking user is used.
The third technical scheme that the present invention solves above-mentioned technical problem is as follows: a kind of wearing based on virtual reality sets Standby, trigger system including above-mentioned destination object based on virtual reality.
The invention has the beneficial effects as follows: the present invention triggers destination object by the presetting action of head, makes man-machine interaction The most independently, real-time without waiting for and handling by force.
The advantage of the aspect that the present invention adds will part be given in the following description, and part will become from the following description Obtain substantially, or recognized by present invention practice.
Accompanying drawing explanation
Fig. 1 is destination object triggering method flow chart based on virtual reality of the present invention;
Fig. 2 is that destination object based on virtual reality of the present invention triggers system construction drawing.
Detailed description of the invention
Being described principle and the feature of the present invention below in conjunction with accompanying drawing, example is served only for explaining the present invention, and Non-for limiting the scope of the present invention.
Fig. 1 is destination object triggering method flow chart based on virtual reality of the present invention.
As it is shown in figure 1, a kind of destination object triggering method based on virtual reality, comprise the steps:
In the man-machine interaction of virtual reality, follow the tracks of user's point of fixation;
When user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head occurs presetting dynamic Make, the most then trigger being set in advance in the interactive operation on described destination object, the most do not trigger described destination object.
In a particular embodiment of the present invention, user's point of fixation is to represent that on the display interface of man-machine interaction sight line is watched attentively The focus of position, in a particular application, such as when using wear-type VR glasses, user opens VR application, after putting on VR glasses Display interface center is user's " point of fixation " (is usually shown as dot or do not shows), is similar to Mus in man-machine interaction Target cursor.
In a particular embodiment of the present invention, of the present invention can be mutual destination object can be expressed as man-machine interaction In a concrete object, can be such as a button, a pictures etc., it is also possible to be not specific to a concrete object, And represent a piece of region that can be mutual.
In a particular embodiment of the present invention, it is judged that whether user's head occurs the original position of presetting action for watching attentively Point starts to rest on the position in certain destination object.
In a particular embodiment of the present invention, when user's point of fixation rest on can be mutual destination object on time, use account If portion does not occur presetting action, the most do not trigger destination object, can return and re-execute the step following the tracks of user's point of fixation.
In a particular embodiment of the present invention, it is judged that user's head whether occur presetting action be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to user and/or acceleration;
By in the presetting action judgment models of the real-time translational speed of user's head and/or acceleration input prebuild, Whether presetting action judgment models output user's head of prebuild there is the judged result of presetting action.
In a particular embodiment of the present invention, can by relevant sensor determine the head space position of user with And the head position following the tracks of user changes, such as, can use gyroscope and/or accelerometer etc., in a particular application, if originally The method of invention uses in VR glasses, and user opens intelligent mobile terminal (including mobile phone, panel computer, palm PC etc.) and takes The APP (the most existing VR panorama player) carried, then puts into intelligent mobile terminal VR glasses, VR glasses is applied to head On, it is possible to 360 degree of viewing contents, user moves up and down head, then the content seen changes along with the change of head, The real scene of the current periphery of perceptual image user, can be by being built in the sensor (top on intelligent mobile terminal or VR glasses Spiral shell instrument and/or accelerometer) determine the head space position of user.
In a particular embodiment of the present invention, be may determine that the head space position of user by relevant sensor, its In can use Eulerian angles representation space position, can become with the head position of real-time tracking user according to the head space position of user Change, the real-time translational speed of head and/or acceleration can be calculated by change in location and sensing data.
In a particular embodiment of the present invention, presetting action includes nodding or shaking the head or other head movements, nods It is relatively common confirmation action, but owing to the people of different regions is accustomed to difference, expression of shaking the head may be used to confirm, it is also possible to Confirmation is represented with other specific actions, accordingly, it is emphasized that, presetting action of the present invention not only comprises nods With shake the head, some other represent confirm trigger head movement also within protection scope of the present invention.
In a particular embodiment of the present invention, for nodding, the presetting action judgment models of prebuild can be the thickest It is divided into acceleration, slows down and reversely accelerate three phases, it is also possible in the middle of accelerating and slowing down, add the downward high-speed mobile stage.Point Head moves up and down with normal header and compares, and when nodding, head translational speed is very fast, and acceleration acutely increases and acceleration maintains Time is shorter, and when normal header moves up and down, speed is slow, and mobile steadily degree of hence speeding up is less, by configuration pin to pre-structure of nodding Relevant parameter in the presetting action judgment models built, can make a distinction nodding to move up and down with normal head.With Reason, for shaking the head, the presetting action judgment models of prebuild substantially rough segmentation can also be acceleration, slows down and reversely accelerate three Stage, it is also possible to add the high-speed mobile stage in the middle of accelerating and slowing down.Shake the head to move left and right with normal header and compare, when shaking the head Head translational speed is very fast, and it is shorter that acceleration acutely increases and acceleration is held time, speed when normal header moves left and right Slowly, mobile steadily degree of hence speeding up is less, by being correlated with in the configuration pin presetting action judgment models to prebuild of shaking the head Parameter, can make a distinction shaking the head to move left and right with normal head.Therefore, by by the real-time translational speed of user's head And/or in the presetting action judgment models of acceleration input prebuild, can obtain whether user's head occurs presetting dynamic The judged result made.
The present invention is that user is in the sight line interaction scenarios of man-machine interaction, it is provided that a kind of new judgement confirms to trigger target pair The mode of elephant, i.e. after user is by sight line Scan orientation to destination object, in conjunction with the head tracking technology of user, by pre-structure The judgment models built judges that user " nods " or " shaking the head " or other actions are confirmed whether to trigger the behaviour of destination object association Making or event, judge whether by the way of confirmation triggers different with current by the sight line time of staying, therefore the present invention is by using Family " nodding " or " shaking the head " naturally or other actions complete the confirmation to destination object, trigger destination object, just look like Increase by one " physical button " to " the sight line mouse " in the VR world, make man-machine interaction the most independently, real-time without Wait and handling by force.
Fig. 2 is that destination object based on virtual reality of the present invention triggers system construction drawing.
Present invention also offers a kind of based on virtual reality destination object as shown in Figure 2 and trigger system, including:
Gaze tracking module, in the man-machine interaction of virtual reality, follows the tracks of user's point of fixation;
Judge module, for when user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head There is presetting action;
Trigger module, for when user's head occurs presetting action, triggering and be set in advance on described destination object Interactive operation.
In a particular embodiment of the present invention, presetting action includes nodding or shaking the head.
In a particular embodiment of the present invention, it is judged that being embodied as of module includes:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By real-time translational speed and/or the presetting action judgment models of acceleration input prebuild of described user's head In, whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
In a particular embodiment of the present invention, the presetting action judgment models of prebuild includes accelerating, slowing down and reverse Accelerate three phases.
In a particular embodiment of the present invention, gyroscope and/or the head position of accelerometer real-time tracking user are used Change.
Present invention also offers a kind of wearable device based on virtual reality, including above-mentioned target pair based on virtual reality As triggering system, wearable device based on virtual reality described here can be VR glasses, the VR helmet etc..
In the description of this specification, reference term " embodiment one ", " embodiment two ", " example ", " concrete example " or The description of " some examples " etc. means that the concrete grammar, device or the feature that combine this embodiment or example description are contained in this In at least one bright embodiment or example.In this manual, the schematic representation of above-mentioned term is necessarily directed to Identical embodiment or example.And, the specific features of description, method, device or feature can be with one or more realities in office Execute in example or example and combine in an appropriate manner.Additionally, in the case of the most conflicting, those skilled in the art can be by Different embodiments or the feature of example and different embodiment or example described in this specification are combined and combine.
The foregoing is only presently preferred embodiments of the present invention, not in order to limit the present invention, all spirit in the present invention and Within principle, any modification, equivalent substitution and improvement etc. made, should be included within the scope of the present invention.

Claims (10)

1. a destination object triggering method based on virtual reality, it is characterised in that comprise the steps:
In the man-machine interaction of virtual reality, follow the tracks of user's point of fixation;
When user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head there is presetting action, if It is then to trigger being set in advance in the interactive operation on described destination object, do not trigger.
Destination object triggering method based on virtual reality the most according to claim 1, it is characterised in that described presetting Action includes nodding or shaking the head.
Destination object triggering method based on virtual reality the most according to claim 1 and 2, it is characterised in that judge to use Account portion whether occur presetting action be embodied as include:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By in the presetting action judgment models of the real-time translational speed of described user's head and/or acceleration input prebuild, Whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Destination object triggering method based on virtual reality the most according to claim 3, it is characterised in that described prebuild Presetting action judgment models include accelerate, slow down and reversely accelerate three phases.
5. a destination object based on virtual reality triggers system, it is characterised in that including:
Gaze tracking module, in the man-machine interaction of virtual reality, follows the tracks of user's point of fixation;
Judge module, for when user's point of fixation rest on can be mutual destination object on time, it is judged that whether user's head occurs Presetting action;
Trigger module, for when user's head occurs presetting action, triggering and be set in advance in the friendship on described destination object Interoperability.
Destination object based on virtual reality the most according to claim 5 triggers system, it is characterised in that described presetting Action includes nodding or shaking the head.
7. according to described in claim 5 or 6 destination object based on virtual reality trigger system, it is characterised in that described in sentence Being embodied as of disconnected module includes:
The head position change of real-time tracking user;
The real-time translational speed of the head position change calculations user's head according to described user and/or acceleration;
By in the presetting action judgment models of the real-time translational speed of described user's head and/or acceleration input prebuild, Whether presetting action judgment models output user's head of described prebuild there is the judged result of presetting action.
Destination object based on virtual reality the most according to claim 7 triggers system, it is characterised in that described prebuild Presetting action judgment models include accelerate, slow down and reversely accelerate three phases.
Destination object based on virtual reality the most according to claim 7 triggers system, it is characterised in that use gyroscope And/or the head position change of accelerometer real-time tracking user.
10. a wearable device based on virtual reality, it is characterised in that include described in any one of claim 5-9 based on The destination object of virtual reality triggers system.
CN201610757729.2A 2016-08-29 2016-08-29 Target object trigger method and system and wearable equipment based on virtual reality Pending CN106325517A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610757729.2A CN106325517A (en) 2016-08-29 2016-08-29 Target object trigger method and system and wearable equipment based on virtual reality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610757729.2A CN106325517A (en) 2016-08-29 2016-08-29 Target object trigger method and system and wearable equipment based on virtual reality

Publications (1)

Publication Number Publication Date
CN106325517A true CN106325517A (en) 2017-01-11

Family

ID=57788414

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610757729.2A Pending CN106325517A (en) 2016-08-29 2016-08-29 Target object trigger method and system and wearable equipment based on virtual reality

Country Status (1)

Country Link
CN (1) CN106325517A (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107025784A (en) * 2017-03-30 2017-08-08 北京奇艺世纪科技有限公司 A kind of remote control, helmet and system
CN107158709A (en) * 2017-05-16 2017-09-15 杭州乐见科技有限公司 A kind of method and apparatus based on game guided-moving
CN107340950A (en) * 2017-07-18 2017-11-10 深圳创维-Rgb电子有限公司 A kind of method of adjust automatically menu interface position, VR equipment and storage medium
CN107358074A (en) * 2017-06-29 2017-11-17 维沃移动通信有限公司 A kind of unlocking method and virtual reality device
CN108470205A (en) * 2018-02-11 2018-08-31 北京光年无限科技有限公司 Head exchange method based on visual human and system
CN108470206A (en) * 2018-02-11 2018-08-31 北京光年无限科技有限公司 Head exchange method based on visual human and system
CN109242976A (en) * 2018-08-02 2019-01-18 实野信息科技(上海)有限公司 A method of based on the automatic rotary display of WebGL virtual reality
CN109696191A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of virtual reality wears the mobile delay measuring method of display equipment
CN109696190A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of rotation delay measuring method of the VR helmet based on gradual change coding
CN109696189A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of rotation delay measuring method of the VR helmet based on encoder
CN109696188A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of rotation delay measuring method of the VR helmet based on laser tube
CN109737997A (en) * 2018-12-29 2019-05-10 北京理工大学 A kind of virtual reality wears the rotation delay measuring method of display equipment
WO2021036624A1 (en) * 2019-08-28 2021-03-04 北京市商汤科技开发有限公司 Interaction method, apparatus and device, and storage medium
WO2023245316A1 (en) * 2022-06-20 2023-12-28 北京小米移动软件有限公司 Human-computer interaction method and device, computer device and storage medium

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101308400A (en) * 2007-05-18 2008-11-19 肖斌 Novel human-machine interaction device based on eye-motion and head motion detection
CN103279253A (en) * 2013-05-23 2013-09-04 广东欧珀移动通信有限公司 Method and terminal device for theme setting
CN103294180A (en) * 2012-03-01 2013-09-11 联想(北京)有限公司 Man-machine interaction control method and electronic terminal
CN103432677A (en) * 2013-09-08 2013-12-11 陈火 Lucidity maintaining device
CN103677232A (en) * 2012-09-17 2014-03-26 联想(北京)有限公司 Information processing method, action recognition device and electronic equipment
CN104106262A (en) * 2012-02-08 2014-10-15 微软公司 Head pose tracking using a depth camera
CN104615231A (en) * 2013-11-01 2015-05-13 中国移动通信集团公司 Determination method for input information, and equipment
CN204814759U (en) * 2013-09-08 2015-12-02 陈火 Wear -type lucidity holdout device
CN105528084A (en) * 2016-01-21 2016-04-27 京东方科技集团股份有限公司 Display control device, display control method thereof and display control system
CN105590015A (en) * 2014-10-24 2016-05-18 中国电信股份有限公司 Information graph hotspot collection method and method, information graph hotspot processing method and device, and information graph hotspot system

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101308400A (en) * 2007-05-18 2008-11-19 肖斌 Novel human-machine interaction device based on eye-motion and head motion detection
CN104106262A (en) * 2012-02-08 2014-10-15 微软公司 Head pose tracking using a depth camera
CN103294180A (en) * 2012-03-01 2013-09-11 联想(北京)有限公司 Man-machine interaction control method and electronic terminal
CN103677232A (en) * 2012-09-17 2014-03-26 联想(北京)有限公司 Information processing method, action recognition device and electronic equipment
CN103279253A (en) * 2013-05-23 2013-09-04 广东欧珀移动通信有限公司 Method and terminal device for theme setting
CN103432677A (en) * 2013-09-08 2013-12-11 陈火 Lucidity maintaining device
CN204814759U (en) * 2013-09-08 2015-12-02 陈火 Wear -type lucidity holdout device
CN104615231A (en) * 2013-11-01 2015-05-13 中国移动通信集团公司 Determination method for input information, and equipment
CN105590015A (en) * 2014-10-24 2016-05-18 中国电信股份有限公司 Information graph hotspot collection method and method, information graph hotspot processing method and device, and information graph hotspot system
CN105528084A (en) * 2016-01-21 2016-04-27 京东方科技集团股份有限公司 Display control device, display control method thereof and display control system

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107025784A (en) * 2017-03-30 2017-08-08 北京奇艺世纪科技有限公司 A kind of remote control, helmet and system
CN107158709A (en) * 2017-05-16 2017-09-15 杭州乐见科技有限公司 A kind of method and apparatus based on game guided-moving
CN107358074A (en) * 2017-06-29 2017-11-17 维沃移动通信有限公司 A kind of unlocking method and virtual reality device
CN107340950A (en) * 2017-07-18 2017-11-10 深圳创维-Rgb电子有限公司 A kind of method of adjust automatically menu interface position, VR equipment and storage medium
CN108470205A (en) * 2018-02-11 2018-08-31 北京光年无限科技有限公司 Head exchange method based on visual human and system
CN108470206A (en) * 2018-02-11 2018-08-31 北京光年无限科技有限公司 Head exchange method based on visual human and system
CN109242976A (en) * 2018-08-02 2019-01-18 实野信息科技(上海)有限公司 A method of based on the automatic rotary display of WebGL virtual reality
CN109696191A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of virtual reality wears the mobile delay measuring method of display equipment
CN109696190A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of rotation delay measuring method of the VR helmet based on gradual change coding
CN109696189A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of rotation delay measuring method of the VR helmet based on encoder
CN109696188A (en) * 2018-12-29 2019-04-30 北京理工大学 A kind of rotation delay measuring method of the VR helmet based on laser tube
CN109737997A (en) * 2018-12-29 2019-05-10 北京理工大学 A kind of virtual reality wears the rotation delay measuring method of display equipment
WO2021036624A1 (en) * 2019-08-28 2021-03-04 北京市商汤科技开发有限公司 Interaction method, apparatus and device, and storage medium
WO2023245316A1 (en) * 2022-06-20 2023-12-28 北京小米移动软件有限公司 Human-computer interaction method and device, computer device and storage medium

Similar Documents

Publication Publication Date Title
CN106325517A (en) Target object trigger method and system and wearable equipment based on virtual reality
US11410392B2 (en) Information display in augmented reality systems
US20090153468A1 (en) Virtual Interface System
WO2015188614A1 (en) Method and device for operating computer and mobile phone in virtual world, and glasses using same
AU2010366331B2 (en) User interface, apparatus and method for gesture recognition
EP3117290B1 (en) Interactive information display
US20140068526A1 (en) Method and apparatus for user interaction
CN107479691B (en) Interaction method, intelligent glasses and storage device thereof
CN107562186B (en) 3D campus navigation method for emotion operation based on attention identification
US10296359B2 (en) Interactive system control apparatus and method
CN110633008A (en) User interaction interpreter
WO2018139203A1 (en) Information processing device, information processing method, and program
US11308694B2 (en) Image processing apparatus and image processing method
US20220262080A1 (en) Interfaces for presenting avatars in three-dimensional environments
CN108369451B (en) Information processing apparatus, information processing method, and computer-readable storage medium
CN110941337A (en) Control method of avatar, terminal device and computer readable storage medium
JPS61199178A (en) Information input system
US20230259265A1 (en) Devices, methods, and graphical user interfaces for navigating and inputting or revising content
GB2535730A (en) Interactive system control apparatus and method
Lee et al. Mouse operation on monitor by interactive analysis of intuitive hand motions
Piumsomboon Natural hand interaction for augmented reality.
US20140300683A1 (en) Data processing method, data processing apparatus and electronic device
EP4275108A1 (en) Interfaces for presenting avatars in three-dimensional environments
CN116670627A (en) Method for grouping user interfaces in an environment
KR20230068752A (en) Method and apparatus for setting reference point of mouse using hand recognition in head-mounted display

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170111

RJ01 Rejection of invention patent application after publication