CN106774820A - The methods, devices and systems that human body attitude is superimposed with virtual scene - Google Patents
The methods, devices and systems that human body attitude is superimposed with virtual scene Download PDFInfo
- Publication number
- CN106774820A CN106774820A CN201610982677.9A CN201610982677A CN106774820A CN 106774820 A CN106774820 A CN 106774820A CN 201610982677 A CN201610982677 A CN 201610982677A CN 106774820 A CN106774820 A CN 106774820A
- Authority
- CN
- China
- Prior art keywords
- human body
- data
- character string
- key point
- equipment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Processing Or Creating Images (AREA)
Abstract
The invention discloses the methods, devices and systems that a kind of human body attitude is superimposed with virtual scene, belong to field of virtual reality, methods described includes:Step 100:Obtain the human body attitude data that somatosensory device is collected;Step 200:It is human body key point data by the human body attitude data preparation;Step 300:The human body key point data are converted into a character string;Step 400:The character string is sent to VR equipment, so that the character string parsing is human body key point data by the VR equipment, and is synchronized on virtual portrait.The present invention enriches the interest of field of virtual reality application and game for carrying out somatic sensation television game, and the present invention is capable of identify that the human action of complexity, reduces network delay, and real-time is good.
Description
Technical field
The present invention relates to field of virtual reality, method, device that a kind of human body attitude is superimposed with virtual scene are particularly related to
And system.
Background technology
Virtual reality (VR, Virtual Reality) technology is a kind of computer that can be created with the experiencing virtual world
Analogue system.It generates a kind of simulated environment using computer, is that a kind of interactive Three-Dimensional Dynamic of Multi-source Information Fusion is regarded
The system emulation of scape and entity behavior makes user be immersed in the environment.
Somatic sensation television game (English:Motion Sensing Game), the electronic game of impression is gone with body.Break through in the past single
The pure mode of operation with handle key-press input, somatic sensation television game is a kind of novel electron operated by limb action change
Game.
Simple somatic sensation television game can only be carried out on virtual reality device (VR equipment) at present, it is impossible to the complicated human body of identification
Action.
The content of the invention
The present invention provides the methods, devices and systems that a kind of human body attitude is superimposed with virtual scene, and the present invention is for carrying out
Somatic sensation television game, enriches the interest of field of virtual reality application and game, and the present invention is capable of identify that the human action of complexity,
Network delay is reduced, real-time is good.
In order to solve the above technical problems, present invention offer technical scheme is as follows:
On the one hand, the present invention provides a kind of method that human body attitude is superimposed with virtual scene, including:
Step 100:Obtain the human body attitude data that somatosensory device is collected;
Step 200:It is human body key point data by the human body attitude data preparation;
Step 300:The human body key point data are converted into a character string;
Step 400:The character string is sent to VR equipment, so that the character string parsing is described by the VR equipment
Human body key point data, and be synchronized on virtual portrait.
Further, the human body key point data bulk be 25, including whole body basic point, vertebra central point, neck,
Head, left shoulder, left elbow, left finesse, left hand, right shoulder, right elbow, right finesse, the right hand, left stern, left knee, left ankle, left foot, the right side
Stern, right knee, right ankle, right crus of diaphragm, shoulder central point, left finesse, left hand refer to, right finesse, the right hand refer to.
Further, the character string includes position data, the position of human body key point of the human body under world coordinate system
Data, the position data of human hands.
Further, in the step 400, the character string is sent to the VR equipment using Socket technologies.
On the other hand, the present invention provides a kind of service end, and the service end includes:
Human body attitude data acquisition module, for obtaining the human body attitude data that somatosensory device is collected;
Human body attitude data preparation module, for being human body key point data by the human body attitude data preparation;
Human body key point data conversion module, for the human body key point data to be converted into a character string;
Character string sending module, for the character string to be sent into VR equipment, so that the VR equipment is by the character
String resolves to the human body key point data, and is synchronized on virtual portrait.
Further, the human body key point data bulk be 25, including whole body basic point, vertebra central point, neck,
Head, left shoulder, left elbow, left finesse, left hand, right shoulder, right elbow, right finesse, the right hand, left stern, left knee, left ankle, left foot, the right side
Stern, right knee, right ankle, right crus of diaphragm, shoulder central point, left finesse, left hand refer to, right finesse, the right hand refer to.
Further, the character string includes position data, the position of human body key point of the human body under world coordinate system
Data, the position data of human hands.
Further, in the character string sending module, the character string is sent to the VR using Socket technologies
Equipment.
Another further aspect, the present invention provides the system that a kind of human body attitude is superimposed with virtual scene, including somatosensory device, VR set
Standby and above-mentioned service end, wherein:
The somatosensory device is connected with the service end, and the service end is connected with VR equipment by wireless network.
The invention has the advantages that:
The method that human body attitude of the invention is superimposed with virtual scene utilizes the mechanism of network service, by VR equipment and service
A LAN is set up between device, somatosensory device is connected to service end and carries out input data collection, while service end will be adopted again
The data for collecting are sent to VR equipment by LAN, and last VR equipment synchronizes to the data for collecting and represents and update,
The action of virtual portrait is controlled, is overlapped with virtual scene, and interaction is carried out with the element in scene, enrich virtual reality neck
The interest that domain is applied and played.
And the present invention is capable of identify that the human action of complexity, separately using the attitude of human body key point digital simulation human body
Outward, human body key point data are converted to character string and are transmitted by the present invention, improve efficiency of transmission, reduce network delay,
Real-time is good.
Brief description of the drawings
Fig. 1 is application scenarios schematic diagram of the invention;
Fig. 2 is the method schematic diagram that human body attitude of the invention is superimposed with virtual scene;
Fig. 3 is the schematic diagram of service end of the invention;
Fig. 4 is the system schematic that human body attitude of the invention is superimposed with virtual scene.
Specific embodiment
To make the technical problem to be solved in the present invention, technical scheme and advantage clearer, below in conjunction with accompanying drawing and tool
Body embodiment is described in detail.
The present invention is applied to the scene that somatic sensation television game is carried out in VR equipment, wherein, somatosensory device is connected with service end, clothes
Business end is connected with VR equipment by network, as shown in Figure 1.Heretofore described service end both can be server, it is also possible to
It is PC, somatosensory device is the equipment for referring to gather human body attitude data, such as body-sensing video camera, and network connection is preferably nothing
Line network connection, VR equipment refers to Android or IOS the end equipments (class such as the VR one head comprising mobile phone, built-in process chip is aobvious
Like equipment etc., the present invention and with the statement of VR equipment).
On the one hand, the present invention provides a kind of method that human body attitude is superimposed with virtual scene, and the method for the present invention is applied
Service end, as shown in Fig. 2 including:
Step 100:Obtain the human body attitude data that somatosensory device is collected.
In this step, somatosensory device, preferably 3D body-sensings video camera, it can realize immediately dynamic seizure, image identification,
The functions such as microphone input, speech recognition, community interactive.Somatosensory device gathers human body attitude data, and sends it to service
End.
Step 200:It is human body key point data by human body attitude data preparation.
The data volume of the human body attitude data of somatosensory device collection is huger numerous and diverse, be unfavorable for transmitting and use, it is necessary to
Detected, screening and filtering, it is human body key point data then to arrange, and human body key point data can simulation human body appearance very well
State.
Step 300:Human body key point data are converted into a character string.
Human body key point data data form be unfavorable for network transmission, in order to improve efficiency of transmission, can by its according to
Certain rule is organized into a character string.
Step 400:Character string is sent to VR equipment, so that character string parsing is human body key point data by VR equipment,
And be synchronized on virtual portrait.
After VR equipment receives character string, human body key point data is resolved to again, be synchronized to virtual in virtual scene
On personage, virtual portrait is set to make corresponding action (action of i.e. extraneous human body).In virtual scene, virtual portrait can be with
Element in scene carries out interaction, reaches the purpose that human body attitude and virtual scene are overlapped.
The method that human body attitude of the invention is superimposed with virtual scene utilizes the mechanism of network service, by VR equipment and service
A LAN is set up between device, somatosensory device is connected to service end and carries out input data collection, while service end will be adopted again
The data for collecting are sent to VR equipment by LAN, and last VR equipment synchronizes to the data for collecting and represents and update,
The action of virtual portrait is controlled, is overlapped with virtual scene, and interaction is carried out with the element in scene, enrich virtual reality neck
The interest that domain is applied and played.
And the present invention is capable of identify that the human action of complexity, separately using the attitude of human body key point digital simulation human body
Outward, human body key point data are converted to character string and are transmitted by the present invention, improve efficiency of transmission, reduce network delay,
Real-time is good.
In order to preferably characterize human body attitude, human body key point data bulk is in 25, including whole body basic point, vertebra
Heart point, neck, head, left shoulder, left elbow, left finesse, left hand, right shoulder, right elbow, right finesse, the right hand, left stern, left knee, a left side
Ankle, left foot, right stern, right knee, right ankle, right crus of diaphragm, shoulder central point, left finesse, left hand refer to, right finesse, the right hand refer to.Above-mentioned 25
Key point is mainly human joint pointses or skeletal structure, and 25 key points (artis, skeletal structure) can simulate human body
Most limb actions and form.
Foregoing character string can have various data forms, it is preferred that character string includes present frame human body in world coordinates
Position data, 25 position datas of key point of present frame human body, the position data of present frame human hands under system.Wherein will
It is convenient in order to do gesture identification that the data of hand individually send.World coordinate system, i.e., with personage position as the origin of coordinates,
Relative position of other objects relative to the point.
Data form is as follows:
km,1.000,0.000,0.000,0.000,0.000,1.000,0.000,1.000,0.000,0.000,1.000,
0.000,0.000,0.000,0.000,1.000|
kb,26300644264040000,6,25,0,0,0,0,0,0|
kh,26300644264040000,6,0,0,0,0,0,0
Wherein km rows represent position data of the present frame human body under world coordinate system;Kb rows represent present frame human body 25
The position data of key point;Kh rows represent the position data of present frame human hands.
In order to further improve the real-time of data transfer, in step 400, character string is sent to using Socket technologies
VR equipment.
Socket, as the Interprocess Communication Mechanism of BSD UNIX, commonly referred to as socket, for describing IP address and end
Mouthful, it is a handle for communication chain, can be used to realize the communication between different virtual machine or different computers.
The present invention is based on above-mentioned principle, using UNet technologies can realize the transmission of character string in Unity engines.
UNet is a network communication module in Unity3D, and bottom is based on Socket technologies, it is packaged, makes
With Unet and high-level API, a network communication environment can be very easily set up.
Unity3D is that developed by Unity Technologies allows player easily to create such as 3 D video trip
The multi-platform comprehensive development of games instrument of the type interaction contents such as play, building visualization, realtime three dimensional animation, is one complete
The professional game engine that face is integrated.
The P2P data syn-chronizations that UNet is realized using udp protocol.Wherein, udp protocol full name is UDP,
It is used for processing data bag as Transmission Control Protocol in network, is a kind of connectionless agreement.It has resource consumption small, treatment
Fireballing advantage, even if losing one or two packet once in a while, will not also have much impact to receiving result.
The present invention uses the Unet communication technologys so that retardance of the invention is low, real-time is good.
It is of course also possible to using engines such as Unreal, CryEngine, its network communication module bottom is also based on
Socket technologies.
On the other hand, the present invention provides a kind of service end (device that human body attitude is superimposed with virtual scene) 1, such as Fig. 3 institutes
Show, service end includes:
Human body attitude data acquisition module 11, for obtaining the human body attitude data that somatosensory device is collected.
Human body attitude data preparation module 12, for being human body key point data by human body attitude data preparation.
Human body key point data conversion module 13, for human body key point data to be converted into a character string.
Character string sending module 14, for character string to be sent into VR equipment, so that VR equipment behaves character string parsing
Body key point data, and be synchronized on virtual portrait.
The method that human body attitude of the invention is superimposed with virtual scene utilizes the mechanism of network service, by VR equipment and service
A LAN is set up between device, somatosensory device is connected to service end and carries out input data collection, while service end will be adopted again
The data for collecting are sent to VR equipment by LAN, and last VR equipment synchronizes to the data for collecting and represents and update,
The action of virtual portrait is controlled, is overlapped with virtual scene, and interaction is carried out with the element in scene, enrich virtual reality neck
The interest that domain is applied and played.
And the present invention is capable of identify that the human action of complexity, separately using the attitude of human body key point digital simulation human body
Outward, human body key point data are converted to character string and are transmitted by the present invention, improve efficiency of transmission, reduce network delay,
Real-time is good.
In order to preferably characterize human body attitude, human body key point data bulk is in 25, including whole body basic point, vertebra
Heart point, neck, head, left shoulder, left elbow, left finesse, left hand, right shoulder, right elbow, right finesse, the right hand, left stern, left knee, a left side
Ankle, left foot, right stern, right knee, right ankle, right crus of diaphragm, shoulder central point, left finesse, left hand refer to, right finesse, the right hand refer to.Above-mentioned 25
Key point is mainly human joint pointses or skeletal structure, and 25 key points (artis, skeletal structure) can simulate human body
Most limb actions and form.
Foregoing character string can have various data forms, it is preferred that character string includes human body under world coordinate system
Position data, the position data of human body key point, the position data of human hands.
In order to further improve the real-time of data transfer, in character string sending module, using Socket technologies by character
String is sent to VR equipment.The present invention uses the Unet communication technologys so that retardance of the invention is low, real-time is good.
Another further aspect, the present invention provides the system that a kind of human body attitude is superimposed with virtual scene, as shown in figure 4, including body
Sense equipment 2, VR equipment 3 and above-mentioned service end 1, wherein:
Somatosensory device 2 is connected with service end 1, and service end 1 is connected with VR equipment 3 by wireless network.
The method that human body attitude of the invention is superimposed with virtual scene utilizes the mechanism of network service, by VR equipment and service
A LAN is set up between device, somatosensory device is connected to service end and carries out input data collection, while service end will be adopted again
The data for collecting are sent to VR equipment by LAN, and last VR equipment synchronizes to the data for collecting and represents and update,
The action of virtual portrait is controlled, is overlapped with virtual scene, and interaction is carried out with the element in scene, enrich virtual reality neck
The interest that domain is applied and played.
And the present invention is capable of identify that the human action of complexity, separately using the attitude of human body key point digital simulation human body
Outward, human body key point data are converted to character string and are transmitted by the present invention, improve efficiency of transmission, reduce network delay,
Real-time is good.
The above is the preferred embodiment of the present invention, it is noted that for those skilled in the art
For, on the premise of principle of the present invention is not departed from, some improvements and modifications can also be made, these improvements and modifications
Should be regarded as protection scope of the present invention.
Claims (9)
1. a kind of method that human body attitude is superimposed with virtual scene, it is characterised in that including:
Step 100:Obtain the human body attitude data that somatosensory device is collected;
Step 200:It is human body key point data by the human body attitude data preparation;
Step 300:The human body key point data are converted into a character string;
Step 400:The character string is sent to VR equipment, so that the character string parsing is the human body by the VR equipment
Crucial point data, and be synchronized on virtual portrait.
2. the method that human body attitude according to claim 1 is superimposed with virtual scene, it is characterised in that the human body is crucial
Points data bulk be 25, including whole body basic point, vertebra central point, neck, head, left shoulder, left elbow, left finesse, left hand,
Right shoulder, right elbow, right finesse, the right hand, left stern, left knee, left ankle, left foot, right stern, right knee, right ankle, right crus of diaphragm, shoulder center
Point, left finesse, left hand refer to, right finesse, the right hand refer to.
3. the method that human body attitude according to claim 2 is superimposed with virtual scene, it is characterised in that the character string bag
Include position data, the position data of human body key point, the position data of human hands of the human body under world coordinate system.
4. the method being superimposed with virtual scene according to any described human body attitudes of claim 1-3, it is characterised in that the step
In rapid 400, the character string is sent to the VR equipment using Socket technologies.
5. a kind of service end, it is characterised in that the service end includes:
Human body attitude data acquisition module, for obtaining the human body attitude data that somatosensory device is collected;
Human body attitude data preparation module, for being human body key point data by the human body attitude data preparation;
Human body key point data conversion module, for the human body key point data to be converted into a character string;
Character string sending module, for the character string to be sent into VR equipment, so that the VR equipment is by the character string solution
It is human body key point data to analyse, and is synchronized on virtual portrait.
6. service end according to claim 5, it is characterised in that the human body key point data bulk is 25, including
Whole body basic point, vertebra central point, neck, head, left shoulder, left elbow, left finesse, left hand, right shoulder, right elbow, right finesse, the right side
Hand, left stern, left knee, left ankle, left foot, right stern, right knee, right ankle, right crus of diaphragm, shoulder central point, left finesse, left hand refer to, the right hand
Wrist, the right hand refer to.
7. service end according to claim 6, it is characterised in that the character string includes human body under world coordinate system
Position data, the position data of human body key point, the position data of human hands.
8. according to any described service ends of claim 5-7, it is characterised in that in the character string sending module, use
The character string is sent to the VR equipment by Socket technologies.
9. the system that a kind of human body attitude is superimposed with virtual scene, it is characterised in that will including somatosensory device, VR equipment and right
Any described service ends of 5-8 are sought, wherein:
The somatosensory device is connected with the service end, and the service end is connected with VR equipment by wireless network.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610982677.9A CN106774820A (en) | 2016-11-08 | 2016-11-08 | The methods, devices and systems that human body attitude is superimposed with virtual scene |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610982677.9A CN106774820A (en) | 2016-11-08 | 2016-11-08 | The methods, devices and systems that human body attitude is superimposed with virtual scene |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106774820A true CN106774820A (en) | 2017-05-31 |
Family
ID=58972417
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610982677.9A Pending CN106774820A (en) | 2016-11-08 | 2016-11-08 | The methods, devices and systems that human body attitude is superimposed with virtual scene |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106774820A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107340869A (en) * | 2017-07-07 | 2017-11-10 | 四川云图瑞科技有限公司 | Virtual reality interactive system based on illusory 4 engine technique |
CN108227931A (en) * | 2018-01-23 | 2018-06-29 | 北京市商汤科技开发有限公司 | For controlling the method for virtual portrait, equipment, system, program and storage medium |
CN109032339A (en) * | 2018-06-29 | 2018-12-18 | 贵州威爱教育科技有限公司 | A kind of method and system that real-time intelligent body-sensing is synchronous |
CN109407825A (en) * | 2018-08-30 | 2019-03-01 | 百度在线网络技术(北京)有限公司 | Interactive approach and device based on virtual objects |
CN110515467A (en) * | 2019-08-30 | 2019-11-29 | 贵州电网有限责任公司 | A kind of multi-person synergy interactive system in virtual reality training field |
CN112070868A (en) * | 2020-09-08 | 2020-12-11 | 北京默契破冰科技有限公司 | Animation playing method based on iOS system, electronic equipment and medium |
US11281898B2 (en) | 2019-06-28 | 2022-03-22 | Fujitsu Limited | Arm action identification method and apparatus and image processing device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102509092A (en) * | 2011-12-12 | 2012-06-20 | 北京华达诺科技有限公司 | Spatial gesture identification method |
CN102695205A (en) * | 2012-05-17 | 2012-09-26 | 江苏中科梦兰电子科技有限公司 | Method for controlling high efficiency transmission of multimedia data in wireless network environment |
CN103049244A (en) * | 2012-12-06 | 2013-04-17 | 北京小米科技有限责任公司 | Method and device for transmitting operating instructions |
CN105373224A (en) * | 2015-10-22 | 2016-03-02 | 山东大学 | Hybrid implementation game system based on pervasive computing, and method thereof |
-
2016
- 2016-11-08 CN CN201610982677.9A patent/CN106774820A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102509092A (en) * | 2011-12-12 | 2012-06-20 | 北京华达诺科技有限公司 | Spatial gesture identification method |
CN102695205A (en) * | 2012-05-17 | 2012-09-26 | 江苏中科梦兰电子科技有限公司 | Method for controlling high efficiency transmission of multimedia data in wireless network environment |
CN103049244A (en) * | 2012-12-06 | 2013-04-17 | 北京小米科技有限责任公司 | Method and device for transmitting operating instructions |
CN105373224A (en) * | 2015-10-22 | 2016-03-02 | 山东大学 | Hybrid implementation game system based on pervasive computing, and method thereof |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107340869A (en) * | 2017-07-07 | 2017-11-10 | 四川云图瑞科技有限公司 | Virtual reality interactive system based on illusory 4 engine technique |
CN107340869B (en) * | 2017-07-07 | 2021-03-30 | 四川云图瑞科技有限公司 | Virtual reality interaction system based on unreal 4 engine technology |
CN108227931A (en) * | 2018-01-23 | 2018-06-29 | 北京市商汤科技开发有限公司 | For controlling the method for virtual portrait, equipment, system, program and storage medium |
CN109032339A (en) * | 2018-06-29 | 2018-12-18 | 贵州威爱教育科技有限公司 | A kind of method and system that real-time intelligent body-sensing is synchronous |
CN109407825A (en) * | 2018-08-30 | 2019-03-01 | 百度在线网络技术(北京)有限公司 | Interactive approach and device based on virtual objects |
US11281898B2 (en) | 2019-06-28 | 2022-03-22 | Fujitsu Limited | Arm action identification method and apparatus and image processing device |
CN110515467A (en) * | 2019-08-30 | 2019-11-29 | 贵州电网有限责任公司 | A kind of multi-person synergy interactive system in virtual reality training field |
CN112070868A (en) * | 2020-09-08 | 2020-12-11 | 北京默契破冰科技有限公司 | Animation playing method based on iOS system, electronic equipment and medium |
CN112070868B (en) * | 2020-09-08 | 2024-04-30 | 北京默契破冰科技有限公司 | Animation playing method based on iOS system, electronic equipment and medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106774820A (en) | The methods, devices and systems that human body attitude is superimposed with virtual scene | |
CN106710590B (en) | Voice interaction system and method with emotion function based on virtual reality environment | |
JP7408792B2 (en) | Scene interaction methods and devices, electronic equipment and computer programs | |
Roth et al. | A simplified inverse kinematic approach for embodied VR applications | |
CN111460874A (en) | Image processing method and apparatus, image device, and storage medium | |
CN109448099A (en) | Rendering method, device, storage medium and the electronic device of picture | |
US8724849B2 (en) | Information processing device, information processing method, program, and information storage medium | |
CN103197757A (en) | Immersion type virtual reality system and implementation method thereof | |
CN108227931A (en) | For controlling the method for virtual portrait, equipment, system, program and storage medium | |
CN112198959A (en) | Virtual reality interaction method, device and system | |
US20090128555A1 (en) | System and method for creating and using live three-dimensional avatars and interworld operability | |
CN106534125A (en) | Method for realizing VR multi-person interaction system on the basis of local area network | |
CN105739703A (en) | Virtual reality somatosensory interaction system and method for wireless head-mounted display equipment | |
CN107203272A (en) | Wearable augmented reality task instruction system and method based on myoelectricity cognition technology | |
CN107943282A (en) | A kind of man-machine interactive system and method based on augmented reality and wearable device | |
CN107818595A (en) | Wearable Instant Interaction System | |
WO2024012459A1 (en) | Method and system for terminal-cloud combined virtual concert rendering for vr terminal | |
CN115337634A (en) | VR (virtual reality) system and method applied to meal games | |
CN114049468A (en) | Display method, device, equipment and storage medium | |
CN106502417A (en) | Mobile virtual real world devices are adapted to the methods, devices and systems of computer peripheral equipment | |
US20220157002A1 (en) | System and method for immersive telecommunications | |
CN110227249A (en) | A kind of upper limb training system | |
Pandzic et al. | Towards natural communication in networked collaborative virtual environments | |
CN112884906A (en) | System and method for realizing multi-person mixed virtual and augmented reality interaction | |
CN107243147A (en) | Boxing training virtual reality system and its implementation based on body-sensing sensor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20170531 |
|
RJ01 | Rejection of invention patent application after publication |