CN107608526A - A kind of virtual reality interactive teaching method - Google Patents
A kind of virtual reality interactive teaching method Download PDFInfo
- Publication number
- CN107608526A CN107608526A CN201711036524.6A CN201711036524A CN107608526A CN 107608526 A CN107608526 A CN 107608526A CN 201711036524 A CN201711036524 A CN 201711036524A CN 107608526 A CN107608526 A CN 107608526A
- Authority
- CN
- China
- Prior art keywords
- data
- depth
- gesture
- computer
- virtual reality
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Abstract
The technical problem to be solved in the present invention is to provide a kind of virtual reality interactive teaching method, the depth map data of depth transducer collection gesture first, it is then transferred to computer, hand depth map data that computer disposal collects simultaneously carries out gesture event corresponding to processing activation, and the gesture event content after activation, which is output on display screen, to be shown.The present invention is using depth transducer collection two-dimensional depth data, computer is handled two-dimensional depth data, then the interactive teaching content after processing is exported and shown into display screen, overall structure of the present invention is simple, cost is low, and processing mode is quick and precisely, acquisition process precision is high, and output image is more comprehensively and accurate.
Description
Technical field
The present invention relates to virtual reality system field, specifically a kind of virtual reality interactive teaching method.
Background technology
Current teaching pattern is also imparted knowledge to students with books mostly, and the content of books is more to be instructed with word and picture,
Not three-dimensional enough, student can only carry out reciting record to the content in books, interactive can not realize because effect is taught.
The content of the invention
The technical problem to be solved in the present invention is to provide a kind of virtual reality interactive teaching method, using virtual reality skill
Art, the purpose of interactive teaching is realized, add the interest of teaching.
The technical scheme is that:
A kind of virtual reality interactive teaching method, first depth transducer gather the depth map data of gesture, are then transferred to meter
The data processing module of calculation machine, then the gesture computing module of computer extract the depth two dimensional image of both hands corresponding region first
Data, then specify a Ge Zheng dynamic queues Q, and set maximum length L, gesture computing module judgment frame dynamic queue Q again length
Degree, when not up to maximum length L, frame data are joined the team, according to the order of joining the team of data in frame dynamic queue Q, and according to a most young waiter in a wineshop or an inn
Multiplication calculates the direction vector of track, and then gesture matching module calculates frame dynamic team according to the data in frame dynamic queue Q
Whether the data in row Q meet the related gesture profile stored in database, when meeting, activate the gesture event, when not meeting,
Gesture event is not activated, the frame data joined the team at first in frame dynamic queue Q are rejected, reload new frame data, finally
Activation gesture event content is output on display screen and shown by computer.
Described depth map data includes(x、y、z)Three coordinate values, x are represented and are horizontally parallel to depth transducer
Axle, the vertical parallel axles in depth transducer of y, z represent the axle perpendicular to depth transducer plane.
The depth two-dimensional image data of described both hands corresponding region concretely comprises the following steps:By all z of depth map data
Value is converted into the scope of the z values of gray level image, i.e. 0-255, and specific step of converting is to be multiplied by z values maximum in depth map data
255 are converted into after one times of value, then other z values are multiplied by identical and are worth to conversion values again in depth map data;Last basis
Gray level image Processing Algorithm obtains the depth two dimensional image of arm corresponding region.
Described depth transducer is communicated to connect by USB3.0 interfaces and computer.
Described depth transducer includes infrared coding light emission module, depth data acquisition module and depth data and passed
Defeated module, infrared coding light emission module project infrared light spot to display screen, depth data acquisition module collection infrared light spot
Reflection light, so as to obtain depth map data, depth map data is transferred to computer and analyzed by depth data transport module
Processing.
Advantages of the present invention:
The present invention is handled two-dimensional depth data using depth transducer collection two-dimensional depth data, computer, then will
Interactive teaching content after processing exports to be shown into display screen, and overall structure of the present invention is simple, and cost is low, and is located
Quick and precisely, acquisition process precision is high for reason mode, and output image is more comprehensively and accurate.
Embodiment
A kind of virtual reality interactive teaching method, has specifically included following steps:
(1), first depth transducer collection gesture depth map data(Including(x、y、z)Three coordinate values, x represent horizontal flat
Row represents the axle perpendicular to depth transducer plane in the axle of depth transducer, the vertical parallel axles in depth transducer of y, z),
Then communicated to connect by USB3.0 interfaces and computer so as to be transferred to the data processing module of computer;
(2), computer gesture computing module extract the depth two-dimensional image data of both hands corresponding region first, specifically will be deep
All z values of degree diagram data are converted into the scope of the z values of gray level image, i.e. 0-255, and specific step of converting is by depth map data
Middle maximum z values are converted into 255 after being multiplied by one times of value, and then other z values are multiplied by identical and are worth to again in depth map data
Conversion values, the depth two dimensional image of arm corresponding region is finally obtained according to gray level image Processing Algorithm;
(3), gesture computing module specify a Ge Zheng dynamic queues Q again, and set maximum length L, gesture computing module judges again
Frame dynamic queue Q length, when not up to maximum length L, frame data are joined the team, according in frame dynamic queue Q data join the team it is suitable
Sequence, and the direction vector of track is calculated according to least square method, then gesture matching module is according to the number in frame dynamic queue Q
According to whether the data calculated in frame dynamic queue Q meet the related gesture profile stored in database, when meeting, activate the hand
Gesture event, when not meeting, gesture event is not activated, the frame data joined the team at first in frame dynamic queue Q are rejected, reloaded
Activation gesture event content is output on display screen and shown by new frame data, last computer.
Wherein, depth transducer includes infrared coding light emission module, depth data acquisition module and depth data and passed
Defeated module, infrared coding light emission module project infrared light spot to display screen, depth data acquisition module collection infrared light spot
Reflection light, so as to obtain depth map data, depth map data is transferred to computer and analyzed by depth data transport module
Processing.
Claims (5)
- A kind of 1. virtual reality interactive teaching method, it is characterised in that:The depth map data of depth transducer collection gesture first, The data processing module of computer is then transferred to, then the gesture computing module of computer extracts both hands corresponding region first Depth two-dimensional image data, then a Ge Zheng dynamic queues Q is specified, and maximum length L is set, gesture computing module again move by judgment frame State queue Q length, when not up to maximum length L, frame data are joined the team, according to the order of joining the team of data in frame dynamic queue Q, and The direction vector of track is calculated according to least square method, then gesture matching module is according to the data in frame dynamic queue Q, meter Whether the data in Suan Zheng dynamic queues Q meet the related gesture profile stored in database, when meeting, activate the gesture thing Part, when not meeting, gesture event is not activated, the frame data joined the team at first in frame dynamic queue Q are rejected, reloaded new Activation gesture event content is output on display screen and shown by frame data, last computer.
- A kind of 2. virtual reality interactive teaching method according to claim 1, it is characterised in that:Described depth map data Include(x、y、z)Three coordinate values, x, which is represented, is horizontally parallel to the axle of depth transducer, and y is vertical parallel in depth transducer Axle, z represent the axle perpendicular to depth transducer plane.
- A kind of 3. virtual reality interactive teaching method according to claim 2, it is characterised in that:Described both hands correspond to area The depth two-dimensional image data in domain concretely comprises the following steps:All z values of depth map data are converted into the model of the z values of gray level image Enclose, i.e. 0-255, specific step of converting is to be converted into 255, Ran Houshen after z values maximum in depth map data are multiplied by into one times of value Other z values are multiplied by identical and are worth to conversion values again in degree diagram data;Arm is finally obtained according to gray level image Processing Algorithm The depth two dimensional image of corresponding region.
- A kind of 4. virtual reality interactive teaching method according to claim 1, it is characterised in that:Described depth transducer Communicated to connect by USB3.0 interfaces and computer.
- A kind of 5. virtual reality interactive teaching method according to claim 1, it is characterised in that:Described depth transducer Include infrared coding light emission module, depth data acquisition module and depth data transport module, infrared coding light transmitting mould Block projects infrared light spot to display screen, and depth data acquisition module gathers the reflection light of infrared light spot, so as to obtain depth map Depth map data is transferred to computer and analyzed and processed by data, depth data transport module.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711036524.6A CN107608526A (en) | 2017-10-30 | 2017-10-30 | A kind of virtual reality interactive teaching method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711036524.6A CN107608526A (en) | 2017-10-30 | 2017-10-30 | A kind of virtual reality interactive teaching method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107608526A true CN107608526A (en) | 2018-01-19 |
Family
ID=61084403
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711036524.6A Pending CN107608526A (en) | 2017-10-30 | 2017-10-30 | A kind of virtual reality interactive teaching method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107608526A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114786037A (en) * | 2022-03-17 | 2022-07-22 | 青岛虚拟现实研究院有限公司 | Self-adaptive coding compression method facing VR projection |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102609974A (en) * | 2012-03-14 | 2012-07-25 | 浙江理工大学 | Virtual viewpoint image generation process on basis of depth map segmentation and rendering |
US20130201105A1 (en) * | 2012-02-02 | 2013-08-08 | Raymond William Ptucha | Method for controlling interactive display system |
CN103488972A (en) * | 2013-09-09 | 2014-01-01 | 西安交通大学 | Method for detection fingertips based on depth information |
CN104915011A (en) * | 2015-06-28 | 2015-09-16 | 合肥金诺数码科技股份有限公司 | Open environment gesture interaction game system |
CN106022266A (en) * | 2016-05-20 | 2016-10-12 | 北京格灵深瞳信息技术有限公司 | Target tracking method and target tracking apparatus |
-
2017
- 2017-10-30 CN CN201711036524.6A patent/CN107608526A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130201105A1 (en) * | 2012-02-02 | 2013-08-08 | Raymond William Ptucha | Method for controlling interactive display system |
CN102609974A (en) * | 2012-03-14 | 2012-07-25 | 浙江理工大学 | Virtual viewpoint image generation process on basis of depth map segmentation and rendering |
CN103488972A (en) * | 2013-09-09 | 2014-01-01 | 西安交通大学 | Method for detection fingertips based on depth information |
CN104915011A (en) * | 2015-06-28 | 2015-09-16 | 合肥金诺数码科技股份有限公司 | Open environment gesture interaction game system |
CN106022266A (en) * | 2016-05-20 | 2016-10-12 | 北京格灵深瞳信息技术有限公司 | Target tracking method and target tracking apparatus |
Non-Patent Citations (2)
Title |
---|
张毅.: "《移动机器人技术基础与制作》", 31 January 2013 * |
王志龙.: "基于运动跟踪的动态手势识别算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114786037A (en) * | 2022-03-17 | 2022-07-22 | 青岛虚拟现实研究院有限公司 | Self-adaptive coding compression method facing VR projection |
CN114786037B (en) * | 2022-03-17 | 2024-04-12 | 青岛虚拟现实研究院有限公司 | VR projection-oriented adaptive coding compression method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11762475B2 (en) | AR scenario-based gesture interaction method, storage medium, and communication terminal | |
Calli et al. | Yale-CMU-Berkeley dataset for robotic manipulation research | |
Alves et al. | Comparing spatial and mobile augmented reality for guiding assembling procedures with task validation | |
CN101923809A (en) | Interactive augment reality jukebox | |
CN101610425B (en) | Method for evaluating stereo image quality and device | |
CN105021124A (en) | Planar component three-dimensional position and normal vector calculation method based on depth map | |
CN109035932A (en) | A kind of VR holography tutoring system | |
CN107452056B (en) | Augmented reality teaching system and control method thereof | |
CN109711472B (en) | Training data generation method and device | |
CN102508363A (en) | Wireless display glasses based on augmented-reality technology and implementation method for wireless display glasses | |
Mai et al. | 3D object detection with SLS-fusion network in foggy weather conditions | |
Tran et al. | Low-cost 3D scene reconstruction for response robots in real-time | |
CN112070782A (en) | Method and device for identifying scene contour, computer readable medium and electronic equipment | |
CN110928414A (en) | Three-dimensional virtual-real fusion experimental system | |
Yao et al. | A high spatial resolution depth sensing method based on binocular structured light | |
CN110389664B (en) | Fire scene simulation analysis device and method based on augmented reality | |
CN107608526A (en) | A kind of virtual reality interactive teaching method | |
CN204406423U (en) | The augmented reality recognition device that a kind of image and Quick Response Code combine | |
MX2019004350A (en) | Cross-boundary interactive transaction system and method therefor. | |
CN109658448A (en) | A kind of product introduction method and system based on body feeling interaction | |
Tadic | Study on automatic electric vehicle charging socket detection using ZED 2i depth sensor | |
De Paolis et al. | Augmented Reality, Virtual Reality, and Computer Graphics: 6th International Conference, AVR 2019, Santa Maria al Bagno, Italy, June 24–27, 2019, Proceedings, Part II | |
CN111399634B (en) | Method and device for recognizing gesture-guided object | |
WO2022160406A1 (en) | Implementation method and system for internet of things practical training system based on augmented reality technology | |
KR20210038451A (en) | Verification method and device for modeling route, unmanned vehicle, and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180119 |