CN107608526A - A kind of virtual reality interactive teaching method - Google Patents

A kind of virtual reality interactive teaching method Download PDF

Info

Publication number
CN107608526A
CN107608526A CN201711036524.6A CN201711036524A CN107608526A CN 107608526 A CN107608526 A CN 107608526A CN 201711036524 A CN201711036524 A CN 201711036524A CN 107608526 A CN107608526 A CN 107608526A
Authority
CN
China
Prior art keywords
data
depth
gesture
computer
virtual reality
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201711036524.6A
Other languages
Chinese (zh)
Inventor
陶斯佳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Anhui Tao Hua Mdt Infotech Ltd
Original Assignee
Anhui Tao Hua Mdt Infotech Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Anhui Tao Hua Mdt Infotech Ltd filed Critical Anhui Tao Hua Mdt Infotech Ltd
Priority to CN201711036524.6A priority Critical patent/CN107608526A/en
Publication of CN107608526A publication Critical patent/CN107608526A/en
Pending legal-status Critical Current

Links

Abstract

The technical problem to be solved in the present invention is to provide a kind of virtual reality interactive teaching method, the depth map data of depth transducer collection gesture first, it is then transferred to computer, hand depth map data that computer disposal collects simultaneously carries out gesture event corresponding to processing activation, and the gesture event content after activation, which is output on display screen, to be shown.The present invention is using depth transducer collection two-dimensional depth data, computer is handled two-dimensional depth data, then the interactive teaching content after processing is exported and shown into display screen, overall structure of the present invention is simple, cost is low, and processing mode is quick and precisely, acquisition process precision is high, and output image is more comprehensively and accurate.

Description

A kind of virtual reality interactive teaching method
Technical field
The present invention relates to virtual reality system field, specifically a kind of virtual reality interactive teaching method.
Background technology
Current teaching pattern is also imparted knowledge to students with books mostly, and the content of books is more to be instructed with word and picture, Not three-dimensional enough, student can only carry out reciting record to the content in books, interactive can not realize because effect is taught.
The content of the invention
The technical problem to be solved in the present invention is to provide a kind of virtual reality interactive teaching method, using virtual reality skill Art, the purpose of interactive teaching is realized, add the interest of teaching.
The technical scheme is that:
A kind of virtual reality interactive teaching method, first depth transducer gather the depth map data of gesture, are then transferred to meter The data processing module of calculation machine, then the gesture computing module of computer extract the depth two dimensional image of both hands corresponding region first Data, then specify a Ge Zheng dynamic queues Q, and set maximum length L, gesture computing module judgment frame dynamic queue Q again length Degree, when not up to maximum length L, frame data are joined the team, according to the order of joining the team of data in frame dynamic queue Q, and according to a most young waiter in a wineshop or an inn Multiplication calculates the direction vector of track, and then gesture matching module calculates frame dynamic team according to the data in frame dynamic queue Q Whether the data in row Q meet the related gesture profile stored in database, when meeting, activate the gesture event, when not meeting, Gesture event is not activated, the frame data joined the team at first in frame dynamic queue Q are rejected, reload new frame data, finally Activation gesture event content is output on display screen and shown by computer.
Described depth map data includes(x、y、z)Three coordinate values, x are represented and are horizontally parallel to depth transducer Axle, the vertical parallel axles in depth transducer of y, z represent the axle perpendicular to depth transducer plane.
The depth two-dimensional image data of described both hands corresponding region concretely comprises the following steps:By all z of depth map data Value is converted into the scope of the z values of gray level image, i.e. 0-255, and specific step of converting is to be multiplied by z values maximum in depth map data 255 are converted into after one times of value, then other z values are multiplied by identical and are worth to conversion values again in depth map data;Last basis Gray level image Processing Algorithm obtains the depth two dimensional image of arm corresponding region.
Described depth transducer is communicated to connect by USB3.0 interfaces and computer.
Described depth transducer includes infrared coding light emission module, depth data acquisition module and depth data and passed Defeated module, infrared coding light emission module project infrared light spot to display screen, depth data acquisition module collection infrared light spot Reflection light, so as to obtain depth map data, depth map data is transferred to computer and analyzed by depth data transport module Processing.
Advantages of the present invention:
The present invention is handled two-dimensional depth data using depth transducer collection two-dimensional depth data, computer, then will Interactive teaching content after processing exports to be shown into display screen, and overall structure of the present invention is simple, and cost is low, and is located Quick and precisely, acquisition process precision is high for reason mode, and output image is more comprehensively and accurate.
Embodiment
A kind of virtual reality interactive teaching method, has specifically included following steps:
(1), first depth transducer collection gesture depth map data(Including(x、y、z)Three coordinate values, x represent horizontal flat Row represents the axle perpendicular to depth transducer plane in the axle of depth transducer, the vertical parallel axles in depth transducer of y, z), Then communicated to connect by USB3.0 interfaces and computer so as to be transferred to the data processing module of computer;
(2), computer gesture computing module extract the depth two-dimensional image data of both hands corresponding region first, specifically will be deep All z values of degree diagram data are converted into the scope of the z values of gray level image, i.e. 0-255, and specific step of converting is by depth map data Middle maximum z values are converted into 255 after being multiplied by one times of value, and then other z values are multiplied by identical and are worth to again in depth map data Conversion values, the depth two dimensional image of arm corresponding region is finally obtained according to gray level image Processing Algorithm;
(3), gesture computing module specify a Ge Zheng dynamic queues Q again, and set maximum length L, gesture computing module judges again Frame dynamic queue Q length, when not up to maximum length L, frame data are joined the team, according in frame dynamic queue Q data join the team it is suitable Sequence, and the direction vector of track is calculated according to least square method, then gesture matching module is according to the number in frame dynamic queue Q According to whether the data calculated in frame dynamic queue Q meet the related gesture profile stored in database, when meeting, activate the hand Gesture event, when not meeting, gesture event is not activated, the frame data joined the team at first in frame dynamic queue Q are rejected, reloaded Activation gesture event content is output on display screen and shown by new frame data, last computer.
Wherein, depth transducer includes infrared coding light emission module, depth data acquisition module and depth data and passed Defeated module, infrared coding light emission module project infrared light spot to display screen, depth data acquisition module collection infrared light spot Reflection light, so as to obtain depth map data, depth map data is transferred to computer and analyzed by depth data transport module Processing.

Claims (5)

  1. A kind of 1. virtual reality interactive teaching method, it is characterised in that:The depth map data of depth transducer collection gesture first, The data processing module of computer is then transferred to, then the gesture computing module of computer extracts both hands corresponding region first Depth two-dimensional image data, then a Ge Zheng dynamic queues Q is specified, and maximum length L is set, gesture computing module again move by judgment frame State queue Q length, when not up to maximum length L, frame data are joined the team, according to the order of joining the team of data in frame dynamic queue Q, and The direction vector of track is calculated according to least square method, then gesture matching module is according to the data in frame dynamic queue Q, meter Whether the data in Suan Zheng dynamic queues Q meet the related gesture profile stored in database, when meeting, activate the gesture thing Part, when not meeting, gesture event is not activated, the frame data joined the team at first in frame dynamic queue Q are rejected, reloaded new Activation gesture event content is output on display screen and shown by frame data, last computer.
  2. A kind of 2. virtual reality interactive teaching method according to claim 1, it is characterised in that:Described depth map data Include(x、y、z)Three coordinate values, x, which is represented, is horizontally parallel to the axle of depth transducer, and y is vertical parallel in depth transducer Axle, z represent the axle perpendicular to depth transducer plane.
  3. A kind of 3. virtual reality interactive teaching method according to claim 2, it is characterised in that:Described both hands correspond to area The depth two-dimensional image data in domain concretely comprises the following steps:All z values of depth map data are converted into the model of the z values of gray level image Enclose, i.e. 0-255, specific step of converting is to be converted into 255, Ran Houshen after z values maximum in depth map data are multiplied by into one times of value Other z values are multiplied by identical and are worth to conversion values again in degree diagram data;Arm is finally obtained according to gray level image Processing Algorithm The depth two dimensional image of corresponding region.
  4. A kind of 4. virtual reality interactive teaching method according to claim 1, it is characterised in that:Described depth transducer Communicated to connect by USB3.0 interfaces and computer.
  5. A kind of 5. virtual reality interactive teaching method according to claim 1, it is characterised in that:Described depth transducer Include infrared coding light emission module, depth data acquisition module and depth data transport module, infrared coding light transmitting mould Block projects infrared light spot to display screen, and depth data acquisition module gathers the reflection light of infrared light spot, so as to obtain depth map Depth map data is transferred to computer and analyzed and processed by data, depth data transport module.
CN201711036524.6A 2017-10-30 2017-10-30 A kind of virtual reality interactive teaching method Pending CN107608526A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711036524.6A CN107608526A (en) 2017-10-30 2017-10-30 A kind of virtual reality interactive teaching method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711036524.6A CN107608526A (en) 2017-10-30 2017-10-30 A kind of virtual reality interactive teaching method

Publications (1)

Publication Number Publication Date
CN107608526A true CN107608526A (en) 2018-01-19

Family

ID=61084403

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711036524.6A Pending CN107608526A (en) 2017-10-30 2017-10-30 A kind of virtual reality interactive teaching method

Country Status (1)

Country Link
CN (1) CN107608526A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114786037A (en) * 2022-03-17 2022-07-22 青岛虚拟现实研究院有限公司 Self-adaptive coding compression method facing VR projection

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102609974A (en) * 2012-03-14 2012-07-25 浙江理工大学 Virtual viewpoint image generation process on basis of depth map segmentation and rendering
US20130201105A1 (en) * 2012-02-02 2013-08-08 Raymond William Ptucha Method for controlling interactive display system
CN103488972A (en) * 2013-09-09 2014-01-01 西安交通大学 Method for detection fingertips based on depth information
CN104915011A (en) * 2015-06-28 2015-09-16 合肥金诺数码科技股份有限公司 Open environment gesture interaction game system
CN106022266A (en) * 2016-05-20 2016-10-12 北京格灵深瞳信息技术有限公司 Target tracking method and target tracking apparatus

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130201105A1 (en) * 2012-02-02 2013-08-08 Raymond William Ptucha Method for controlling interactive display system
CN102609974A (en) * 2012-03-14 2012-07-25 浙江理工大学 Virtual viewpoint image generation process on basis of depth map segmentation and rendering
CN103488972A (en) * 2013-09-09 2014-01-01 西安交通大学 Method for detection fingertips based on depth information
CN104915011A (en) * 2015-06-28 2015-09-16 合肥金诺数码科技股份有限公司 Open environment gesture interaction game system
CN106022266A (en) * 2016-05-20 2016-10-12 北京格灵深瞳信息技术有限公司 Target tracking method and target tracking apparatus

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
张毅.: "《移动机器人技术基础与制作》", 31 January 2013 *
王志龙.: "基于运动跟踪的动态手势识别算法研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114786037A (en) * 2022-03-17 2022-07-22 青岛虚拟现实研究院有限公司 Self-adaptive coding compression method facing VR projection
CN114786037B (en) * 2022-03-17 2024-04-12 青岛虚拟现实研究院有限公司 VR projection-oriented adaptive coding compression method

Similar Documents

Publication Publication Date Title
US11762475B2 (en) AR scenario-based gesture interaction method, storage medium, and communication terminal
Calli et al. Yale-CMU-Berkeley dataset for robotic manipulation research
Alves et al. Comparing spatial and mobile augmented reality for guiding assembling procedures with task validation
CN101923809A (en) Interactive augment reality jukebox
CN101610425B (en) Method for evaluating stereo image quality and device
CN105021124A (en) Planar component three-dimensional position and normal vector calculation method based on depth map
CN109035932A (en) A kind of VR holography tutoring system
CN107452056B (en) Augmented reality teaching system and control method thereof
CN109711472B (en) Training data generation method and device
CN102508363A (en) Wireless display glasses based on augmented-reality technology and implementation method for wireless display glasses
Mai et al. 3D object detection with SLS-fusion network in foggy weather conditions
Tran et al. Low-cost 3D scene reconstruction for response robots in real-time
CN112070782A (en) Method and device for identifying scene contour, computer readable medium and electronic equipment
CN110928414A (en) Three-dimensional virtual-real fusion experimental system
Yao et al. A high spatial resolution depth sensing method based on binocular structured light
CN110389664B (en) Fire scene simulation analysis device and method based on augmented reality
CN107608526A (en) A kind of virtual reality interactive teaching method
CN204406423U (en) The augmented reality recognition device that a kind of image and Quick Response Code combine
MX2019004350A (en) Cross-boundary interactive transaction system and method therefor.
CN109658448A (en) A kind of product introduction method and system based on body feeling interaction
Tadic Study on automatic electric vehicle charging socket detection using ZED 2i depth sensor
De Paolis et al. Augmented Reality, Virtual Reality, and Computer Graphics: 6th International Conference, AVR 2019, Santa Maria al Bagno, Italy, June 24–27, 2019, Proceedings, Part II
CN111399634B (en) Method and device for recognizing gesture-guided object
WO2022160406A1 (en) Implementation method and system for internet of things practical training system based on augmented reality technology
KR20210038451A (en) Verification method and device for modeling route, unmanned vehicle, and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20180119