CN109395375A - A kind of 3d gaming method of interface interacted based on augmented reality and movement - Google Patents

A kind of 3d gaming method of interface interacted based on augmented reality and movement Download PDF

Info

Publication number
CN109395375A
CN109395375A CN201811086686.5A CN201811086686A CN109395375A CN 109395375 A CN109395375 A CN 109395375A CN 201811086686 A CN201811086686 A CN 201811086686A CN 109395375 A CN109395375 A CN 109395375A
Authority
CN
China
Prior art keywords
data
hand
augmented reality
mobile
coordinate system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811086686.5A
Other languages
Chinese (zh)
Inventor
杜广龙
李斌冰
李方
张平
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201811086686.5A priority Critical patent/CN109395375A/en
Publication of CN109395375A publication Critical patent/CN109395375A/en
Pending legal-status Critical Current

Links

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/212Input arrangements for video game devices characterised by their sensors, purposes or types using sensors worn by the player, e.g. for measuring heart beat or leg activity
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/428Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Cardiology (AREA)
  • General Health & Medical Sciences (AREA)
  • Heart & Thoracic Surgery (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention proposes a kind of based on augmented reality and the mobile 3d gaming method of interface interacted.This method is divided into four steps: data collection, and data calculate, and data transmission, hand is rebuild.Data collection section is completed the acquisition to hand position and bearing data by Leap Motion;In data calculating section, the PC connecting with Leap Motion carries out coordinate system conversion to the hand data that acquisition comes and reduces noise data using Interval Kalman filter device and section particle filter, calculates hand direction;In TCP data segment, PC sends the data handled well to by server the client on HoloLens;In scene phase of regeneration, hand is rebuild out by HoloLens using the data received in virtual scene.The present invention can capture hand motion not against any handheld device, improve man-machine interactive compatibility, enhance the sense of reality of game.

Description

A kind of 3d gaming method of interface interacted based on augmented reality and movement
Technical field
The present invention relates to a kind of human-computer interaction technologies, and in particular to a kind of side that human-computer interaction is carried out based on hand motion Method.
Background technique
Human-computer interaction is always the historical pith of development of computer.From earliest machine language, mouse finally Mark, keyboard, and now universal touch screen etc..Keep man-machine interactive process more intuitive naturally, existing for computer The development and application of every field has tremendous influence.
In recent years, human-computer interaction develops a variety of different interaction channels, including speech recognition, bio-identification and body-sensing are handed over Mutually.Speech recognition application identifies robot in intelligent sounds such as Cortana, Siri;Bio-identification mainly utilizes human body distinctive The physiological characteristics such as fingerprint, iris complete the functions such as mobile phone unlock, daily payment;In terms of body feeling interaction, Kinect is known as drawing The product in epoch, as a 3-D camera, it can capture the depth of human body and be moved in the case where not wearing any equipment Make.
And in aspects of game play, main interactive mode is to complete by game paddle or mouse and keyboard to game object Control.AR or VR game newly developed needs handheld device to control game object.Although some game using Kinect into Row body feeling interaction, but this interaction can be completed without any handheld device, but Kinect can not be moved, it means that dynamically catch It is fixed for catching range.In order to keep player more intuitive naturally, we are captured using Leap Motion when controlling game object Player's hand motion, i.e. player can complete interactive process by naked hand, without any handheld device.On the other hand, it passes The scope of activities of the game interaction of system is excessively limited to, we devise a kind of moveable wearable device, no matter being in player Where game interaction can be carried out.
Summary of the invention
In order to realize that naked hand interaction and the moveable function of equipment, the present invention provide a kind of based on augmented reality and movement Interactive 3d gaming interface method.
The purpose of the present invention is achieved through the following technical solutions.
A kind of 3d gaming method of interface interacted based on augmented reality and movement, is captured by Leap Motion Hand motion simultaneously obtains coordinate data, carries out coordinate system conversion to data, reduces the white noise of data, calculated by computer Hand finally is rebuild out by hand direction in virtual scene.
Further, method includes the following steps:
S1, hand original coordinates and bearing data are obtained;
S2, coordinate system conversion is carried out, calculates accurate hand position and direction;
S3, the client sent by server treated data in AR equipment;
S4, AR equipment reconstruct hand model in virtual scene.
Further, step S1 obtains the three-dimensional coordinate information and bearing data of hand using Leap Motion.
Further, the original right hand cartesian coordinate that step S2 is obtained Leap Motion using cartesian coordinate system System is converted into world coordinate system, and is mapped in the Virtual Space that AR equipment is established;Use the area Interval Kalman filter Qi He Between particle filter removal due to calculating mass data and the white noise that generates, make calculated hand position data and direction number According to more accurate.
Further, the server at the end PC transmits the data calculated to the client in AR equipment in step S3 End.
Further, AR equipment used in step S4 is HoloLens, and HoloLens is by the data handled well virtual It rebuilds out in scene.
The present invention in conjunction with AR glasses HoloLens, incites somebody to action gesture induction device Leap Motion using cartesian coordinate system Hand coordinate system is converted into world coordinate system and projects in the virtual scene of AR glasses building, allows equipment with user Movement and move, expand the range of AR scene of game.The present invention directly acquires the number of coordinates for surveying naked hand using Leap Motion According to direction of rotation data, and reduced due to long-time data stacking using Interval Kalman filter device and section particle filter The white noise of generation, accurately calculates hand position and direction, enhances the Stability and veracity of interactive system.
Compared with prior art, the present invention having the following advantages that and technical effect:
1, body feeling interaction is carried out without handheld device, can directly captures naked manual work, makes the interactive process of game more Add intuitive naturally, giving user more true game experiencing.
2, equipment is removable.The disadvantage of excessive, the mobile inconvenience of traditional body-sensing equipment volume is overcome, the user's of increase Scope of activities.
Detailed description of the invention
Fig. 1 is a kind of based on augmented reality and the mobile 3d gaming method of interface flow chart interacted.
Specific embodiment
Specific implementation of the invention is described further below in conjunction with attached drawing and example, but implementation and protection of the invention It is without being limited thereto.If it is noted that having process (such as Interval Kalman filter and particle filter of not special detailed description below Algorithm), it is that those skilled in the art can refer to prior art understanding or realize.
At this stage there are three types of main man-machine interaction modes: 1) mouse;2) touch screen;3) body-sensing.And in the development side of game Face, the game of mouse interactive class and touch screen interactive class game are more, but both interactive modes are for based on AR technological development It is not applicable for game.And game this respect is being carried out by body feeling interaction, and mainly rely on Kinect --- it is a 3-D cameras capture human action completes interaction.However, Kinect device is more huge, need to be fixedly mounted, therefore player Scope of activities is restricted.The characteristics of AR technology is that virtual 3-D object is shown on the basis of realistic space, and this requires AR to swim Play can obtain more living scenes, avoid the dull and dull of game, and the irremovable feature of Kinect leads to its nothing The exploitation of method adaptation AR game.
It is provided in this embodiment a kind of removable based on augmented reality in order to solve the above-mentioned problems of the prior art Game interaction mode.Using Leap Motion and HoloLens small in size, light-weight as interactive device, with this reality The existing moveable feature of equipment.Leap Motion captures hand data, carries out coordinate conversion to data by the end PC and accurately estimates The data handled well are reconstructed to come in Virtual Space by meter using HoloLens.
As shown in Figure 1.The entire method of this example includes following four step:
S1, the three-dimensional coordinate information and bearing data that hand is obtained using Leap Motion;
S2, world coordinates is converted into using the original right-handed Cartesian coordinate system that cartesian coordinate system obtains Leap Motion System, and be mapped in the Virtual Space that HoloLens is established;It is gone using Interval Kalman filter device and section particle filter Except the white noise generated due to calculating mass data, keep calculated hand position data and bearing data more accurate;
S3, the client sent by server treated data on HoloLens;
S4, HoloLens reconstruct hand model in virtual scene.
The 3d gaming method of interface interacted based on augmented reality and movement of the present embodiment.This method is divided into four Step: data collection, data calculate, and data transmission, hand is rebuild.Data collection section is completed by Leap Motion to hand The acquisition of position and direction data;In data calculating section, hand data that the PC connecting with Leap Motion carrys out acquisition into Row coordinate system converts and uses Interval Kalman filter device and section particle filter to reduce noise data, calculates hand direction; In TCP data segment, PC sends the data handled well to by server the client on HoloLens;In scene rebuilding Hand is rebuild out by stage, HoloLens using the data received in virtual scene.The present invention is not against any Handheld device can capture hand motion, improve man-machine interactive compatibility, enhance the sense of reality of game.

Claims (6)

1. a kind of based on augmented reality and the mobile 3d gaming method of interface interacted, which is characterized in that pass through Leap Motion captures hand motion and obtains coordinate data, carries out coordinate system conversion to data, reduces the white noise of data, pass through meter Calculation machine calculates hand direction, finally rebuilds out in virtual scene by hand.
2. according to claim 1 a kind of based on augmented reality and the mobile 3d gaming interface method interacted, feature Be method includes the following steps:
S1, hand original coordinates and bearing data are obtained;
S2, coordinate system conversion is carried out, calculates accurate hand position and direction;
S3, the client sent by server treated data in AR equipment;
S4, AR equipment reconstruct hand model in virtual scene.
3. according to claim 1 a kind of based on augmented reality and the mobile 3d gaming interface method interacted, feature It is that step S1 obtains the three-dimensional coordinate information and bearing data of hand using Leap Motion.
4. according to claim 1 a kind of based on augmented reality and the mobile 3d gaming interface method interacted, feature It is that step S2 is converted into the world using the original right-handed Cartesian coordinate system that cartesian coordinate system obtains Leap Motion and sits Mark system, and be mapped in the Virtual Space that AR equipment is established;It is gone using Interval Kalman filter device and section particle filter Except the white noise generated due to calculating mass data, keep calculated hand position data and bearing data more accurate.
5. according to claim 1 a kind of based on augmented reality and the mobile 3d gaming interface method interacted, feature It is that the server at the end PC in step S3 transmits the data calculated to the client in AR equipment.
6. according to claim 1 a kind of based on augmented reality and the mobile 3d gaming interface method interacted, feature It is that AR equipment used in step S4 is HoloLens, HoloLens structure again in virtual scene by the data handled well It builds out.
CN201811086686.5A 2018-09-18 2018-09-18 A kind of 3d gaming method of interface interacted based on augmented reality and movement Pending CN109395375A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811086686.5A CN109395375A (en) 2018-09-18 2018-09-18 A kind of 3d gaming method of interface interacted based on augmented reality and movement

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811086686.5A CN109395375A (en) 2018-09-18 2018-09-18 A kind of 3d gaming method of interface interacted based on augmented reality and movement

Publications (1)

Publication Number Publication Date
CN109395375A true CN109395375A (en) 2019-03-01

Family

ID=65464968

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811086686.5A Pending CN109395375A (en) 2018-09-18 2018-09-18 A kind of 3d gaming method of interface interacted based on augmented reality and movement

Country Status (1)

Country Link
CN (1) CN109395375A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110070777A (en) * 2019-06-13 2019-07-30 大连民族大学 A kind of Hezhe's fish-skin draws simulation training system and implementation method
CN110561424A (en) * 2019-07-28 2019-12-13 华南理工大学 online robot kinematic calibration method based on multi-sensor hybrid filter
CN110877335A (en) * 2019-09-27 2020-03-13 华南理工大学 Self-adaptive unmarked mechanical arm track tracking method based on hybrid filter
CN112000219A (en) * 2020-03-30 2020-11-27 华南理工大学 Movable gesture interaction device and method for augmented reality game

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101807114A (en) * 2010-04-02 2010-08-18 浙江大学 Natural interactive method based on three-dimensional gestures
US20120270653A1 (en) * 2011-04-25 2012-10-25 Imimtek, Inc. Method and system to create three-dimensional mapping in a two-dimensional game
CN104007819A (en) * 2014-05-06 2014-08-27 清华大学 Gesture recognition method and device and Leap Motion system
CN106444861A (en) * 2016-11-21 2017-02-22 清华大学深圳研究生院 Space robot teleoperation system based on three-dimensional gestures
CN108334199A (en) * 2018-02-12 2018-07-27 华南理工大学 The multi-modal exchange method of movable type based on augmented reality and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101807114A (en) * 2010-04-02 2010-08-18 浙江大学 Natural interactive method based on three-dimensional gestures
US20120270653A1 (en) * 2011-04-25 2012-10-25 Imimtek, Inc. Method and system to create three-dimensional mapping in a two-dimensional game
CN104007819A (en) * 2014-05-06 2014-08-27 清华大学 Gesture recognition method and device and Leap Motion system
CN106444861A (en) * 2016-11-21 2017-02-22 清华大学深圳研究生院 Space robot teleoperation system based on three-dimensional gestures
CN108334199A (en) * 2018-02-12 2018-07-27 华南理工大学 The multi-modal exchange method of movable type based on augmented reality and device

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110070777A (en) * 2019-06-13 2019-07-30 大连民族大学 A kind of Hezhe's fish-skin draws simulation training system and implementation method
CN110070777B (en) * 2019-06-13 2021-07-09 大连民族大学 Huchizhui fish skin painting simulation training system and implementation method
CN110561424A (en) * 2019-07-28 2019-12-13 华南理工大学 online robot kinematic calibration method based on multi-sensor hybrid filter
CN110877335A (en) * 2019-09-27 2020-03-13 华南理工大学 Self-adaptive unmarked mechanical arm track tracking method based on hybrid filter
CN112000219A (en) * 2020-03-30 2020-11-27 华南理工大学 Movable gesture interaction device and method for augmented reality game
CN112000219B (en) * 2020-03-30 2022-06-14 华南理工大学 Movable gesture interaction method for augmented reality game

Similar Documents

Publication Publication Date Title
CN109003301B (en) Human body posture estimation method based on OpenPose and Kinect and rehabilitation training system
CN109395375A (en) A kind of 3d gaming method of interface interacted based on augmented reality and movement
CN107632699B (en) Natural human-machine interaction system based on the fusion of more perception datas
Ren et al. Depth camera based hand gesture recognition and its applications in human-computer-interaction
WO2021169839A1 (en) Action restoration method and device based on skeleton key points
Wang et al. Real-time hand-tracking with a color glove
CN102638653B (en) Automatic face tracing method on basis of Kinect
Segen et al. Human-computer interaction using gesture recognition and 3D hand tracking
CN1304931C (en) Head carried stereo vision hand gesture identifying device
CN110728739B (en) Virtual human control and interaction method based on video stream
WO2023071964A1 (en) Data processing method and apparatus, and electronic device and computer-readable storage medium
CN108509026B (en) Remote maintenance support system and method based on enhanced interaction mode
CN109885163A (en) A kind of more people's interactive cooperation method and systems of virtual reality
JP2004094288A (en) Instructed position detecting device and autonomous robot
WO2008029467A1 (en) Gesture input system, method and program
CN104090660A (en) Motion collecting and feedback method and system based on stereoscopic vision
CN102831380A (en) Body action identification method and system based on depth image induction
Leibe et al. Toward spontaneous interaction with the perceptive workbench
KR20170086317A (en) Apparatus and Method for Generating 3D Character Motion via Timing Transfer
CN102306053B (en) Virtual touch screen-based man-machine interaction method and device and electronic equipment
CN104517100B (en) Gesture pre-judging method and system
CN102830798A (en) Mark-free hand tracking method of single-arm robot based on Kinect
CN115129164B (en) Interaction control method and system based on virtual reality and virtual reality equipment
CN111596767A (en) Gesture capturing method and device based on virtual reality
CN106598211A (en) Gesture interaction system and recognition method for multi-camera based wearable helmet

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20190301

RJ01 Rejection of invention patent application after publication