CN108334199A - The multi-modal exchange method of movable type based on augmented reality and device - Google Patents
The multi-modal exchange method of movable type based on augmented reality and device Download PDFInfo
- Publication number
- CN108334199A CN108334199A CN201810144421.XA CN201810144421A CN108334199A CN 108334199 A CN108334199 A CN 108334199A CN 201810144421 A CN201810144421 A CN 201810144421A CN 108334199 A CN108334199 A CN 108334199A
- Authority
- CN
- China
- Prior art keywords
- augmented reality
- modal
- gesture
- virtual
- interactive
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000003190 augmentative effect Effects 0.000 title claims abstract description 128
- 238000000034 method Methods 0.000 title claims abstract description 33
- 230000002452 interceptive effect Effects 0.000 claims abstract description 74
- 230000003993 interaction Effects 0.000 claims abstract description 51
- 230000004927 fusion Effects 0.000 claims abstract description 29
- 230000000694 effects Effects 0.000 claims abstract description 9
- 238000007500 overflow downdraw method Methods 0.000 claims abstract description 8
- 239000011159 matrix material Substances 0.000 claims description 25
- 230000002708 enhancing effect Effects 0.000 claims description 11
- 230000009466 transformation Effects 0.000 claims description 11
- 230000033001 locomotion Effects 0.000 claims description 9
- 238000013519 translation Methods 0.000 claims description 9
- 230000007246 mechanism Effects 0.000 claims description 5
- 238000009877 rendering Methods 0.000 claims description 5
- 230000005540 biological transmission Effects 0.000 claims description 3
- 238000004891 communication Methods 0.000 claims description 3
- 238000010276 construction Methods 0.000 claims description 3
- 230000008569 process Effects 0.000 claims description 3
- 230000017105 transposition Effects 0.000 claims description 3
- 229910017435 S2 In Inorganic materials 0.000 claims 2
- 108091006146 Channels Proteins 0.000 description 8
- 238000005516 engineering process Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 8
- 230000009471 action Effects 0.000 description 7
- 238000005259 measurement Methods 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 241000208340 Araliaceae Species 0.000 description 2
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 description 2
- 235000003140 Panax quinquefolius Nutrition 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 235000008434 ginseng Nutrition 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 description 1
- 241001269238 Data Species 0.000 description 1
- 241000406668 Loxodonta cyclotis Species 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000000465 moulding Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000002203 pretreatment Methods 0.000 description 1
- 239000000047 product Substances 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 210000000697 sensory organ Anatomy 0.000 description 1
- 238000013179 statistical model Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/183—Speech classification or search using natural language modelling using context dependencies, e.g. language models
- G10L15/187—Phonemic context, e.g. pronunciation rules, phonotactical constraints or phoneme n-grams
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The multi-modal exchange method of movable type and device that the invention discloses a kind of based on augmented reality, this approach includes the following steps:Show that human-computer interaction interface, augmented reality virtual scene include the interactive information such as dummy object by augmented reality mode;User sends interactive instruction by way of gesture and voice, by multi-modal fusion method, understands different modalities semanteme, and merge the modal data of gesture and voice, generates multi-modal fusion interactive instruction;After user's interactive instruction effect, exercising result returns in augmented reality virtual scene, is fed back into row information by the variation of scene.The inventive system comprises gesture sensor, PC machine, microphone, optics through mode augmented realities to show equipment, WiFi routers.The present invention provides a kind of method and devices of combination augmented reality and multi-modal interaction, embody thought focusing on people, naturally intuitive, reduce Learning work load, improve interactive efficiency.
Description
Technical field
The present invention relates to human-computer interaction technique fields, and in particular to a kind of multi-modal interaction of movable type based on augmented reality
Method and device.
Background technology
Augmented reality (Augmented Reality, AR) is disappearing with the rapid development of computer technology, in recent years
Expense causes huge concern in the market, and various products emerge one after another, and have started one vision tide of revolution.Augmented reality skill
Art is a kind of technology merging real scene with virtual scene, its purpose is by computer graphical, image processing techniques
Realize the synthesis of outdoor scene (display environment or user image) and empty scape (virtual environment or dummy object that computer generates).
Similarly, multi-modal human-computer interaction technology is also that current field of human-computer interaction is studied with being widely studied.It is multi-modal
Man-machine interaction mode applies multiple natural interactive styles, so that the perception mode of people is fully used, multiple interaction modalities are not with
Same interactive mode cooperating is realized and more freely and is naturally communicated.Multi-modal interaction is not merely using multiple logical
Road is independently completed task, but is integrated the interactive information between user's difference channel by Multimodal Integration technology, passes through
Interaction between multiple channels forms final interaction and is intended to, and is properly completed task.
However, although current augmented reality possesses the visual display mode different from tradition, in a manner of holographic
More information are provided, wearable AR equipment also has good mobility and portability, but its shortage nature is intuitive, high
The interactive mode of effect interacts so that user experience is bad typically only by controller or simple voice or gesture.
And current multi-modal interactive mode can unify different sense organ mode and realize that nature is intuitive, efficiently interact, but be only applied to
Desk device lacks good portability and mobility.
Invention content
The purpose of the present invention is to solve drawbacks described above in the prior art, provide a kind of movement based on augmented reality
The multi-modal exchange method of formula and device merge multiple perception mode, and information feedback is realized by augmented reality, and enhancing is existing
The enhancement of real technology, portability and the interactivity of multi-modal interaction organically combine, and realize intuitive nature, low Learning work load, height
Interactive efficiency has portability, ambulant man-machine interaction mode simultaneously.
According to disclosed embodiment, it is multi-modal that the first aspect of the present invention discloses a kind of movable type based on augmented reality
Exchange method, the multi-modal exchange method include the following steps:
S1, human-computer interaction interface, the interactive information of augmented reality virtual scene are shown by augmented reality mode;
S2, user pass through gesture and the multi-modal interactive mode of voice and the virtual interacting pair in augmented reality virtual scene
As interacting;
S3, by multi-modal fusion method, understand different modalities semanteme, and merge the modal data of gesture and voice, produce
Raw multi-modal fusion interactive instruction;
After S4, user's interactive instruction effect, exercising result returns in augmented reality virtual scene, passes through the variation of scene
It is fed back into row information.
Further, the virtual reality scene described in step S1 includes virtual interacting object and virtual information object,
In, the virtual interacting object possesses multi-modal interaction capabilities and Informational Expression ability;The virtual information object, is gathered around
There is Informational Expression ability.
Further, interactive instruction operation is sent by gesture in step S2, wherein the interactive object of gesture is that enhancing is existing
Virtual interacting object in real virtual scene, interactive mode include:Virtual interacting object is implemented to click, drag or touch behaviour
Make.
Further, to realize that user passes through the virtual interacting object in gesture and augmented reality virtual scene in step S2
Interaction is needed to realize being registrated between gesture sensor coordinate system and augmented reality virtual scene coordinate system, be obtained between the two
Coordinate transformation relation, calculate gesture sensor and augmented reality using Zhang Zhengyou standardizations and show and join outside the interior participation of equipment,
The lens model of Zhang Zhengyou standardizations is as follows:
Wherein, s is scale factor, [u, v, 1]TFor pixel planes coordinate, [Xw,Yw,Zw,1]TFor the coordinate of world coordinate system
Point, [R, T] i.e.Matrix is outer parameter, and R is spin matrix, and t is translation vector, the transposition of T representing matrixes,For intrinsic parameter K, F is the focal length of video camera,[u0,v0]TFor coordinate of the camera coordinate system origin in image coordinate system, dx and dy are picture
The length of side of element, unit mm, K=K1K2;
According to homography, the relationship between image obtained by plane reference plate and video camera is as follows:
Wherein r1、r2、r3Expansion for spin matrix R in x, y, z direction, it is assumed that the point on plane reference plate is sat in the world
Z coordinate in mark system is 0, then formula (2) homography relationship is simplified as:
Wherein, K [r1r2T] it is homography matrix H, it enablesAbove-mentioned formula can be reduced toWherein:
H=[h1 h2 h3]=λ K [r1 r2 t] (10)
According to the characteristic of spin matrix, constrained as follows:With ‖ r1‖=‖ r2‖=1, can according to formula (4)
Know:
Formula (5), which is substituted into above-mentioned constraint, to be obtained:
I.e. each homography matrix can provide 2 equations, and internal reference matrix includes 5 parameters, it is desirable that solution at least needs 3
A homography matrix, it is therefore desirable to which the picture of three width plane reference plates obtains three groups of formulas (6) to calculate internal reference, further according to formula
(5) the outer relationship joined is participated in calculates outer parameter.
Further, the multi-modal fusion method described in step S3 uses the hierarchical fusion model of oriented mission.
Further, which is characterized in that the realization process of the hierarchical fusion model of the oriented mission is as follows:Pass through word
The input form in the unified different channels of method layer, the same content in different channels is expressed using same primitive;From morphology layer
Primitive information is divided into the primitive, the primitive for indicating object, the primitive for indicating object properties for indicating order according to syntax gauge;It is semantic
Layer utilizes task-driven mechanism, finally by primitive combination at various specific tasks.
Further, carrying out information feedback method by augmented reality virtual scene in step S4 includes:By virtually believing
Cease object display text and graphical information;Pass through the status and appearance of virtual interacting object.
According to disclosed embodiment, it is multi-modal that the second aspect of the present invention discloses a kind of movable type based on augmented reality
Interactive device, the multi-modal interactive device include gesture sensor, PC machine, microphone, augmented reality show equipment,
In, the gesture sensor is mounted on the augmented reality by support construction and shows in equipment that data-interface passes through USB
The mode of data line is connected with the PC machine, hand gesture location and posture for capturing controller;
The microphone is mounted on the augmented reality and shows in equipment, data-interface by USB data line with
PC machine is connected, the phonetic control command for capturing controller;
The augmented reality shows that equipment for rendering and showing augmented reality virtual scene, passes through augmented reality skill
Art is superimposed virtual scene in true environment, provides the auxiliary information that real world can not obtain, user is to real world for enhancing
Sensing capability and interaction capabilities with real world;
The PC machine, data for identification from gesture mode and speech modality simultaneously carry out multi-modal fusion, described
The PC machine interaction results that instruct multi-modal fusion show equipment by wireless network transmissions to the augmented reality, pass through
Virtual objects in augmented reality virtual scene realize the feedback of interactive information.
Further, the multi-modal interactive device further includes WiFi routers, and the PC machine is route by WiFi
Device and the augmented reality carry out wireless communication between showing equipment.
Further, the gesture sensor uses Leap Motion, the augmented reality to show that equipment uses
HoloLens。
The present invention has the following advantages and effects with respect to the prior art:
1, the present invention provides a kind of multi-modal exchange method of movable type based on augmented reality, can effectively organize gesture
Mode and speech modality, it is different from the interactive mode of conventional serial, it realizes the parallel and cooperating operation between different modalities, realizes
More naturally intuitive interactive mode.
2, the co-registration of coordinate systems used between equipment is shown by gesture sensor and augmented reality, can realize gesture with it is virtual
Direct interaction between scene realizes the high efficiency interactive of user and augmented reality without extras such as controllers.
3, the implementing result that multi-modal fusion interacts is fed back to by augmented reality in virtual scene, utilizes enhancing
The enhancement of reality technology provides three-dimensional, intuitive feedback information.
4, a kind of multi-modal interactive device of movable type based on augmented reality provided by the invention, is increased using optics through mode
Strong reality head-mounted display HoloLens and gesture sensor Leap Motion, is combined the two by bindiny mechanism, is realized
The portability and mobility of the interactive device so that user also can normally implement to interact outdoors, under mobile working environment.
Description of the drawings
Fig. 1 is a kind of interaction flow of multi-modal exchange method of movable type based on augmented reality in the embodiment of the present invention
Figure;
Fig. 2 is a kind of interaction diagrams of specific implementation scene in the embodiment of the present invention;
Fig. 3 is layering task model structure chart in the embodiment of the present invention;
Fig. 4 is general task slot structure figure in the embodiment of the present invention;
Fig. 5 is multi-modal fusion algorithm flow chart in the embodiment of the present invention;
Fig. 6 is the system assumption diagram that task model is layered in the embodiment of the present invention;
Fig. 7 is a kind of composition figure of the multi-modal interactive device of movable type based on augmented reality in the embodiment of the present invention.
Specific implementation mode
In order to make the object, technical scheme and advantages of the embodiment of the invention clearer, below in conjunction with the embodiment of the present invention
In attached drawing, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that described embodiment is
A part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, those of ordinary skill in the art
The every other embodiment obtained without making creative work, shall fall within the protection scope of the present invention.
Embodiment
A kind of multi-modal exchange method of movable type and device based on augmented reality are provided respectively in the present embodiment, realized
Naturally intuitive, low Learning work load, high interactive efficiency have portability, ambulant man-machine interaction mode simultaneously.It is existing by enhancing
Real mode shows human-computer interaction interface, and augmented reality virtual scene includes the interactive information such as dummy object;User by gesture and
The mode of voice sends interactive instruction, by multi-modal fusion method, understands different modalities semanteme, and merges gesture and voice
Modal data generates multi-modal fusion interactive instruction;After user's interactive instruction effect, it is virtual that exercising result returns to augmented reality
In scene, fed back into row information by the variation of scene.
As shown in fig. 7, a kind of multi-modal interactive device of movable type based on augmented reality provided in the present embodiment includes
Gesture sensor, PC machine, microphone, optics through mode augmented reality show equipment, WiFi routers, wherein:
Gesture sensor, gesture sensor (i.e. Leap Motion in attached drawing 7) are mounted on augmented reality by support construction
Show in equipment, data-interface is connected by way of USB data line with PC machine, for capture the hand gesture location of controller with
Posture.
PC machine is connected by USB data line with gesture sensor, microphone, comes from gesture mode and voice for identification
The data of mode simultaneously carry out multi-modal fusion;The interaction results that multi-modal fusion instructs are passed through into wireless network transmissions to optical lens
It crosses formula augmented reality and shows equipment, the feedback of interactive information is realized by the virtual objects in augmented reality virtual scene;And lead to
It crosses between WiFi routers show equipment with augmented reality and is communicated.
Microphone is mounted on augmented reality and shows in equipment that data-interface is connected by USB data line with PC machine, uses
In the phonetic control command of capture controller;
Augmented reality shows that equipment, augmented reality show that equipment (i.e. HoloLens in attached drawing 7) is responsible for rendering and display increases
Strong virtual reality scene is superimposed virtual scene by augmented reality in true environment, and providing real world can not obtain
Auxiliary information, enhancing user to the sensing capability of real world and with the interaction capabilities of real world;
WiFi routers provide the wireless network environment that PC machine shows communication between devices with augmented reality.
The device can be divided into following functions module again according to functional characteristics:Augmented reality display module, gesture are defeated
Enter module, voice input module and multi-modal understanding and Fusion Module.
Augmented reality display module:It is responsible for rendering and showing augmented reality virtual scene, is mainly enhanced by optics through mode
Reality glasses form, and Microsoft's HoloLens equipment is used in the present embodiment.By augmented reality, it is superimposed in true environment
Virtual scene, provides the auxiliary information that real world can not obtain, enhancing user to the sensing capability of real world and with it is true
The interaction capabilities in the world;Its own possesses and maintains a virtual scene coordinate system, is same as rendering and handling virtual scene object;
Simultaneously because the multi-modal interactive device of movable type based on augmented reality should have portability and mobility, augmented reality is aobvious
Show that module shows equipment using wear-type augmented reality.
Gesture input module:The acquisition and further processing, the present embodiment for being responsible for gesture interaction data use Leap
Motion gesture sensors.Gesture input module is based on binocular camera Depth Imaging, for realizing following steps:It collects and obtains
Hand images.Hand Gesture Segmentation is carried out to image, gesture is split from image background.Gesture model is established, a system is passed through
The parameter of row describes gesture.Gesture feature is extracted, corresponding characteristic parameter parameter is extracted from gesture according to the model of foundation;Its
Owned gesture input module coordinate system, for describing gesture and gesture feature data;Due to described based on augmented reality
Mobile multi-modal interactive device should have portability and mobility, and the gesture sensor that gesture input module uses will be installed extremely
The wear-type augmented reality of augmented reality display module shows that upper side, induction range follow wear-type augmented reality to show
Equipment moves, it is ensured that the normal work under situation of movement.
Voice input module:It is responsible for the acquisition of interactive voice data and further processing, is mainly made of microphone.It is right
After voice carries out the pretreatments such as preemphasis, end-point detection, the redundancy in voice data is removed, its feature such as mel cepstrum system is extracted
Number, and statistical model is trained with this, sound bank is obtained, recognition result is obtained finally by pattern match.Equally, in order to ensure
The portability and mobility of the multi-modal interactive device of movable type based on augmented reality, the Mike that voice input module uses
Wind is mounted on augmented reality and shows in equipment.
Multi-modal understanding and Fusion Module:It is responsible for the raw information from mechanical floor carrying out unification in multi-modal understanding part
Processing, meaning is identical and input that form is different is same is indicated for identical information, to grammer layer provide with equipment without
The information of pass, i.e. Interaction function;And multi-modal fusion part is then the task model based on layering, is appointed according to what user was planned
Business fills the Interaction function in different channels into corresponding task slot, is finally fused into goal task.It is multi-modal to understand and melt
Molding block is mainly realized in PC machine.
The workflow of the multi-modal exchange method of movable type based on augmented reality based on above-mentioned apparatus and function module
As shown in Figure 1, including the following steps:
S1, show that human-computer interaction interface, augmented reality virtual scene include the friendships such as dummy object by augmented reality mode
Mutual information.
Step S1 is by augmented reality display module by information with the virtual interacting object of virtual scene and virtual display
The mode of object is shown;
S2, user pass through gesture and the multi-modal interactive mode of voice and the virtual interacting pair in augmented reality virtual scene
As interacting.
The step is virtual using multichannel interaction manner and augmented reality by gesture input module and voice input module
Virtual interacting object in scene interacts;
In step s 2, the gesture of user is needed with virtual interacting object interaction directly in virtual scene by gesture number
It is converted to augmented reality virtual scene coordinate system according to from the coordinate system of gesture input module.
It realizes that user is interactive by the virtual interacting object in gesture and augmented reality virtual scene, needs to realize gesture
Being registrated between sensor coordinate system and augmented reality virtual scene coordinate system, obtains coordinate transformation relation between the two.
The interior participation of gesture input module and augmented reality display module is calculated in the present embodiment using Zhang Zhengyou standardizations
The lens model of outer ginseng, Zhang Zhengyou standardizations is as follows:
Wherein, s is scale factor, [u, v, 1]TFor pixel planes coordinate, [Xw,Yw,Zw,1]TFor the coordinate of world coordinate system
Point, [R, T] i.e.Matrix is outer parameter, and R is spin matrix, and t is translation vector, the transposition of T representing matrixes,For intrinsic parameter K, F is the focal length of video camera,[u0,v0]TFor coordinate of the camera coordinate system origin in image coordinate system, dx and dy are picture
The length of side of element, unit mm, K=K1K2。
Since Zhang Zhengyou standardizations are that one kind being based on the tessellated calibration of plane, it is flat to another to be transformed to a plane
The projection mapping in face, as homography.
According to homography, the relationship between image obtained by plane reference plate and video camera is as follows:
Wherein r1、r2、r3Expansion for spin matrix R in x, y, z direction, it is assumed that the point on plane reference plate is sat in the world
Z coordinate in mark system is 0, then formula (2) homography relationship is simplified as:
Wherein, K [r1 r2T] it is homography matrix H, it enablesAbove-mentioned formula can simplify
ForWherein:
H=[h1 h2 h3]=λ K [r1 r2 t] (4)
According to the characteristic of spin matrix, it is easy to get following constraint:With ‖ r1‖=‖ r2‖=1, according to formula (4)
It is readily apparent that:
Formula (5), which is substituted into above-mentioned constraint, to be obtained:
I.e. each homography matrix can provide 2 equations, and internal reference matrix includes 5 parameters, it is desirable that solution at least needs 3
A homography matrix, it is therefore desirable to which the picture of three width plane reference plates is to calculate internal reference, further according to the pass joined outside interior participation
System calculates outer parameter.
Obtain after joining outside the interior participation of augmented reality display module and gesture sensor, you can with both calculate coordinate system it
Between transformational relation.
If PvsFor the space coordinate that certain in augmented reality display module is put, pvsFor projection coordinate of this in image plane,
HvsFor the internal reference of augmented reality display module, can be obtained according to national forest park in Xiaokeng:
pvs=HvsPvs (7)
Similarly, it can obtain:
ph=HhPh (8)
Wherein set PhFor the space coordinate that certain in gesture sensor is put, phFor projection coordinate of this in image plane, HhFor
The internal reference of gesture sensor.Set up an office pvsWith point phFor same point in space, therefore it is real to pass through the transformation of coordinate rotation and translation
Now transformation of this o'clock between two coordinate systems, is denoted as Pvs=RPh+ T, wherein R are rotation transformation, and T is translation transformation.Simultaneously should
The P of pointvsAnd PhExpression can be obtained by coordinate transform according to global coordinate system, that is, scaling board coordinate, such as formula Pvs=RvsP
+TvsAnd Ph=RhP+Th, wherein Rvs、Rh、TvsAnd ThIt is that equipment video camera and hand are shown from global coordinate system to augmented reality respectively
The rotation transformation of gesture sensor and translation transformation, value can be obtained from the outer ginseng matrix of calibration.Slightly being converted according to upper two formula can
To obtain:
According to coordinates of targets transformation relation Pvs=RPh+ T can be obtained:
Transformation relation can be obtained according to formula (10), finally coordinate can be transformed into enhancing from gesture sensor coordinate system
Reality is shown in equipment camera coordinate system, realizes the registration of the two coordinate system.
S3, by multi-modal fusion method, understand different modalities semanteme, and merge the modal data of gesture and voice, produce
Raw multi-modal fusion interactive instruction.
The initial data of gesture input module and voice input module is transferred to multi-modal understanding and Fusion Module by the step
It is handled and is merged, the interactive task of user is generated according to multi-channel information;
In step s3, it needs to carry out being processed and converted to Interaction function to coming from different each modal datas of input module,
Primitive is classified and combined according to semanteme, forms final interactive task.
The mechanical floor expression of voice input module speech recognition is character string, and gesture input module device layer is then coordinate
Information and click information are handled by multi-modal understanding, and the information of different modalities is indicated using common data structure, is formed
Interaction function.
As shown in figure 3, task model of the multi-modal fusion part using layering, task model is made by the bridge of task
With people carrys out the behavior of organically organizational computing machine with the identity of mission planning person, to turn the random function of computer
Become the implementation method for target, is exactly briefly the intention of people to be communicated to computer by this form of task, and divide
It is mechanical floor, morphology from specific facility information to the final semantic abstraction to be filled that the thought of layer, which is then by modal information,
4 layers of layer, grammer layer and semantic layer.
Due to use be oriented mission multi-modal fusion model, need to define certain structure and appoint to describe interaction
Business.As shown in figure 4, general task structure is made of task action and a series of task parameters, and the present embodiment according to
Task structure is defined as the form of task action, Structure of object attribute and parameter, a certain generic task structure by specific implementation scene
Referred to as task slot.Structure of object attribute is for referring to the object to be interacted, in the interactive process in the present embodiment:Voice is defeated
Enter in " measuring distance here thereunto ", " here " and " there " is the primitive for indicating Structure of object attribute, for indicating
The location information of object.Task action is the core of task structure, to connect different object properties and parameter, in the present embodiment
" measurement " of voice input is just task action.Parameter is the required information of task action, such as " measurement " in the present embodiment
It needs " distance " parameter to supplement, different tasks can be combined by changing parameter.It should be noted that different task actions
Different task structures may be corresponded to, as " the measuring distance here thereunto " in the present embodiment needs one task of filling dynamic
Work, two Structure of object attribute and a parameter, and task " marking this position " then only needs to fill a task and one
Structure of object attribute.It may determine that whether a task should submit execution according to the occupied state of task slot, if task
There is still a need for the further inputs that other information then waits for user for slot, submit explanation to hold immediately if task slot has been filled with and finishes
Row.
Show that multi-modal fusion algorithm flow, step include referring to Fig. 5, Fig. 5:
Step S31 classifies to the incoming event received, if extracting task key word, goes to step S32.Otherwise, it presses
Input channel is respectively filled in Speech time queue or gesture event queue, and forms parameter stack;
Step S32 generates corresponding task slot according to task key word, is put into Interaction context;
Step S33 with the syntax rule of the task in temporal correlation, current Interaction context is about in event queue
Shu Jinhang task slots are filled, and judge whether that filling is complete, if completely, submitting explanation to execute, otherwise returning to waiting for further filling out
It fills;
Step S34 explains execution to task, and empties context.If explaining failure, task can not be executed, be dished out
Exception simultaneously records context.
Referring to the architecture of Fig. 6, Fig. 6 layering task model for showing to use in the present embodiment, structure includes:Event
Subsystem is managed, the input information of input module is responsible for receiving and managing;Event input by user is responsible in work event queue;
Event-action translation table is responsible for mapping the event of work event queue and action;Event argument translation table;Event object attribute
Translation table;Interaction context is responsible for current and historic task slot information;Order integrator is responsible for that task slot will be filled, whole
Synthesis task.
After S4, user's interactive instruction effect, exercising result returns in augmented reality virtual scene, passes through the variation of scene
It is fed back into row information.
It, below will tool in order to preferably describe the movable type multi-modal exchange method proposed by the present invention based on augmented reality
Body introduces step of the method for the present invention in specific application scenarios.Fig. 2 shows a kind of specific implementation scenes of the present embodiment
Interaction flow, the exchange method obtains distance between two points on map through the invention for the specific implementation scene description,
Its key step includes:
Information is shown object with the virtual interacting object of virtual scene and virtually by step R1, augmented reality display module
Mode is shown.Specifically, the augmented reality virtual scene of the specific implementation scene of the present embodiment includes virtual map and letter
Cease overview.Virtual map is virtual interacting object, and user can be interacted by multi-modal interactive mode.And information profile is
Virtual information object, for showing area, weather, traffic conditions etc. where the map display area;
Step R2, user are said by voice:" measuring distance here thereunto ", while user clicks virtual field with hand
Two positions of the virtual map of scape, respectively as beginning and end;
Step R3, voice input module and gesture input module pass the voice of the user of step R2 and gesture initial data
Enter multi-modal understanding and Fusion Module, convert raw data into Interaction function, and is divided into according to syntax rule different classes of
Primitive finally utilizes task-driven mechanism by different primitive combinations, forms final interactive task, friendship is executed by the system;
Range measurement is back in virtual scene by step R4 after system completes range measurement task.Specifically, information
Feedback system can be shown for virtual information object in a manner of text, or the feedback user by way of voice broadcast.
A kind of multi-modal exchange method of movable type based on augmented reality that the present embodiment proposes includes the following steps:Pass through
Augmented reality mode shows that human-computer interaction interface, augmented reality virtual scene include the interactive information such as dummy object;User passes through
Gesture and the mode of voice send interactive instruction, by multi-modal fusion method, understand different modalities semanteme, and merge gesture with
The modal data of voice generates multi-modal fusion interactive instruction;After user's interactive instruction effect, it is existing that exercising result returns to enhancing
In real virtual scene, fed back into row information by the variation of scene.A kind of movement based on augmented reality that the present embodiment proposes
The multi-modal interactive device of formula includes gesture sensor, PC machine, microphone, optics through mode augmented reality display equipment, the roads WiFi
By device.
In the present embodiment, by augmented reality by information with the virtual interacting object of virtual scene and virtual display pair
The mode of elephant is shown.User is empty with augmented reality using multichannel interaction manner by gesture input module and voice input module
Virtual interacting object in quasi- scene interacts.According to interaction generate initial data, it is multi-modal understanding and Fusion Module into
Row realizes the fusion of multi-modal channel interaction, generates the interactive task of user.After interactive task friendship has been executed by the system, execute
As a result augmented reality virtual scene is returned, realizes interactive feedback.By using the information enhancement of augmented reality, extension
User obtains the dimension of information.The parallel of a variety of effect channels and cooperation are realized further through multi-modal fusion technology, provide simultaneously
Integrated and flexible natural interactive style.The interactivity for relying on the enhancement and multi-modal interaction of augmented reality, carries
A kind of naturally intuitive, efficient interactive mode is supplied.
Simultaneously again because of a kind of multi-modal interactive device of movable type based on augmented reality provided by the invention, using optics
Through mode augmented reality head-mounted display HoloLens and gesture sensor Leap Motion, by bindiny mechanism by the two
In conjunction with realizing the portability and mobility of the interactive device so that user outdoors, under mobile working environment also can be normal
Implement interaction, provides good portability and mobility.
The above embodiment is a preferred embodiment of the present invention, but embodiments of the present invention are not by above-described embodiment
Limitation, it is other it is any without departing from the spirit and principles of the present invention made by changes, modifications, substitutions, combinations, simplifications,
Equivalent substitute mode is should be, is included within the scope of the present invention.
Claims (10)
1. a kind of multi-modal exchange method of movable type based on augmented reality, which is characterized in that the multi-modal exchange method
Include the following steps:
S1, human-computer interaction interface, the interactive information of augmented reality virtual scene are shown by augmented reality mode;
S2, user by gesture and the multi-modal interactive mode of voice and the virtual interacting object in augmented reality virtual scene into
Row interaction;
S3, by multi-modal fusion method, understand different modalities semanteme, and merge the modal data of gesture and voice, generate more
Modality fusion interactive instruction;
After S4, user's interactive instruction effect, exercising result returns in augmented reality virtual scene, is carried out by the variation of scene
Information is fed back.
2. the movable type multi-modal exchange method according to claim 1 based on augmented reality, which is characterized in that step S1
Described in augmented reality virtual scene include virtual interacting object and virtual information object, wherein the virtual interacting pair
As possessing multi-modal interaction capabilities and Informational Expression ability;The virtual information object, possesses Informational Expression ability.
3. the movable type multi-modal exchange method according to claim 1 based on augmented reality, which is characterized in that step S2
In pass through gesture send interactive instruction operation, wherein the interactive object of gesture be augmented reality virtual scene in virtual interacting
Object, interactive mode include:Virtual interacting object is implemented to click, drag or touch operation.
4. the movable type multi-modal exchange method according to claim 3 based on augmented reality, which is characterized in that step S2
In to realize that user is interactive by the virtual interacting object in gesture and augmented reality virtual scene, need to realize gesture sensor
Being registrated between coordinate system and augmented reality virtual scene coordinate system, obtains coordinate transformation relation between the two, using just
Friendly standardization calculates gesture sensor and shown with augmented reality joins outside the interior participation of equipment, and the lens model of Zhang Zhengyou standardizations is such as
Under:
Wherein, s is scale factor, [u, v, 1]TFor pixel planes coordinate, [Xw,Yw,Zw,1]TFor the coordinate points of world coordinate system,
[R, T] i.e.Matrix is outer parameter, and R is spin matrix, and t is translation vector, the transposition of T representing matrixes,For intrinsic parameter K, F is the focal length of video camera,[u0,v0]TFor coordinate of the camera coordinate system origin in image coordinate system, dx and dy are picture
The length of side of element, unit mm, K=K1K2;
According to homography, the relationship between image obtained by plane reference plate and video camera is as follows:
Wherein r1、r2、r3Expansion for spin matrix R in x, y, z direction, it is assumed that the point on plane reference plate is in world coordinate system
In Z coordinate be 0, then formula (2) homography relationship be simplified as:
Wherein, K [r1 r2T] it is homography matrix H, it enablesAbove-mentioned formula can be reduced toWherein:
H=[h1 h2 h3]=λ K [r1 r2 t] (4)
According to the characteristic of spin matrix, constrained as follows:With ‖ r1‖=‖ r2‖=1, according to formula (4):
Formula (5), which is substituted into above-mentioned constraint, to be obtained:
I.e. each homography matrix can provide 2 equations, and internal reference matrix includes 5 parameters, it is desirable that solution at least needs 3 lists
Answering property matrix, it is therefore desirable to which the picture of three width plane reference plates obtains three groups of formulas (6) to calculate internal reference, further according to formula (5)
The relationship joined outside interior participation calculates outer parameter.
5. the movable type multi-modal exchange method according to claim 1 based on augmented reality, which is characterized in that step S3
Described in multi-modal fusion method use oriented mission hierarchical fusion model.
6. the movable type multi-modal exchange method according to claim 5 based on augmented reality, which is characterized in that described
The realization process of the hierarchical fusion model of oriented mission is as follows:By the input form in the unified different channels of morphology layer, using same
One primitive expresses the same content in different channels;Primitive information from morphology layer is divided into according to syntax gauge and indicates order
Primitive, the primitive for indicating object, the primitive for indicating object properties;Semantic layer utilizes task-driven mechanism, finally by primitive combination
At various specific tasks.
7. the movable type multi-modal exchange method according to claim 1 based on augmented reality, which is characterized in that step S4
In by augmented reality virtual scene carry out information feedback method include:Believed by virtual information object display text and figure
Breath;Pass through the status and appearance of virtual interacting object.
8. a kind of multi-modal interactive device of movable type based on augmented reality, which is characterized in that the multi-modal interactive device
Equipment is shown including gesture sensor, PC machine, microphone, augmented reality, wherein the gesture sensor is pacified by support construction
It shows in equipment that data-interface is connected by way of USB data line with the PC machine mounted in the augmented reality, uses
In the hand gesture location and posture of capture controller;
The microphone is mounted on the augmented reality and shows in equipment that data-interface passes through USB data line and PC machine
It is connected, the phonetic control command for capturing controller;
The augmented reality shows equipment for rendering and showing augmented reality virtual scene, by augmented reality,
It is superimposed virtual scene in true environment, the auxiliary information that real world can not obtain, sense of the enhancing user to real world are provided
Know ability and the interaction capabilities with real world;
The PC machine, data for identification from gesture mode and speech modality simultaneously carry out multi-modal fusion, the PC
The interaction results that machine instructs multi-modal fusion show equipment by wireless network transmissions to the augmented reality, pass through enhancing
Virtual objects in virtual reality scene realize the feedback of interactive information.
9. the movable type multi-modal interactive device according to claim 8 based on augmented reality, which is characterized in that described
Multi-modal interactive device further includes WiFi routers, and the PC machine is shown with the augmented reality by WiFi routers and set
Wireless communication is carried out between standby.
10. the movable type multi-modal interactive device according to claim 8 based on augmented reality, which is characterized in that described
Gesture sensor use Leap Motion, the augmented reality show equipment use HoloLens.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810144421.XA CN108334199A (en) | 2018-02-12 | 2018-02-12 | The multi-modal exchange method of movable type based on augmented reality and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810144421.XA CN108334199A (en) | 2018-02-12 | 2018-02-12 | The multi-modal exchange method of movable type based on augmented reality and device |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108334199A true CN108334199A (en) | 2018-07-27 |
Family
ID=62929256
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810144421.XA Pending CN108334199A (en) | 2018-02-12 | 2018-02-12 | The multi-modal exchange method of movable type based on augmented reality and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108334199A (en) |
Cited By (34)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109214006A (en) * | 2018-09-18 | 2019-01-15 | 中国科学技术大学 | The natural language inference method that the hierarchical semantic of image enhancement indicates |
CN109343703A (en) * | 2018-09-10 | 2019-02-15 | 中国科学院计算机网络信息中心 | Information processing method, device, system, storage medium and processor |
CN109395375A (en) * | 2018-09-18 | 2019-03-01 | 华南理工大学 | A kind of 3d gaming method of interface interacted based on augmented reality and movement |
CN109461351A (en) * | 2018-09-28 | 2019-03-12 | 中国科学院苏州生物医学工程技术研究所 | The augmented reality game training system of three screens interaction |
CN109766003A (en) * | 2018-12-29 | 2019-05-17 | 北京诺亦腾科技有限公司 | Object methods of exhibiting and device in a kind of VR scene |
CN109782907A (en) * | 2018-12-28 | 2019-05-21 | 西安交通大学 | A kind of virtual filling coorinated training system based on polyhybird real world devices |
CN109976519A (en) * | 2019-03-14 | 2019-07-05 | 浙江工业大学 | A kind of interactive display unit and its interactive display method based on augmented reality |
CN110109541A (en) * | 2019-04-25 | 2019-08-09 | 广州智伴人工智能科技有限公司 | A kind of method of multi-modal interaction |
CN110286762A (en) * | 2019-06-21 | 2019-09-27 | 济南大学 | A kind of Virtual Experiment Platform Based with multi-modal information processing function |
CN110288016A (en) * | 2019-06-21 | 2019-09-27 | 济南大学 | The multi-modal intention fusion method of one kind and application |
CN110471531A (en) * | 2019-08-14 | 2019-11-19 | 上海乂学教育科技有限公司 | Multi-modal interactive system and method in virtual reality |
CN111104470A (en) * | 2019-11-19 | 2020-05-05 | 青岛海信网络科技股份有限公司 | Method and system for linkage of electronic sand table and emergency platform |
CN111124116A (en) * | 2019-12-18 | 2020-05-08 | 佛山科学技术学院 | Method and system for interacting with remote object in virtual reality |
CN111124236A (en) * | 2018-10-30 | 2020-05-08 | 阿里巴巴集团控股有限公司 | Data processing method, device and machine readable medium |
CN111167115A (en) * | 2018-11-09 | 2020-05-19 | 致伸科技股份有限公司 | Interactive game system |
CN111367407A (en) * | 2020-02-24 | 2020-07-03 | Oppo(重庆)智能科技有限公司 | Intelligent glasses interaction method, intelligent glasses interaction device and intelligent glasses |
CN111724485A (en) * | 2020-06-11 | 2020-09-29 | 浙江商汤科技开发有限公司 | Method, device, electronic equipment and storage medium for realizing virtual-real fusion |
CN111832656A (en) * | 2020-07-17 | 2020-10-27 | 复旦大学 | Medical human-computer interaction assistance system and computer-readable storage medium containing the same |
CN112000219A (en) * | 2020-03-30 | 2020-11-27 | 华南理工大学 | Movable gesture interaction device and method for augmented reality game |
CN112069834A (en) * | 2020-09-02 | 2020-12-11 | 中国航空无线电电子研究所 | Integration method of multi-channel control instruction |
CN112148120A (en) * | 2020-08-18 | 2020-12-29 | 华为技术有限公司 | Method, equipment and storage medium for displaying virtual interface |
CN112486322A (en) * | 2020-12-07 | 2021-03-12 | 济南浪潮高新科技投资发展有限公司 | Multimodal AR (augmented reality) glasses interaction system based on voice recognition and gesture recognition |
CN112527112A (en) * | 2020-12-08 | 2021-03-19 | 中国空气动力研究与发展中心计算空气动力研究所 | Multi-channel immersive flow field visualization man-machine interaction method |
CN113476835A (en) * | 2020-10-22 | 2021-10-08 | 青岛海信电子产业控股股份有限公司 | Picture display method and device |
CN113589929A (en) * | 2021-07-29 | 2021-11-02 | 和舆图(北京)科技有限公司 | Spatial distance measuring method and system based on HoloLens equipment |
US11176910B2 (en) | 2018-08-22 | 2021-11-16 | Google Llc | Smartphone providing radar-based proxemic context |
CN113703583A (en) * | 2021-09-08 | 2021-11-26 | 厦门元馨智能科技有限公司 | Multi-mode cross fusion virtual image fusion system, method and device |
US11204694B2 (en) | 2018-08-24 | 2021-12-21 | Google Llc | Radar system facilitating ease and accuracy of user interactions with a user interface |
CN114167994A (en) * | 2022-02-11 | 2022-03-11 | 北京亮亮视野科技有限公司 | Knowledge base adding method, device, equipment and medium |
US11314312B2 (en) | 2018-10-22 | 2022-04-26 | Google Llc | Smartphone-based radar system for determining user intention in a lower-power mode |
CN114454814A (en) * | 2022-01-26 | 2022-05-10 | 深圳时空科技集团有限公司 | Human-computer interaction method and device for augmented reality |
CN114842146A (en) * | 2022-05-10 | 2022-08-02 | 中国民用航空飞行学院 | Civil aviation engine maintenance manual and work card modeling method and storable medium |
US11435468B2 (en) | 2018-08-22 | 2022-09-06 | Google Llc | Radar-based gesture enhancement for voice interfaces |
WO2022183775A1 (en) * | 2021-03-05 | 2022-09-09 | 华中师范大学 | Method for fusing multiple locomotion mechanisms in hybrid reinforcement teaching scene |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102339129A (en) * | 2011-09-19 | 2012-02-01 | 北京航空航天大学 | Multichannel human-computer interaction method based on voice and gestures |
CN104615243A (en) * | 2015-01-15 | 2015-05-13 | 深圳市掌网立体时代视讯技术有限公司 | Head-wearable type multi-channel interaction system and multi-channel interaction method |
CN106997236A (en) * | 2016-01-25 | 2017-08-01 | 亮风台(上海)信息科技有限公司 | Based on the multi-modal method and apparatus for inputting and interacting |
CN107194972A (en) * | 2017-05-16 | 2017-09-22 | 成都通甲优博科技有限责任公司 | A kind of camera marking method and system |
-
2018
- 2018-02-12 CN CN201810144421.XA patent/CN108334199A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102339129A (en) * | 2011-09-19 | 2012-02-01 | 北京航空航天大学 | Multichannel human-computer interaction method based on voice and gestures |
CN104615243A (en) * | 2015-01-15 | 2015-05-13 | 深圳市掌网立体时代视讯技术有限公司 | Head-wearable type multi-channel interaction system and multi-channel interaction method |
CN106997236A (en) * | 2016-01-25 | 2017-08-01 | 亮风台(上海)信息科技有限公司 | Based on the multi-modal method and apparatus for inputting and interacting |
CN107194972A (en) * | 2017-05-16 | 2017-09-22 | 成都通甲优博科技有限责任公司 | A kind of camera marking method and system |
Non-Patent Citations (1)
Title |
---|
张翼: "《基于图像特征的增强现实装配技术研究》", 《中国优秀硕士学位论文全文数据库信息科技辑》 * |
Cited By (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11435468B2 (en) | 2018-08-22 | 2022-09-06 | Google Llc | Radar-based gesture enhancement for voice interfaces |
US11176910B2 (en) | 2018-08-22 | 2021-11-16 | Google Llc | Smartphone providing radar-based proxemic context |
US11204694B2 (en) | 2018-08-24 | 2021-12-21 | Google Llc | Radar system facilitating ease and accuracy of user interactions with a user interface |
CN109343703A (en) * | 2018-09-10 | 2019-02-15 | 中国科学院计算机网络信息中心 | Information processing method, device, system, storage medium and processor |
CN109395375A (en) * | 2018-09-18 | 2019-03-01 | 华南理工大学 | A kind of 3d gaming method of interface interacted based on augmented reality and movement |
CN109214006A (en) * | 2018-09-18 | 2019-01-15 | 中国科学技术大学 | The natural language inference method that the hierarchical semantic of image enhancement indicates |
CN109461351A (en) * | 2018-09-28 | 2019-03-12 | 中国科学院苏州生物医学工程技术研究所 | The augmented reality game training system of three screens interaction |
CN109461351B (en) * | 2018-09-28 | 2021-04-02 | 中国科学院苏州生物医学工程技术研究所 | Three-screen interactive augmented reality game training system |
US11314312B2 (en) | 2018-10-22 | 2022-04-26 | Google Llc | Smartphone-based radar system for determining user intention in a lower-power mode |
CN111124236B (en) * | 2018-10-30 | 2023-04-28 | 斑马智行网络(香港)有限公司 | Data processing method, device and machine-readable medium |
CN111124236A (en) * | 2018-10-30 | 2020-05-08 | 阿里巴巴集团控股有限公司 | Data processing method, device and machine readable medium |
CN111167115A (en) * | 2018-11-09 | 2020-05-19 | 致伸科技股份有限公司 | Interactive game system |
CN109782907A (en) * | 2018-12-28 | 2019-05-21 | 西安交通大学 | A kind of virtual filling coorinated training system based on polyhybird real world devices |
CN109766003A (en) * | 2018-12-29 | 2019-05-17 | 北京诺亦腾科技有限公司 | Object methods of exhibiting and device in a kind of VR scene |
CN109976519A (en) * | 2019-03-14 | 2019-07-05 | 浙江工业大学 | A kind of interactive display unit and its interactive display method based on augmented reality |
CN109976519B (en) * | 2019-03-14 | 2022-05-03 | 浙江工业大学 | Interactive display device based on augmented reality and interactive display method thereof |
CN110109541B (en) * | 2019-04-25 | 2022-04-05 | 广州智伴人工智能科技有限公司 | Multi-modal interaction method |
CN110109541A (en) * | 2019-04-25 | 2019-08-09 | 广州智伴人工智能科技有限公司 | A kind of method of multi-modal interaction |
CN110286762B (en) * | 2019-06-21 | 2022-11-04 | 济南大学 | Virtual experiment platform with multi-mode information processing function |
CN110288016A (en) * | 2019-06-21 | 2019-09-27 | 济南大学 | The multi-modal intention fusion method of one kind and application |
CN110286762A (en) * | 2019-06-21 | 2019-09-27 | 济南大学 | A kind of Virtual Experiment Platform Based with multi-modal information processing function |
CN110288016B (en) * | 2019-06-21 | 2021-09-28 | 济南大学 | Multi-modal intention fusion method and application |
CN110471531A (en) * | 2019-08-14 | 2019-11-19 | 上海乂学教育科技有限公司 | Multi-modal interactive system and method in virtual reality |
CN111104470A (en) * | 2019-11-19 | 2020-05-05 | 青岛海信网络科技股份有限公司 | Method and system for linkage of electronic sand table and emergency platform |
CN111124116A (en) * | 2019-12-18 | 2020-05-08 | 佛山科学技术学院 | Method and system for interacting with remote object in virtual reality |
CN111367407B (en) * | 2020-02-24 | 2023-10-10 | Oppo(重庆)智能科技有限公司 | Intelligent glasses interaction method, intelligent glasses interaction device and intelligent glasses |
CN111367407A (en) * | 2020-02-24 | 2020-07-03 | Oppo(重庆)智能科技有限公司 | Intelligent glasses interaction method, intelligent glasses interaction device and intelligent glasses |
CN112000219A (en) * | 2020-03-30 | 2020-11-27 | 华南理工大学 | Movable gesture interaction device and method for augmented reality game |
CN112000219B (en) * | 2020-03-30 | 2022-06-14 | 华南理工大学 | Movable gesture interaction method for augmented reality game |
CN111724485A (en) * | 2020-06-11 | 2020-09-29 | 浙江商汤科技开发有限公司 | Method, device, electronic equipment and storage medium for realizing virtual-real fusion |
CN111832656A (en) * | 2020-07-17 | 2020-10-27 | 复旦大学 | Medical human-computer interaction assistance system and computer-readable storage medium containing the same |
CN112148120A (en) * | 2020-08-18 | 2020-12-29 | 华为技术有限公司 | Method, equipment and storage medium for displaying virtual interface |
CN112069834A (en) * | 2020-09-02 | 2020-12-11 | 中国航空无线电电子研究所 | Integration method of multi-channel control instruction |
CN113476835A (en) * | 2020-10-22 | 2021-10-08 | 青岛海信电子产业控股股份有限公司 | Picture display method and device |
CN112486322A (en) * | 2020-12-07 | 2021-03-12 | 济南浪潮高新科技投资发展有限公司 | Multimodal AR (augmented reality) glasses interaction system based on voice recognition and gesture recognition |
CN112527112A (en) * | 2020-12-08 | 2021-03-19 | 中国空气动力研究与发展中心计算空气动力研究所 | Multi-channel immersive flow field visualization man-machine interaction method |
WO2022183775A1 (en) * | 2021-03-05 | 2022-09-09 | 华中师范大学 | Method for fusing multiple locomotion mechanisms in hybrid reinforcement teaching scene |
CN113589929A (en) * | 2021-07-29 | 2021-11-02 | 和舆图(北京)科技有限公司 | Spatial distance measuring method and system based on HoloLens equipment |
CN113703583A (en) * | 2021-09-08 | 2021-11-26 | 厦门元馨智能科技有限公司 | Multi-mode cross fusion virtual image fusion system, method and device |
CN114454814A (en) * | 2022-01-26 | 2022-05-10 | 深圳时空科技集团有限公司 | Human-computer interaction method and device for augmented reality |
CN114454814B (en) * | 2022-01-26 | 2023-08-11 | 深圳时空数字科技有限公司 | Human-computer interaction method and device for augmented reality |
CN114167994B (en) * | 2022-02-11 | 2022-06-28 | 北京亮亮视野科技有限公司 | Knowledge base adding method, device, equipment and medium |
CN114167994A (en) * | 2022-02-11 | 2022-03-11 | 北京亮亮视野科技有限公司 | Knowledge base adding method, device, equipment and medium |
CN114842146A (en) * | 2022-05-10 | 2022-08-02 | 中国民用航空飞行学院 | Civil aviation engine maintenance manual and work card modeling method and storable medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108334199A (en) | The multi-modal exchange method of movable type based on augmented reality and device | |
CN105075246B (en) | The method that Tele-immersion formula is experienced is provided using mirror metaphor | |
CN107491174A (en) | Method, apparatus, system and electronic equipment for remote assistance | |
Zhou et al. | Trends in augmented reality tracking, interaction and display: A review of ten years of ISMAR | |
Dai | Virtual reality for industrial applications | |
CN106157359B (en) | Design method of virtual scene experience system | |
Wang | Augmented reality in architecture and design: potentials and challenges for application | |
Azuma | Overview of augmented reality | |
US20140176607A1 (en) | Simulation system for mixed reality content | |
MacIntyre et al. | Future multimedia user interfaces | |
US20150301596A1 (en) | Method, System, and Computer for Identifying Object in Augmented Reality | |
CN107566793A (en) | Method, apparatus, system and electronic equipment for remote assistance | |
CN106648098B (en) | AR projection method and system for user-defined scene | |
US11727238B2 (en) | Augmented camera for improved spatial localization and spatial orientation determination | |
CN103516808A (en) | Virtual/reality interaction platform provided with mobile terminal for intelligent exhibition hall | |
Kim et al. | AR interfacing with prototype 3D applications based on user-centered interactivity | |
Jo et al. | Chili: viewpoint control and on-video drawing for mobile video calls | |
Zhang et al. | The Application of Folk Art with Virtual Reality Technology in Visual Communication. | |
CN105975259A (en) | Implementation method and device of 3D (Three-dimensional) space user interface | |
CN113989462A (en) | Railway signal indoor equipment maintenance system based on augmented reality | |
Siegl et al. | An augmented reality human–computer interface for object localization in a cognitive vision system | |
CN106909222A (en) | VR display methods and device | |
Zhang et al. | Virtual Museum Scene Design Based on VRAR Realistic Interaction under PMC Artificial Intelligence Model | |
RE | Low cost augmented reality for industrial problems | |
US20240169568A1 (en) | Method, device, and computer program product for room layout |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180727 |
|
RJ01 | Rejection of invention patent application after publication |