CN110211240A - A kind of augmented reality method for exempting from sign-on ID - Google Patents

A kind of augmented reality method for exempting from sign-on ID Download PDF

Info

Publication number
CN110211240A
CN110211240A CN201910467466.5A CN201910467466A CN110211240A CN 110211240 A CN110211240 A CN 110211240A CN 201910467466 A CN201910467466 A CN 201910467466A CN 110211240 A CN110211240 A CN 110211240A
Authority
CN
China
Prior art keywords
information
data
model
augmented reality
client
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910467466.5A
Other languages
Chinese (zh)
Other versions
CN110211240B (en
Inventor
张元�
张乐
王智豪
焦世超
田杰
马珩钧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
North University of China
Original Assignee
North University of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by North University of China filed Critical North University of China
Priority to CN201910467466.5A priority Critical patent/CN110211240B/en
Publication of CN110211240A publication Critical patent/CN110211240A/en
Application granted granted Critical
Publication of CN110211240B publication Critical patent/CN110211240B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • Software Systems (AREA)
  • Processing Or Creating Images (AREA)
  • Information Transfer Between Computers (AREA)

Abstract

The invention belongs to augmented reality fields, disclose a kind of augmented reality method for exempting from sign-on ID, this method is with C/S (client-server end) framework, information transmission is carried out using udp protocol, client provides the dynamic loading function of human-computer interaction, information collection and dummy model, the information received is carried out identification classification by the convolutional neural networks of transfer learning training by server-side, dummy model is provided, to realize the effect of augmented reality.Human-computer interaction includes the present invention using by dummy object model upload service end and by client dynamically load, and while the memory needed for reducing client application, the load and interaction of a variety of models can also be realized in the case where not updating client;Solve the problems, such as that traditional augmented reality is demanding to new scene bad adaptability, and exploitation high to mark dependence.This method is applicable in the augmented reality application of a large amount of dummy models, especially in engineering model field.

Description

A kind of augmented reality method for exempting from sign-on ID
Technical field
The invention belongs to augmented reality fields, and in particular to a kind of augmented reality method for exempting from sign-on ID.
Background technique
Augmented reality, it is a kind of by " seamless " the integrated new technology of real world information and virtual world information, It is entity information (visual information, sound, the taste that script is difficult to experience in the certain time spatial dimension of real world Road, tactile etc.), it by science and technology such as computers, is superimposed again after analog simulation, by virtual Information application to real world, quilt Human sensory is perceived, to reach the sensory experience of exceeding reality.True environment and virtual object are added in real time Same picture or space exist simultaneously.
The specific implementation of augmented reality method and is placed to virtual information by carrying out three-dimensional registration in real world It on three-dimensional site, is finally shown by display equipment, developing AR by this method, (Augmented Reality, enhancing are existing It is real) application item, it needs to be placed in identification information and virtual information in advance, causes to identify specific mark in use Information, and new virtual information can not be loaded, it can not be interacted with new virtual information.In addition, AR (Augmented Reality, augmented reality) objective factors such as angle, distance and ambient of equipment can all influence the tracking of identification information The effect of identification and model load.
Summary of the invention
It is needed for above-mentioned tradition AR (Augmented Reality, augmented reality) technology identification information and virtual information It to be placed in advance, the identification of identification information, tracking can be by the interference of extraneous factor, thus the problem of influencing user experience, herein Propose a kind of augmented reality method for exempting from sign-on ID.This method is suitable for the virtual information of dynamically load target information, uses In augmented reality application.
In order to solve the above technical problems, the technical solution adopted by the present invention are as follows:
A kind of augmented reality method for exempting from sign-on ID, the augmented reality method are using C/S (Client-Server) The framework at client-server end carries out information transmission using udp protocol;Client provides human-computer interaction, information collection and void The dynamic loading function of analog model, server-side know the information received by the convolutional neural networks of transfer learning training Do not classify, dummy model is provided, to realize the effect of augmented reality.
By dummy object model upload service end and by client dynamically load in the present invention, client application institute is being reduced While needing memory, the load and interaction of a variety of models can also be realized in the case where not updating client;
The convolutional neural networks identification nicety of grading of transfer learning training is high, speed is fast, and higher generalization ability makes Identification object is also no longer limited to specific identifier;
In terms of human-computer interaction, interacted in a manner of by conveniently gesture, voice and staring with model, Yong Hucao Make comfortable natural.
Further, the human-computer interaction, information collection of the client and the dynamic loading function of dummy model include Information collection function stares identification function, gesture identification function, speech identifying function, space reflection function and dynamically load function Energy.
Still further, the information collection function of the client, stare identification function, gesture identification function, voice know Other function, space reflection function and dynamic loading function are realized using following steps:
The information collection function of client described in step C1.;It is taken pictures using holographic glasses HoloLens to target, it will Shooting result is saved with the format of Jpg;The picture that will take pictures is converted to Sprite format and is loaded onto provided by Uity3D engine It is shown on UI carrier;
Client described in step C2. stares identification function;It is described stare identification be based on eye movement tracer technique, for Track and selected holographic object, according to user's head position and direction by Unity3D engine Physical Raycast object Ricoh's line obtains the feedback of collision result, the letter of position and colliding object including the point of impingement after colliding with holographic object Breath, to realize the tracking of holographic object in scene and select;Client application can be realized by staring identification function The selected and movement of dummy object.
The gesture identification function of client described in step C3.;The gesture identification is to identify and track the position of user hand With capture input gesture while state, system automatic trigger is fed back accordingly, to manipulate the virtual objects in scene;
The speech identifying function of client described in step C4.;The speech recognition is user with the method for voice and holography Object interaction, speech recognition are realized and keyword and corresponding feedback behavior is arranged in client application, When user says keyword, client application responds preset feedback behavior;
The space reflection functional development of client described in step C5.;The space reflection is by virtual world and real generation Boundary is superimposed, is realized by following methods;
Step C5.1 is by using the holographic glasses HoloLens depth camera being equipped with and environment sensing camera, scanning Environmental data and built-in triangulation around user obtain real world to realize the modelling and digitlization of real world Digital physical space information;
Step C5.2 calculates whether digital physical space obtained above can place virtual hologram object in real time;By visitor The space reflection function at family end, the spatial position of dummy model no longer by identification information in real world position constraint;It adopts Method with space reflection can get rid of the physical location limitation of mark, keep virtual information more quasi- for identifying tracking Really, it is reasonably combined with real world.
The dynamic loading function of client models described in step C6.;The dynamically load of the model is using holographic glasses HoloLens loads the method for dummy model by access server to realize.
Further, the input gesture in the step 3 includes Air-tap, Navigation gesture and Bloom Three kinds.
Further, in the step C6 server store dummy model be by Unity3D engine, in advance will be empty The compressed package that analog model and script are packaged as AssetBundle is uploaded to server;Holographic glasses HoloLens is according to service It is after the identification of end to download the AssetBundle compressed package of corresponding model as a result, accessing server and decompress, thus implementation model Dynamically load.The type of object is thousands of in daily life, is difficult to shift to an earlier date on all objects in client application Merging, and holographic glasses are compared with high performance computer, rendering capability, memory and performance are all extremely limited, so wearing Formula augmented reality glasses can not carry a large amount of model for load, therefore present invention employs the dynamically loads from server The method of model.
Further, the server-side carries out the information received by the convolutional neural networks of transfer learning training Identification classification, provides dummy model;The convolutional neural networks of the transfer learning training, are realized especially by following steps:
Step S1. establishes sample data set;Good sample data set is the basis of information Classification and Identification, passes through internet Channel obtains sample image, and sample image is rotated by 90 ° by quantitative proportion, rotates 180 °, horizontal mirror image and vertical mirror image Exptended sample data set is operated, after expanding, is finally made as the data set identified for information;
Step S2. carries out model training on the data set for information identification that step S1 finally makes, from inhomogeneity Each sample image for selecting 70%-80% at random is training dataset in not, and remaining sample image is as test data set, repeatedly Generation number is 40-100 times, carries out model training;
Step S3. is trained by the accuracy rate that penalty values, over-fitting ratio and test data are classified come scoring model Effect;Wherein, shown in the accuracy rate such as formula (1) of test data classification
In formula (1), ExactQuantity indicates the correct quantity of test data classification results, and TotalQuantity is indicated The total quantity of test data;The accuracy rate of test data classification is higher, indicates that the effect of network model classification is better;
Obtained by cross entropy loss function of the penalty values by Softmax, as shown in formula (2)
Wherein,1{yi=j } refer to indicative function, its value is 1 when value is true within " { } ", otherwise for 0;Penalty values indicate that the training result of network model is better closer to 0;
Shown in over-fitting ratio such as formula (3)
In formula, TrainAcc indicates training data accuracy rate, and TrainAcc is as shown in Equation 4
In formula, TrainExactQuantity is the correct quantity of training data classification results, TrainTotalQuantity is the total quantity of training data.Over-fitting ratio indicates the extensive energy of network model closer to 1 Power is better.
Further, the model training process of the step S2 are as follows:
Step S2.1AlexNet network model carries out pre-training on the data set of ImageNet, will by the step The parameter initialization of AlexNet network model;
Step S2.2 is since last three layers of AlexNet network model are configured as 1000 classification, so by last three The full articulamentum of layer carries out re -training to adapt to new classification, and the parameter of new full articulamentum, Lai Shiying are retained by the step The class categories for the data set that step S1 is established;
Step S2.3 joins first five layer of convolutional layer and its corresponding pond layer, the activation primitive and model in step S2.1 Several full articulamentums and its parameter in step S2.2 combine, and are finely adjusted to complete the training of model.
Further, the information transmission is by the information of transmitting terminal by processing, is carried out using udp protocol to information Transmission, the information received is handled and is restored by receiving end;The maximum number of byte that can be transmitted according to its single first will pass Defeated information is pre-processed, transmitting terminal according to file absolute path go obtain picture, then to picture carry out data encoding, Data are cut and the operation of addition header information, add file type, file data length, number of data packets in header information And data number, receiving end according to data header information complete data decoding and recombination, and verify whether null data, if having Packet loss by verification information back and applies for that transmitting terminal numbers retransmission according to header data to packet loss information.
Further, the information processing of the transmitting terminal includes the following steps:
Step F1. encodes the information of transmission, and according to its information type, type content is encoded, and will compile In the file type of the result insertion header of code;
Step F2. counts the result length after the information coding of transmission, and the resultant content of statistics is compiled Code, will be in the file size of the result insertion header of coding;
Result equalization after the information coding of transmission is divided into multiple groups by step F3., and by the content of total number packets amount into Row coding, will be in the number of data packets of the result insertion header of coding;
Data group after segmentation is numbered step F4. in order, and number content is encoded, the result of coding It is inserted into the data number of header;
Step F5. repeats step F1 to step F4, pre-processes and retransmits in order to the information after segmentation, it is ensured that information is not It a large amount of simultaneously can send.
Further, the information processing of the receiving end includes the following steps:
Step R1. decodes the datagram header data received, by file type, IP and source port number classify, together When create new thread and go to receive new information;
Step R2. is created according to header file length and is received container, while receiving multiple file sizes, and representative has multiple same Type file;
Step R3. inserts data content according to data number in header and block length (file size/total number packets amount) Enter the corresponding position in container, so that it is guaranteed that file content sequence;
Step R4. verifies received content, if there is free information in container, represents data packetloss, then passes through step The IP and source port number that rapid R1 is recorded are fed back, and the volume where file type, file size and container sky information is passed through Number position, application transmitting terminal retransmit information corresponding to the number;
Step R5. repeats step R1 to step R4, and the information in container is corresponded to decode to lay equal stress on by file type and is write, is restored File.
Compared with prior art, the invention has the following advantages:
1. the present invention use a kind of augmented reality method for exempting from sign-on ID, by dummy object model upload service end and by Client dynamically load can also be real in the case where not updating client while the memory needed for reducing client application The load and interaction of existing a variety of models;Solve traditional augmented reality to new scene bad adaptability, to mark rely on it is high with And the demanding problem of exploitation.
2. the convolutional neural networks identification nicety of grading that training is completed is high, speed is fast, good generalization ability to know Other object is also no longer limited to specific identifier.
3. sample data set quantitative requirement is being greatly lowered using the method training convolutional neural networks of transfer learning Meanwhile the training time is also reduced, the more rapid adaptation application scenarios of network model are facilitated and classified service is provided.
4. substituting the mark tracking in traditional augmented reality method using the method for space reflection, the reality of mark can be got rid of The limitation of border position combines virtual information more accurately, reasonably with real world.
5. being interacted in a manner of by conveniently gesture, voice and staring with model, user in terms of human-computer interaction Operation is comfortable natural.
Detailed description of the invention
Fig. 1 is system architecture diagram of the invention;
Fig. 2 is client framework figure;
Fig. 3 is udp protocol header information figure;
Fig. 4 is server architecture figure;
Fig. 5 is 1 sample data set of embodiment;
Fig. 6 is 1 sample extending method of embodiment;
Fig. 7 is 1 model training result of example;
Fig. 8 is to implement 1 pair of 99 formula tank to take pictures;
Fig. 9 is the information transmission of embodiment 1;
Figure 10 is the information identification classification of embodiment 1;
Figure 11 is the dynamically load of the virtual information of embodiment 1.
Specific embodiment
Further technical solution of the present invention is clearly and completely described in the following with reference to the drawings and specific embodiments, Obviously, described embodiment is only a part of the embodiments of the present invention, instead of all the embodiments.Based in the present invention Embodiment, every other embodiment obtained by those of ordinary skill in the art without making creative efforts, It shall fall within the protection scope of the present invention.
Embodiment 1:
As shown in Figure 1, one of the present embodiment exempts from the augmented reality method of sign-on ID, the augmented reality method is Using the framework at client-server end, information transmission is carried out using udp protocol;Client provide human-computer interaction, information collection with And the dynamic loading function of dummy model, server-side by the information received by transfer learning training convolutional neural networks into Row identification classification, provides dummy model, to realize the effect of augmented reality.Human-computer interaction, the information collection of the client And the dynamic loading function of dummy model includes information collection function, stares identification function, gesture identification function, speech recognition Function, space reflection function and dynamic loading function.
The function realization for client, server-side and information transmission is described in detail separately below.
Client:
The architecture diagram of client is as shown in Fig. 2, the function of client is implemented by the following steps:
The information collection function of client described in step C1.;It is taken pictures using holographic glasses HoloLens to target, it will Shooting result is saved with the format of Jpg;The picture that will take pictures is converted to Sprite format and is loaded onto provided by Uity3D engine It is shown on UI carrier;Directly find out on UI taking pictures as a result, the photo bad to effect, takes pictures again to facilitate.
Client described in step C2. stares identification function;It is described stare identification be based on eye movement tracer technique, for Track and selected holographic object, according to user's head position and direction by Unity3D engine Physical Raycast object Ricoh's line obtains the feedback of collision result, the letter of position and colliding object including the point of impingement after colliding with holographic object Breath, to realize the tracking of holographic object in scene and select;The application program of client stares identification function and shows virtual object The selected and movement of body;
The gesture identification function of client described in step C3.;The gesture identification is to identify and track the position of user hand With capture input gesture while state, system automatic trigger is fed back accordingly, to manipulate the virtual objects in scene;Input Gesture includes Air-tap, Navigation gesture and tri- kinds of Bloom.
The speech identifying function of client described in step C4.;The speech recognition is user with the method for voice and holography Object interaction, speech recognition are realized and keyword and corresponding feedback behavior is arranged in client application, When user says keyword, client application responds preset feedback behavior;In the present embodiment, speech recognition and gesture The concrete operations instruction of identification and respondent behavior are as shown in table 1.
1 speech recognition of table and the concrete operations instruction of gesture identification and respondent behavior
The space reflection functional development of client described in step C5.;The space reflection is by virtual world and real generation Boundary is superimposed, is realized by following methods;
Step C5.1 is by using the holographic glasses HoloLens depth camera being equipped with and environment sensing camera, scanning Environmental data and built-in triangulation around user obtain real world to realize the modelling and digitlization of real world Digital physical space information;
Step C5.2 calculates whether digital physical space obtained above can place virtual hologram object in real time;By visitor The space reflection function at family end, the spatial position of dummy model no longer by identification information in real world position constraint;
The dynamic loading function of client models described in step C6.;The dynamically load of the model is using holographic glasses HoloLens loads the method for dummy model by access server to realize.The dynamically load of model is HoloLens wear-type What augmented reality glasses were realized by access server load dummy model.The type of object is thousands of in daily life, It is difficult for all objects to be placed in advance in augmented reality application program, and HoloLens is compared with high performance computer, Its rendering capability, memory and performance are all extremely limited, so HoloLens can not carry a large amount of model for load.For The problem is present invention employs the method for the dynamically load model from server, by Unity3D engine, by dummy model and Script is packaged as the compressed package of AssetBundle, is uploaded to server, and HoloLens identified according to server-side after as a result, Access server is downloaded the AssetBundle compressed package of corresponding model and is decompressed, thus the dynamically load of implementation model.
Information transmission:
The information transmission is to transmit the information of transmitting terminal, receiving end to information using udp protocol by processing The information received is handled and restored;The information of transmission is carried out according to the maximum number of byte that its single can transmit first Pretreatment, transmitting terminal according to file absolute path go obtain picture, then to picture carry out data encoding, data cutting and The operation of header information is added, file type, file data length, number of data packets and data is added in header information and compiles Number, receiving end according to data header information complete data decoding and recombination, and verify whether null data, if there is packet loss, high-ranking officers It tests information back and applies for that transmitting terminal numbers packet loss information according to header data and retransmit.Udp protocol header information such as Fig. 3 It is shown.
The information processing of the transmitting terminal includes the following steps:
Step F1. encodes the information of transmission, and according to its information type, type content is encoded, and will compile In the file type of the result insertion header of code;
Step F2. counts the result length after the information coding of transmission, and the resultant content of statistics is compiled Code, will be in the file size of the result insertion header of coding;
Result equalization after the information coding of transmission is divided into multiple groups by step F3., and by the content of total number packets amount into Row coding, will be in the number of data packets of the result insertion header of coding;
Data group after segmentation is numbered step F4. in order, and number content is encoded, the result of coding It is inserted into the data number of header;
Step F5. repeats step F1 to step F4, pre-processes and retransmits in order to the information after segmentation, it is ensured that information is not It a large amount of simultaneously can send.
The information processing of the receiving end includes the following steps:
Step R1. decodes the datagram header data received, by file type, IP and source port number classify, together When create new thread and go to receive new information;
Step R2. is created according to header file length and is received container, while receiving multiple file sizes, and representative has multiple same Type file;
Step R3. inserts data content according to data number in header and block length (file size/total number packets amount) Enter the corresponding position in container, so that it is guaranteed that file content sequence;
Step R4. verifies received content, if there is free information in container, represents data packetloss, then passes through step The IP and source port number that rapid R1 is recorded are fed back, and the volume where file type, file size and container sky information is passed through Number position, application transmitting terminal retransmit information corresponding to the number;
Step R5. repeats step R1 to step R4, and the information in container is corresponded to decode to lay equal stress on by file type and is write, is restored File.
Server-side:
The architecture diagram of server-side is as shown in Figure 4.The information received is passed through the volume of transfer learning training by the server-side Product neural network carries out identification classification, provides dummy model;The present invention includes tank, panzer according to existing armoring class model And the identifications such as fighting vehicle classification, and server is built using Apache, existing dummy model is uploaded on server.Institute The convolutional neural networks for stating transfer learning training, are realized especially by following steps:
Step S1. establishes sample data set;Good sample data set is the basis of information Classification and Identification, and the present invention passes through Internet approach has collected 15 class tanks and panzer image pattern, after arrangement and mark, obtains amounting to 1444 images Sample, specimen types and distributed number are as shown in Figure 5.In order to over-fitting occur when mitigating and avoid model training, reduce The influence of the recognition effect brought by Different categories of samples data bulk is unevenly distributed, the sample obtained by internet channels Image is rotated by 90 ° by quantitative proportion, rotate 180 °, horizontal mirror image and vertical mirror image operation exptended sample data set, such as Fig. 6 (a)-(e) shown in.After expanding, data set total quantity reaches 9012, is finally made as tank armor identification Data set;
Step S2. carries out model training on the data set for information identification that step S1 finally makes, from inhomogeneity Each 75% sample image of selecting at random is training dataset in not, and remaining sample image is as test data set, iteration time Number is 48 times, carries out model training;Training process are as follows:
Step S2.1 AlexNet network model carries out pre-training on the data set of ImageNet, will by the step The parameter initialization of AlexNet network model;
Step S2.2 is since last three layers of AlexNet network model are configured as 1000 classification, so by last three The full articulamentum of layer carries out re -training to adapt to new classification, and the parameter of new full articulamentum, Lai Shiying are retained by the step The class categories of the data set for the tank armor identification that step S1 is established;
Step S2.3 joins first five layer of convolutional layer and its corresponding pond layer, the activation primitive and model in step S2.1 Several full articulamentums and its parameter in step S2.2 combine, and are finely adjusted to complete the training of model.
Step S3. is trained by the accuracy rate that penalty values, over-fitting ratio and test data are classified come scoring model Effect;Wherein, shown in the accuracy rate such as formula (1) of test data classification
In formula (1), ExactQuantity indicates the correct quantity of test data classification results, and TotalQuantity is indicated The total quantity of test data;The accuracy rate of test data classification is higher, indicates that the effect of network model classification is better.
Obtained by cross entropy loss function of the penalty values by Softmax, as shown in formula (2)
Wherein,1{yi=j } refer to indicative function, its value is 1 when value is true within " { } ", otherwise for 0;Penalty values indicate that the training result of network model is better closer to 0.
Shown in over-fitting ratio such as formula (3)
In formula, TrainAcc indicates training data accuracy rate, and TrainAcc is as shown in Equation 4
In formula, TrainExactQuantity is the correct quantity of training data classification results, TrainTotalQuantity is the total quantity of training data.Over-fitting ratio indicates the extensive energy of network model closer to 1 Power is better.Network model is as shown in Figure 7 by transfer learning training training result.The accuracy rate average value of final test is 97.51%, and the over-fitting ratio of model is basically stable at 1.03 or so, and it is good to show that the network model of present invention training has Good generalization ability.
Practical operation is carried out using system, HoloLens is connected into Wifi, with server-side according to IP address and port numbers It realizes information transmission, starts test macro functions.HoloLens takes pictures result as shown in figure 8, Fig. 9 is service client information Reception.As can be seen that picture packet loss phenomenon solves substantially by carrying out information checking in application layer from Fig. 8 and Fig. 9, Service client information classification results are as shown in Figure 10, and Figure 11 is the dynamically load figure of virtual information.It can be seen from fig. 10 that this hair The convolutional neural networks identification nicety of grading of bright middle transfer learning training is high, speed is fast, and higher generalization ability to identify Object is also no longer limited to specific identifier, and 99 formula tank models are loaded out by Figure 11, and mobile by the method for space reflection On to shelf (dotted line is tripod in Fig. 9).It can be seen from fig. 11 that using the method for space reflection for identifying tracking, The physical location limitation of mark can be got rid of.C/S framework in through the invention, by dummy object model upload service end and by Client dynamically load can also be in the case where not updating client while the memory needed for reducing client application Realize the load and interaction of a variety of models.
Embodiment 2:
The difference of embodiment 2 and embodiment 1 is only that step S2.
Embodiment 2 is that each 70% sample image selected at random is training dataset, remaining sample from different classes of For image as test data set, the number of iterations is 40 times, carries out model training;As a result consistent with embodiment 1.
Embodiment 3:
The difference of embodiment 3 and embodiment 1 is only that step S2.
Embodiment 3 is that each 80% sample image selected at random is training dataset, remaining sample from different classes of For image as test data set, the number of iterations is 100 times, carries out model training;As a result consistent with embodiment 1.
This method is applicable in the augmented reality application of a large amount of dummy models, especially in engineering model field.
Although the present invention is described in detail referring to the foregoing embodiments, for those skilled in the art, It is still possible to modify the technical solutions described in the foregoing embodiments, or part of technical characteristic is carried out etc. With replacement, all within the spirits and principles of the present invention, any modification, equivalent replacement, improvement and so on should be included in this Within the protection scope of invention.

Claims (10)

1. a kind of augmented reality method for exempting from sign-on ID, it is characterised in that: the augmented reality method is using client-clothes The framework at business end carries out information transmission using udp protocol;Client provides human-computer interaction, information collection and dummy model The information received is carried out identification classification by the convolutional neural networks of transfer learning training by dynamic loading function, server-side, Dummy model is provided, to realize the effect of augmented reality.
2. a kind of augmented reality method for exempting from sign-on ID according to claim 1, it is characterised in that: the client Human-computer interaction, information collection and dummy model dynamic loading function include information collection function, stare identification function, hand Gesture identification function, speech identifying function, space reflection function and dynamic loading function.
3. a kind of augmented reality method for exempting from sign-on ID according to claim 2, it is characterised in that: the client Information collection function, stare identification function, gesture identification function, speech identifying function, space reflection function and dynamically load Function is realized using following steps:
The information collection function of client described in step C1.;It is taken pictures, will be shot to target using holographic glasses HoloLens As a result it is saved with the format of Jpg;The picture that will take pictures is converted to Sprite format and is loaded onto the load of UI provided by Uity3D engine It is shown on body;
Client described in step C2. stares identification function;It is described stare identification be based on eye movement tracer technique, for tracking and Selected holographic object, according to user's head position and direction by Unity3D engine Physical Raycast physical light Line, after being collided with holographic object, obtain collision result feedback, the information of position and colliding object including the point of impingement, from And it realizes the tracking of holographic object in scene and selectes;
The gesture identification function of client described in step C3.;The gesture identification is position and the shape for identifying and tracking user hand Capture input gesture while state, system automatic trigger is fed back accordingly, to manipulate the virtual objects in scene;
The speech identifying function of client described in step C4.;The speech recognition is user with the method and holographic object of voice Interaction, speech recognition be by keyword being arranged in client application and corresponding feedback behavior and is realized, when with When keyword is said at family, client application responds preset feedback behavior;
The space reflection functional development of client described in step C5.;The space reflection is by virtual world and real world phase Superposition, is realized by following methods;
Step C5.1 scans user by using the holographic glasses HoloLens depth camera being equipped with and environment sensing camera Around environmental data and built-in triangulation to realize the modelling and digitlization of real world obtain the number of real world Word physical space information;
Step C5.2 calculates whether digital physical space obtained above can place virtual hologram object in real time;By client Space reflection function, the spatial position of dummy model no longer by identification information in real world position constraint;
The dynamic loading function of client models described in step C6.;The dynamically load of the model is using holographic glasses HoloLens loads the method for dummy model by access server to realize.
4. a kind of augmented reality method for exempting from sign-on ID according to claim 3, it is characterised in that: the step C3 In input gesture include Air-tap, Navigation gesture and tri- kinds of Bloom.
5. a kind of augmented reality method for exempting from sign-on ID according to claim 3, it is characterised in that: the step C6 The dummy model of middle server storage is in advance to be packaged as dummy model and script by Unity3D engine The compressed package of AssetBundle is uploaded to server;Holographic glasses HoloLens identified according to server-side after as a result, access clothes Business device is downloaded the AssetBundle compressed package of corresponding model and is decompressed, thus the dynamically load of implementation model.
6. a kind of augmented reality method for exempting from sign-on ID according to claim 1, it is characterised in that: the server-side The information received is subjected to identification classification by the convolutional neural networks of transfer learning training, dummy model is provided;It is described to move The convolutional neural networks for moving learning training, are realized especially by following steps:
Step S1. establishes sample data set;Good sample data set is the basis of information Classification and Identification, passes through internet channels Sample image is obtained, sample image is rotated by 90 ° by quantitative proportion, rotates 180 °, horizontal mirror image and vertical mirror image operation Exptended sample data set is finally made as the data set identified for information after expanding;
Step S2. carries out model training on the data set for information identification that step S1 finally makes, from different classes of Each sample image for selecting 70%-80% at random is training dataset, and remaining sample image is as test data set, iteration time Number is 40-100 times, carries out model training;
The effect that step S3. is trained by the accuracy rate that penalty values, over-fitting ratio and test data are classified come scoring model; Wherein, shown in the accuracy rate such as formula (1) of test data classification
In formula (1), ExactQuantity indicates the correct quantity of test data classification results, and TotalQuantity indicates test The total quantity of data;The accuracy rate of test data classification is higher, indicates that the effect of network model classification is better;
Obtained by cross entropy loss function of the penalty values by Softmax, as shown in formula (2)
Wherein,1{yi=j } refer to indicative function, otherwise it is 0 that when value is true in " { } ", its value, which is 1,;Damage Mistake value indicates that the training result of network model is better closer to 0;
Shown in over-fitting ratio such as formula (3)
In formula, TrainAcc indicates training data accuracy rate, and TrainAcc is as shown in Equation 4
In formula, TrainExactQuantity is the correct quantity of training data classification results, and TrainTotalQuantity is The total quantity of training data;Over-fitting ratio indicates that the generalization ability of network model is better closer to 1.
7. a kind of augmented reality method for exempting from sign-on ID according to claim 6, it is characterised in that: the step S2 Model training process are as follows:
Step S2.1 AlexNet network model carries out pre-training on the data set of ImageNet, will by the step The parameter initialization of AlexNet network model;
Step S2.2 since last three layers of AlexNet network model are configured as 1000 classification, so will last three layers entirely Articulamentum carries out re -training to adapt to new classification, and the parameter of new full articulamentum, Lai Shiying step are retained by the step The class categories for the data set that S1 is established;
Step S2.3 by step S2.1 first five layer of convolutional layer and its corresponding pond layer, activation primitive and model parameter with Full articulamentum and its parameter in step S2.2 combine, and are finely adjusted to complete the training of model.
8. a kind of augmented reality method for exempting from sign-on ID according to claim 1, it is characterised in that: the information passes Defeated is to transmit the information of transmitting terminal to information using udp protocol by processing, and receiving end carries out the information received Processing and reduction;The maximum number of byte that can be transmitted according to its single first pre-processes the information of transmission, in transmitting terminal root It goes to obtain picture according to file absolute path, data encoding then is carried out to picture, data are cut and the behaviour of addition header information Make, file type, file data length, number of data packets and data number is added in header information, receiving end is according to data Header information completes data decoding and recombination, and verifies whether null data by verification information back and is applied sending out if there is packet loss Sending end is numbered packet loss information according to header data and is retransmitted.
9. a kind of augmented reality method for exempting from sign-on ID according to claim 8, it is characterised in that: the transmitting terminal Information processing include the following steps:
Step F1. encodes the information of transmission, and according to its information type, type content is encoded, and by coding As a result it is inserted into the file type of header;
Step F2. counts the result length after the information coding of transmission, and the resultant content of statistics is encoded, will In the file size of the result insertion header of coding;
Result equalization after the information coding of transmission is divided into multiple groups by step F3., and the content of total number packets amount is compiled Code, will be in the number of data packets of the result insertion header of coding;
Data group after segmentation is numbered step F4. in order, and number content is encoded, the result insertion of coding In the data number of header;
Step F5. repeats step F1 to step F4, pre-processes and retransmits in order to the information after segmentation, it is ensured that information will not be same Shi great Liang is sent.
10. a kind of augmented reality method for exempting from sign-on ID according to claim 8, it is characterised in that: the reception The information processing at end includes the following steps:
Step R1. decodes the datagram header data received, by file type, IP and source port number classify, create simultaneously New thread is built to go to receive new information;
Step R2. is created according to header file length and is received container, while receiving multiple file sizes, and representative has multiple same types File;
Data content is inserted into and holds according to data number in header and block length (file size/total number packets amount) by step R3. Corresponding position in device, so that it is guaranteed that file content sequence;
Step R4. verifies received content, if there is free information in container, represents data packetloss, then passes through step R1 The IP and source port number recorded is fed back, and the number position where file type, file size and container sky information is passed through It sets, application transmitting terminal retransmits information corresponding to the number;
Step R5. repeats step R1 to step R4, and the information in container is corresponded to decode to lay equal stress on by file type and is write, original text is gone back Part.
CN201910467466.5A 2019-05-31 2019-05-31 Registration-free identification augmented reality method Active CN110211240B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910467466.5A CN110211240B (en) 2019-05-31 2019-05-31 Registration-free identification augmented reality method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910467466.5A CN110211240B (en) 2019-05-31 2019-05-31 Registration-free identification augmented reality method

Publications (2)

Publication Number Publication Date
CN110211240A true CN110211240A (en) 2019-09-06
CN110211240B CN110211240B (en) 2022-10-21

Family

ID=67789867

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910467466.5A Active CN110211240B (en) 2019-05-31 2019-05-31 Registration-free identification augmented reality method

Country Status (1)

Country Link
CN (1) CN110211240B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110782525A (en) * 2019-11-08 2020-02-11 腾讯科技(深圳)有限公司 Method, apparatus and medium for identifying virtual object in virtual environment
CN112486322A (en) * 2020-12-07 2021-03-12 济南浪潮高新科技投资发展有限公司 Multimodal AR (augmented reality) glasses interaction system based on voice recognition and gesture recognition
CN114205429A (en) * 2021-12-14 2022-03-18 深圳壹账通智能科技有限公司 Voice packet processing method, system, equipment and storage medium based on UDP protocol
CN117689508A (en) * 2023-12-19 2024-03-12 杭州露电数字科技集团有限公司 Intelligent teaching aid method and system based on MR equipment

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102739872A (en) * 2012-07-13 2012-10-17 苏州梦想人软件科技有限公司 Mobile terminal, and augmented reality method used for mobile terminal
CN106373198A (en) * 2016-09-18 2017-02-01 福州大学 Method for realizing augmented reality
CN106846237A (en) * 2017-02-28 2017-06-13 山西辰涵影视文化传媒有限公司 A kind of enhancing implementation method based on Unity3D
CN107015642A (en) * 2017-03-13 2017-08-04 武汉秀宝软件有限公司 A kind of method of data synchronization and system based on augmented reality
CN107331220A (en) * 2017-09-01 2017-11-07 国网辽宁省电力有限公司锦州供电公司 Transformer O&M simulation training system and method based on augmented reality
CN107451661A (en) * 2017-06-29 2017-12-08 西安电子科技大学 A kind of neutral net transfer learning method based on virtual image data collection
US20180012411A1 (en) * 2016-07-11 2018-01-11 Gravity Jack, Inc. Augmented Reality Methods and Devices
CN108416846A (en) * 2018-03-16 2018-08-17 北京邮电大学 It is a kind of without the three-dimensional registration algorithm of mark

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102739872A (en) * 2012-07-13 2012-10-17 苏州梦想人软件科技有限公司 Mobile terminal, and augmented reality method used for mobile terminal
US20180012411A1 (en) * 2016-07-11 2018-01-11 Gravity Jack, Inc. Augmented Reality Methods and Devices
CN106373198A (en) * 2016-09-18 2017-02-01 福州大学 Method for realizing augmented reality
CN106846237A (en) * 2017-02-28 2017-06-13 山西辰涵影视文化传媒有限公司 A kind of enhancing implementation method based on Unity3D
CN107015642A (en) * 2017-03-13 2017-08-04 武汉秀宝软件有限公司 A kind of method of data synchronization and system based on augmented reality
CN107451661A (en) * 2017-06-29 2017-12-08 西安电子科技大学 A kind of neutral net transfer learning method based on virtual image data collection
CN107331220A (en) * 2017-09-01 2017-11-07 国网辽宁省电力有限公司锦州供电公司 Transformer O&M simulation training system and method based on augmented reality
CN108416846A (en) * 2018-03-16 2018-08-17 北京邮电大学 It is a kind of without the three-dimensional registration algorithm of mark

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
宋亦凡: "基于移动端的无标识增强现实系统研究与开发", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *
张乐 等: "一种免注册标识的增强现实方法", 《科学技术与工程》 *
张乐: "基于HoloLens的增强现实识别系统", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *
戚纯: "基于混合现实技术的数字博物馆应用研究", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110782525A (en) * 2019-11-08 2020-02-11 腾讯科技(深圳)有限公司 Method, apparatus and medium for identifying virtual object in virtual environment
CN112486322A (en) * 2020-12-07 2021-03-12 济南浪潮高新科技投资发展有限公司 Multimodal AR (augmented reality) glasses interaction system based on voice recognition and gesture recognition
CN114205429A (en) * 2021-12-14 2022-03-18 深圳壹账通智能科技有限公司 Voice packet processing method, system, equipment and storage medium based on UDP protocol
CN117689508A (en) * 2023-12-19 2024-03-12 杭州露电数字科技集团有限公司 Intelligent teaching aid method and system based on MR equipment

Also Published As

Publication number Publication date
CN110211240B (en) 2022-10-21

Similar Documents

Publication Publication Date Title
CN110211240A (en) A kind of augmented reality method for exempting from sign-on ID
US10109101B2 (en) Re-utilization of render assets for video compression
CN110679152B (en) Method and system for generating fused reality scene
KR102318772B1 (en) Domain Separation Neural Networks
US10699471B2 (en) Methods and systems for rendering frames based on a virtual entity description frame of a virtual scene
Rapprich et al. Application of modern technologies in popularization of the Czech volcanic geoheritage
CN108764453B (en) Modeling method and action prediction system for multi-agent synchronous game
IL305340A (en) High resolution neural rendering
CN108236784A (en) The training method and device of model, storage medium, electronic device
CN113269858A (en) Virtual scene rendering method and device, computer equipment and storage medium
US11232623B2 (en) Method and system for creating a neural net based lossy renderer
CN117649478B (en) Model training method, image processing method and electronic equipment
KR102342738B1 (en) Method and apparatus for generating mega size augmented reality image information
US11544894B2 (en) Latency-resilient cloud rendering
CN117809084A (en) Image recognition model training method, interaction method and device based on image recognition
CN108520532A (en) Identify the method and device of movement direction of object in video
Meloni et al. Messing up 3d virtual environments: Transferable adversarial 3d objects
CN116561578A (en) Sample generation method, device, equipment and storage medium
CN110458164A (en) Image processing method, device, equipment and computer readable storage medium
CN116983660A (en) Image processing method and related device
CN116030040B (en) Data processing method, device, equipment and medium
CN115984094B (en) Face safety generation method and equipment based on multi-loss constraint visual angle consistency
CN117036652B (en) Layout information generation method, model training method, device and electronic equipment
US20230419655A1 (en) Controlled image manipulations using semantic labels
KR102667003B1 (en) Method and system for training image generation model using content information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant