CN110110647A - The method, apparatus and storage medium that information is shown are carried out based on AR equipment - Google Patents
The method, apparatus and storage medium that information is shown are carried out based on AR equipment Download PDFInfo
- Publication number
- CN110110647A CN110110647A CN201910363552.1A CN201910363552A CN110110647A CN 110110647 A CN110110647 A CN 110110647A CN 201910363552 A CN201910363552 A CN 201910363552A CN 110110647 A CN110110647 A CN 110110647A
- Authority
- CN
- China
- Prior art keywords
- information
- equipment
- required movement
- spatial position
- limbs
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Computer Graphics (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Processing Or Creating Images (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
The method that information is shown is carried out based on AR equipment the disclosure is directed to a kind of, belongs to intelligent terminal technical field.The method is executed by the augmented reality AR equipment comprising image collection assembly, which comprises obtains the first location information that instruction refers to spatial position;When this with reference to spatial position be in AR equipment within sweep of the eye when, pass through image collection assembly acquisition AR equipment limbs image within the vision;According to collected limbs image, the real space position of the corresponding limbs of required movement that are including in limbs image is determined;It obtains with reference to the matching relationship between spatial position and real space position;Prompt information is shown according to matching relationship.The disclosure can prompt whether required movement is executed correctly by the prompt information, allow user when individually taking exercise using the AR helmet, the requirement point of required movement can more accurately be grasped, make itself made athletic performance more standard, and then obtains more accurate exercise guidance.
Description
Technical field
This disclosure relates to intelligent terminal technical field, in particular to a kind of method shown based on AR equipment progress information,
Device and storage medium.
Background technique
In life of today, due to the impact of body-building upsurge, more and more people are added in the troop of body-building.With
The development of intelligent terminal technical field, many fitness enthusiasts can in conjunction with the field technology come learn body-building movement.
For the crowd individually taken exercise, in the related art, usually pass through the technologies such as multimedia and multi-functional aided education
Body-building is played in mobile phone to learn relevant body-building movement for example, user passes through the application program of mobile phone-downloaded relative motion
Video recording or the body-building of virtual portrait movement oneself are imitated corresponding body-building movement and are practiced, to reach the mesh of motion exercise
's.
Summary of the invention
The embodiment of the present disclosure, which provides, a kind of carries out the method, apparatus that shows of information and storage medium based on AR equipment.Institute
It is as follows to state technical solution:
According to the first aspect of the embodiments of the present disclosure, a kind of method shown based on AR equipment progress information, institute are provided
Method is stated to be executed by the augmented reality AR equipment comprising image collection assembly, which comprises
Obtain the first location information being used to indicate with reference to spatial position;The reference spatial position is required movement by just
Really execute when, the corresponding limbs of the required movement spatial position;
When it is described with reference to spatial position be in the AR equipment within sweep of the eye when, adopted by described image acquisition component
Collect the limbs image within the vision of the AR equipment;
According to the collected limbs image, determine that including in the limbs image, the required movement is corresponding
The real space position of limbs;
The matching relationship with reference between spatial position and the real space position is obtained, the matching relationship includes
Matching mismatches;
Prompt information is shown according to the matching relationship, and the prompt information is for prompting the required movement whether by just
Really execute.
Optionally, the acquisition is used to indicate the first location information with reference to spatial position, comprising:
Inquire it is pre-stored, when the required movement is executed correctly, the corresponding limbs of the required movement with refer to
Determine the coordinate information in the space coordinates that indicia patterns position is reference;
The coordinate information inquired is retrieved as the first location information.
Optionally, before acquisition is used to indicate with reference to the first location information of spatial position, further includes:
The modeling information of the required movement is obtained, the modeling information includes the action message of the required movement, with
And location information of the corresponding dummy model of the required movement in the space coordinates;
The dummy model is created according to the action message of required movement;
According to location information of the dummy model in the space coordinates, obtains in the dummy model, is described
Coordinate information of the corresponding limbs of required movement in the space coordinates;
It stores in the dummy model, coordinate letter of the corresponding limbs of the required movement in the space coordinates
Breath.
Optionally, in the location information according to the dummy model in the space coordinates, the virtual mould is obtained
In type, the corresponding limbs of the required movement are before the coordinate information in the space coordinates, further includes:
Shape parameter is obtained, the shape parameter is used to indicate the figure of the wearer of the AR equipment;
The dummy model is zoomed in and out according to the shape parameter;
The location information according to the dummy model in the space coordinates, obtain in the dummy model,
Coordinate information of the corresponding limbs of the required movement in the space coordinates, comprising:
According to location information of the dummy model in the space coordinates after scaling, the dummy model is obtained
In, coordinate information of the corresponding limbs of the required movement in the space coordinates.
Optionally, it is described when it is described with reference to spatial position be in the AR equipment within sweep of the eye when, pass through the figure
Before the limbs image within the vision for acquiring the AR equipment as acquisition component, further includes:
It obtains second location information and Viewing-angle information, the second location information is used to indicate the AR equipment in the sky
Between spatial position in coordinate system, the Viewing-angle information is used to indicate the view directions of the AR equipment;
According to the second location information and the Viewing-angle information, the field range for obtaining the AR equipment is corresponding described
Space coordinate boundary in space coordinates;
When the corresponding space coordinate of the first location information is in the space coordinate border inner, the ginseng is determined
It examines spatial position and is in the AR equipment within sweep of the eye.
Optionally, the acquisition second location information and Viewing-angle information, comprising:
According to the AR equipment in previous moment to the moving parameter between current time, and, the AR equipment is preceding
The historical position information at one moment obtains the second location information;The historical position information is opened from initial position message
Begin, successively the moving parameter according to the AR equipment between each adjacent moment determines;
According to the AR equipment in previous moment to the attitudes vibration parameter between current time, and, the AR equipment
The Viewing-angle information described in the historical angel acquisition of information of previous moment;The historical angel information is opened from initial Viewing-angle information
Begin, successively the attitudes vibration parameter according to the AR equipment between each adjacent moment determines.
Optionally, before the acquisition second location information and Viewing-angle information, further includes:
It include the tag image of the specified indicia patterns by the acquisition of described image acquisition component;
According to the collected tag image, the AR equipment and institute at the time of acquiring the tag image are determined
State the relative positional relationship between specified indicia patterns;
According to opposite position at the time of acquiring the tag image, between the AR equipment and the specified indicia patterns
Set initial position message described in Relation acquisition and the initial Viewing-angle information.
Optionally, the specified indicia patterns are arranged on Yoga mat surface.
It is optionally, described that prompt information is shown according to the matching relationship, comprising:
It is corresponding at the reference spatial position when the reference spatial position and the real space position mismatch
Show the first prompt information, first prompt information is for prompting the required movement not to be executed correctly;
It is corresponding to be shown at the reference spatial position when reference spatial position and the real space location matches
Show the second prompt information, second prompt information is for prompting the required movement to be executed correctly.
Optionally, first prompt information and second prompt information are the patterns that shape is identical and color is different.
According to the second aspect of an embodiment of the present disclosure, a kind of device shown based on AR equipment progress information, institute are provided
Device is stated to contain in the augmented reality AR equipment of image collection assembly, described device includes:
First obtains module, for obtaining the first location information being used to indicate with reference to spatial position;It is described to refer to space
When position is that required movement is executed correctly, the spatial position of the corresponding limbs of the required movement;
Image capture module, for when it is described with reference to spatial position be in the AR equipment within sweep of the eye when, pass through
Described image acquisition component acquires the limbs image within the vision of the AR equipment;
Position determination module, for determining include in the limbs image, institute according to the collected limbs image
State the real space position of the corresponding limbs of required movement;
Second obtains module, closes for obtaining the matching with reference between spatial position and the real space position
System, the matching relationship include matching or mismatch;
Information display module, for showing prompt information according to the matching relationship, the prompt information is for prompting institute
State whether required movement is executed correctly.
Optionally, described first module is obtained, is specifically used for,
Inquire it is pre-stored, when the required movement is executed correctly, the corresponding limbs of the required movement with refer to
Determine the coordinate information in the space coordinates that indicia patterns position is reference;
The coordinate information inquired is retrieved as the first location information.
Optionally, described device further include:
Third obtains module, for obtaining the first position being used to indicate with reference to spatial position in the first acquisition module
Before information, the modeling information of the required movement is obtained, the modeling information includes the action message of the required movement, with
And location information of the corresponding dummy model of the required movement in the space coordinates;
Creation module, for creating the dummy model according to the action message of required movement;
4th obtains module, for the location information according to the dummy model in the space coordinates, obtains institute
It states in dummy model, coordinate information of the corresponding limbs of the required movement in the space coordinates;
Memory module, for storing in the dummy model, the corresponding limbs of the required movement are in the space coordinate
Coordinate information in system.
Optionally, described device further include:
5th obtains module, is used in the 4th acquisition module according to the dummy model in the space coordinates
Location information, obtain in the dummy model, coordinate of the corresponding limbs of the required movement in the space coordinates
Before information, shape parameter is obtained, the shape parameter is used to indicate the figure of the wearer of the AR equipment;
Zoom module, in the location information according to the dummy model in the space coordinates, described in acquisition
In dummy model, the corresponding limbs of the required movement before the coordinate information in the space coordinates, according to the body
Shape parameter zooms in and out the dummy model;
It is described 4th obtain module, specifically for according to the dummy model after scaling in the space coordinates
Location information obtains in the dummy model, coordinate letter of the corresponding limbs of the required movement in the space coordinates
Breath.
Optionally, described device further include:
6th obtains module, for being in the AR equipment with reference to spatial position when described in described image acquisition module
When within sweep of the eye, before the limbs image within the vision that the AR equipment is acquired by described image acquisition component, obtain
Second location information and Viewing-angle information are taken, the second location information is used to indicate the AR equipment in the space coordinates
Spatial position, the Viewing-angle information is used to indicate the view directions of the AR equipment;
7th obtains module, for obtaining the AR equipment according to the second location information and the Viewing-angle information
The corresponding space coordinate boundary in the space coordinates of field range;
First determining module, for being in the space coordinate boundary when the corresponding space coordinate of the first location information
When internal, determine and described be in the AR equipment within sweep of the eye with reference to spatial position.
Optionally, the described 6th module is obtained, is specifically used for,
According to the AR equipment in previous moment to the moving parameter between current time, and, the AR equipment is preceding
The historical position information at one moment obtains the second location information;The historical position information is opened from initial position message
Begin, successively the moving parameter according to the AR equipment between each adjacent moment determines;
According to the AR equipment in previous moment to the attitudes vibration parameter between current time, and, the AR equipment
The Viewing-angle information described in the historical angel acquisition of information of previous moment;The historical angel information is opened from initial Viewing-angle information
Begin, successively the attitudes vibration parameter according to the AR equipment between each adjacent moment determines.
Optionally, described device further include:
Acquisition module, for passing through institute before the described 6th obtains module acquisition second location information and Viewing-angle information
State the tag image that image collection assembly acquisition includes the specified indicia patterns;
Second determining module, for according to the collected tag image, determine acquire the tag image when
It carves, the relative positional relationship between the AR equipment and the specified indicia patterns;
8th obtains module, for basis at the time of acquiring the tag image, the AR equipment and the specified mark
Remember that the relative positional relationship between pattern obtains the initial position message and the initial Viewing-angle information.
Optionally, the specified indicia patterns are arranged on Yoga mat surface.
Optionally, the information display module, is specifically used for,
It is corresponding at the reference spatial position when the reference spatial position and the real space position mismatch
Show the first prompt information, first prompt information is for prompting the required movement not to be executed correctly;
It is corresponding to be shown at the reference spatial position when reference spatial position and the real space location matches
Show the second prompt information, second prompt information is for prompting the required movement to be executed correctly.
Optionally, first prompt information and second prompt information are the patterns that shape is identical and color is different.
According to the third aspect of an embodiment of the present disclosure, a kind of device shown based on AR equipment progress information, institute are provided
Device is stated to contain in the augmented reality AR equipment of image collection assembly, described device includes:
Processor;
For storing the memory of the executable instruction of the processor;
Wherein, the processor is configured to:
Obtain the first location information being used to indicate with reference to spatial position;The reference spatial position is required movement by just
Really execute when, the corresponding limbs of the required movement spatial position;
When it is described with reference to spatial position be in the AR equipment within sweep of the eye when, adopted by described image acquisition component
Collect the limbs image within the vision of the AR equipment;
According to the collected limbs image, determine that including in the limbs image, the required movement is corresponding
The real space position of limbs;
The matching relationship with reference between spatial position and the real space position is obtained, the matching relationship includes
Matching mismatches;
Prompt information is shown according to the matching relationship, and the prompt information is for prompting the required movement whether by just
Really execute.
According to a fourth aspect of embodiments of the present disclosure, a kind of computer readable storage medium is provided, the computer can
It reads to include executable instruction in storage medium, processor calls the executable instruction to realize above-mentioned first aspect and first
The method that information is shown is carried out based on AR equipment described in the arbitrarily optional implementation of aspect.
The technical solution that the embodiment of the present disclosure provides is include at least the following beneficial effects:
The first location information of spatial position is referred to by obtaining instruction;When this refers to the view that spatial position is in AR equipment
When in wild range, the limbs image within the vision of AR equipment is acquired by image collection assembly;According to collected limbs
Image determines the real space position of the corresponding limbs of required movement that are including in limbs image;Obtain with reference to spatial position and
Matching relationship between real space position;Prompt information is shown according to matching relationship.The disclosure can pass through the prompt information
Whether prompt required movement is executed correctly, and allows user when individually taking exercise using the AR helmet, can be more
The requirement point for accurately grasping required movement makes athletic performance more standard made by itself, and then obtains more accurately
Exercise guidance.
It should be understood that above general description and following detailed description be only it is exemplary and explanatory, not
The disclosure can be limited.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows the implementation for meeting the disclosure
Example, and together with specification for explaining the principles of this disclosure.
Fig. 1 is a kind of composed structure schematic diagram for camera shooting type AR helmet that the embodiment of the present disclosure provides;
Fig. 2 is a kind of method flow diagram shown based on AR equipment progress information that the embodiment of the present disclosure provides;
Fig. 3 is a kind of interface schematic diagram for space coordinates that the embodiment of the present disclosure provides;
Fig. 4 is a kind of method flow diagram shown based on AR equipment progress information that the embodiment of the present disclosure provides;
Fig. 5 is a kind of action schematic diagram for required movement that the embodiment of the present disclosure is related to;
Fig. 6 is a kind of interface schematic diagram for creation user shape parameter that the embodiment of the present disclosure is related to;
Fig. 7 is a kind of schematic diagram of a scenario for acquisition limbs image that the embodiment of the present disclosure is related to;
Fig. 8 is a kind of schematic diagram of a scenario for displaying prompt information that the embodiment of the present disclosure is related to;
Fig. 9 is a kind of schematic diagram of a scenario for displaying prompt information that the embodiment of the present disclosure is related to;
Figure 10 is a kind of schematic diagram of a scenario for displaying prompt information that the embodiment of the present disclosure is related to;
Figure 11 is a kind of illustrative block diagram that the device that information is shown is carried out based on AR equipment of the disclosure one;
Figure 12 is shown according to an exemplary embodiment a kind of for carrying out the device that information is shown based on AR equipment
Block diagram.
Specific embodiment
Example embodiments are described in detail here, and the example is illustrated in the accompanying drawings.Following description is related to
When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment
Described in embodiment do not represent all implementations consistent with this disclosure.On the contrary, they be only with it is such as appended
The example of the consistent device and method of some aspects be described in detail in claims, the disclosure.
The scheme that the disclosure provides can be used for people in daily life and individually carry out in the scene of fitness training, in order to just
In understanding, some nouns that the embodiment of the present disclosure is related to simply are introduced first below.
Augmented reality (Augmented Reality, AR): being position and the angle that one kind calculates camera image in real time
And add the technology of respective image.This technology can be by the modes such as projecting or directly displaying, in the display screen of eyeglass
Middle that image frame is superimposed upon in real world, operator can be interacted by equipment with the image frame of display.
AR technology can include but is not limited to following 3 kinds: projection AR, display type AR, camera shooting type AR according to projection pattern.
Wherein, be in the projection device of projection AR with transparent glasses lens, can be by the image projecting of virtual world to the transparent glasses lens
On;It is that can directly be shown virtual world on the display screen with transparent display screen in the projection device of display type AR
Show;Camera shooting type AR is also known as puppet AR, the image collection assembly of such equipment can be set by taking the scene of real world at such
By real world and virtual world Overlapping display on standby display screen.
Referring to FIG. 1, it illustrates a kind of composed structures of camera shooting type AR helmet of embodiment of the present disclosure offer to show
It is intended to.As shown in Figure 1, wherein containing video card 110, memory 120, processor 130, external equipment 140, sensor 150, figure
As acquisition component 160 and display screen 170.
Optionally, which can be scanned real-life scene by image collection assembly 160
Or shooting, the data of scanning or shooting is sent to processor 130, is shown on display screen 170 after the processing of processor 130
Show.Wherein, image collection assembly 160 can wrap containing camera, and optionally, which can be depth camera, specifically
Quantity can be provided by developer.Pre-stored image data in memory 120 can also be handled by processor 130
It is shown on display screen 170 afterwards, further, pre-stored image data can be with Overlapping display in image in memory 120
In the image of the scanning of acquisition component 160 or shooting, it is shown on display screen 170 jointly.
Optionally, the motion image information of various movements and wanting for various athletic performances be can store in memory 120
It asks a little, also can store the account information etc. of corresponding user.In addition, memory 120 or integrated being locally stored are set
Standby or expanding storage apparatus, such as: pluggable storage card etc., it is not limited in this embodiment of the present disclosure.
Optionally, external equipment 140 can include but is not limited to image recognition technology, Gesture Recognition, eye tracking
One of technology or multiple combinations.For example possesses the computer equipment of image recognition technology, possesses the intelligence of Gesture Recognition
Energy gloves, the eye tracker for possessing eye tracking technology etc..
Optionally, sensor 150 can include but is not limited to acceleration transducer, gyro sensor, gravity sensor,
One of temperature sensor, infrared sensor or multiple combinations.Sensor 150 can find out user institute by principle of induction
The movement done can finally be shown on display screen 170 by the movement by a series of processing of entire AR helmet, be in
Now arrive user at the moment.
In the related technology, most users are to carry out body-building forging by existing multimedia and multi-functional assisted teaching system
Refining.For example, user can be allowed to watch virtual portrait by the behavior and movement etc. of player plays virtual portrait
Behavior and movement etc., and then user can be instructed to learn these behaviors in player and movement etc..However, using this teaching
Mode goes to be learnt, and user can not accurately grasp the action criteria point of virtual portrait, can not learn the movement of oneself
It is whether correct, accurate, and then such tutoring system cannot be to the guidance that user is accurately acted.
To solve the problems, such as that above-mentioned the relevant technologies exist, carry out what information was shown based on AR equipment present disclose provides a kind of
Method can be applied in the AR equipment in embodiment shown in FIG. 1.Referring to FIG. 2, provided it illustrates the embodiment of the present disclosure
A kind of to carry out the method flow diagram that information is shown based on AR equipment, this method can be by the augmented reality comprising image collection assembly
AR equipment executes, as shown in Fig. 2, this method may include steps of:
In step 201, the first location information being used to indicate with reference to spatial position is obtained.
Wherein, when this is that required movement is executed correctly with reference to spatial position, the space of the corresponding limbs of the required movement
Position.
In step 202, when reference spatial position be in AR equipment within sweep of the eye when, adopted by image collection assembly
Collect the limbs image within the vision of AR equipment.
In step 203, according to collected limbs image, determine that including in the limbs image, required movement is corresponding
Limbs real space position.
In step 204, this is obtained with reference to the matching relationship between spatial position and the real space position.
Wherein, which includes matching or mismatches.
In step 205, prompt information is shown according to the matching relationship.
Wherein, the prompt information is for prompting whether required movement is executed correctly.
Optionally, above-mentioned acquisition is used to indicate the first location information with reference to spatial position, comprising:
Inquire it is pre-stored, when required movement is executed correctly, the corresponding limbs of required movement with specify label figure
Case position is the coordinate information in the space coordinates of reference;
The coordinate information inquired is retrieved as first location information.
Optionally, before above-mentioned acquisition is used to indicate with reference to the first location information of spatial position, this method further include:
The modeling information of required movement is obtained, modeling information includes the action message and required movement pair of required movement
Location information of the dummy model answered in space coordinates;
Dummy model is created according to the action message of required movement;
According to location information of the dummy model in space coordinates, obtain in dummy model, the corresponding limb of required movement
Coordinate information of the body in space coordinates;
It stores in dummy model, coordinate information of the corresponding limbs of required movement in space coordinates.
Optionally, it in the above-mentioned location information according to dummy model in space coordinates, obtains in dummy model, specify
Before acting coordinate information of the corresponding limbs in space coordinates, this method further include:
Shape parameter is obtained, which is used to indicate the figure of the wearer of AR equipment;
Dummy model is zoomed in and out according to shape parameter;
The above-mentioned location information according to dummy model in space coordinates obtains in dummy model, required movement corresponds to
Coordinate information of the limbs in space coordinates, comprising:
According to location information of the dummy model after scaling in space coordinates, obtain in dummy model, required movement
Coordinate information of the corresponding limbs in space coordinates.
Optionally, when reference spatial position be in AR equipment within sweep of the eye when, AR is acquired by image collection assembly
Before the limbs image within the vision of equipment, this method further include:
It obtains second location information and Viewing-angle information, the second location information is used to indicate AR equipment in space coordinates
Spatial position, which is used to indicate the view directions of AR equipment;
According to second location information and Viewing-angle information, the corresponding sky in space coordinates of field range of AR equipment is obtained
Between coordinate system;
When the corresponding space coordinate of first location information is in space coordinate border inner, determine with reference at spatial position
Within sweep of the eye in AR equipment.
Optionally, above-mentioned acquisition second location information and Viewing-angle information, comprising:
According to AR equipment in previous moment to the moving parameter between current time, and, AR equipment is in previous moment
Historical position information obtains second location information;The historical position information is successively set according to AR since initial position message
What the standby moving parameter between each adjacent moment determined;
According to AR equipment in previous moment to the attitudes vibration parameter between current time, and, AR equipment is when previous
The historical angel acquisition of information Viewing-angle information at quarter;The historical angel information is successively set according to AR since initial Viewing-angle information
What the standby attitudes vibration parameter between each adjacent moment determined.
Optionally, before above-mentioned acquisition second location information and Viewing-angle information, this method further include:
Tag image by image collection assembly acquisition comprising specified indicia patterns;
According to collected tag image, determine at the time of acquire tag image, AR equipment and specified indicia patterns it
Between relative positional relationship;
According at the time of acquiring tag image, the relative positional relationship between AR equipment and specified indicia patterns is obtained just
Beginning location information and initial Viewing-angle information.
Optionally, above-mentioned specified indicia patterns are arranged on Yoga mat surface.
It is optionally, above-mentioned that prompt information is shown according to matching relationship, comprising:
When reference spatial position and real space position mismatch, corresponding the first prompt of the display at reference spatial position
Information, the first prompt information is for prompting required movement not to be executed correctly;
When reference spatial position and real space location matches, corresponding the second prompt of the display at reference spatial position letter
Breath, the second prompt information is for prompting required movement to be executed correctly.
Optionally, first prompt information and the second prompt information are the patterns that shape is identical and color is different.
In conclusion the disclosure can prompt whether required movement is executed correctly by the prompt information, so that user
The requirement point that when individually taking exercise using the AR helmet, can more accurately can grasp required movement, makes itself
Made athletic performance more standard, and then obtain more accurate exercise guidance.
In a kind of mode in the cards, user can open the AR equipment, pass through the Image Acquisition group in AR equipment
Part identifies specified indicia patterns, at this point, the AR equipment can determine itself phase between the specified indicia patterns
To positional relationship, and record at the time of identify the specified indicia patterns accordingly;According to identify the specified indicia patterns when
Carve and the relative positional relationship between the AR equipment and the specified indicia patterns obtain the AR equipment initial position message and
Initial Viewing-angle information.Wherein, which is used to indicate the inceptive direction of the view directions of AR equipment.
In a kind of mode in the cards, which be can establish after identifying the specified indicia patterns with this
Specified indicia patterns position is the space coordinates of reference, referring to FIG. 3, it illustrates one that the embodiment of the present disclosure provides
The interface schematic diagram of kind space coordinates specifies indicia patterns 302 as shown in figure 3, wherein containing AR equipment 301, and space is sat
The reference axis 303 of system is marked, which specifies indicia patterns 302 to carry out identifying it this by the image collection assembly of itself
Afterwards, specified the distance between the indicia patterns and AR equipment are got, are sat to obtain the AR equipment from the space in foundation
Position coordinates in mark system, the position coordinates of the AR equipment in the space coordinates of foundation obtained at this time are set as the AR
Standby initial position message.Optionally, which can be the combination of single or multiple cameras, optionally,
The specified indicia patterns can be set in Yoga mat surface, desktop etc., can also be pasted by a kind of paster by user any
One position that AR equipment can be allowed to recognize.Optionally, which can be two dimensional code, picture, in bar code
One or several kinds of combinations.
In a kind of mode in the cards, as shown in figure 3, view directions 304 are wherein further comprised, area of visual field boundary
305, which, can also will be corresponding while determining itself relative positional relationship between the specified indicia patterns
View directions 304 are decided, that is, obtain the view directions of the AR equipment in the space coordinates.Further, which sets
It is standby the corresponding space of the field range of the AR equipment to be determined according to the initial Viewing-angle information and initial position message of acquisition
Coordinate system, i.e., area of visual field boundary 305 shown in Fig. 3.
When AR equipment is when the movement with wearer is moved, which can sensor by itself and place
Device etc. is managed, the location information and Viewing-angle information at each moment of the AR equipment in moving process are got, correspondingly, can also
To obtain corresponding area of visual field boundary of each moment.
In a kind of mode in the cards, above-mentioned initial position message is got in the AR equipment and initial visual angle is believed
After breath, with the use of the AR equipment, which can carry out information and show.Referring to FIG. 4, it illustrates disclosure realities
The a kind of of example offer is provided and carries out the method flow diagram that information is shown based on AR equipment, this method can be by including image collection assembly
Augmented reality AR equipment execute, as shown in figure 4, this method may include steps of:
In step 401, the first location information being used to indicate with reference to spatial position is obtained.
Wherein, when this is that required movement is executed correctly with reference to spatial position, the space bit of the corresponding limbs of required movement
It sets.
In a kind of mode in the cards, which can inquire pre-stored, and required movement is executed correctly
When, the corresponding limbs of required movement with specify indicia patterns position be refer to space coordinates in coordinate information,
And the coordinate information inquired is retrieved as first location information.
In a kind of mode in the cards, which can first obtain the modeling of required movement before step 401
Information, wherein the modeling information includes that the action message of required movement and the corresponding dummy model of required movement are sat in space
Location information in mark system.And the AR equipment can create the virtual mould according to the action message of the required movement got
Type, and obtained in the dummy model according to location information of the corresponding dummy model of required movement in space coordinates, should
Coordinate information of the corresponding limbs of required movement in space coordinates.
Optionally, which can be a set of yoga exercise movement, a set of military sports fist acts, a set of muscular training is dynamic
Make etc.;Optionally, user can choose required movement of the corresponding a set of movement as the AR equipment, the AR equipment
Get corresponding action message.Optionally, the action message of the required movement can be the operating point information of the dummy model,
The either corresponding identification information of the required movement, for example, the AR equipment stores identification information and required movement in memory
Corresponding table please refers to table 1, wherein containing identification information corresponding relationship corresponding with required movement.
Mark 1 | Movement 1 |
Mark 2 | Movement 2 |
…… | …… |
Table 1
After user selects corresponding a set of movement, the AR equipment is available to arrive corresponding identification information, which sets
It is standby that corresponding movement can be got by the identification information, and then obtain the corresponding action message of the dummy model.
In a kind of mode in the cards, position letter of the corresponding dummy model of the required movement in space coordinates
Breath can be location information of the reference center of the corresponding dummy model of the required movement in space coordinates.Referring to FIG. 5,
It illustrates a kind of action schematic diagrams for required movement that the embodiment of the present disclosure is related to, as shown in figure 5, wherein containing virtual mould
Type 501, operating point 502, reference center 503, location information of the corresponding dummy model of the required movement in space coordinates
It can be location information of the reference center 503 of the corresponding dummy model of the required movement in space coordinates.Further,
By taking Fig. 5 as an example, the modeling information which obtains the required movement may include the action message of the required movement --- and it is dynamic
Location information of the reference center 503 of work point 502 and dummy model in space coordinates, exists according to the reference center 503
Location information in space coordinates, so that it is determined that the location information of the operating point 502 in space coordinates, the operating point
502 location information in space coordinates is into the dummy model, and the corresponding limbs of required movement are in space coordinates
Interior coordinate information.It is above-mentioned to can be in the dummy model with reference to spatial position in a kind of mode in the cards, it specifies dynamic
Make the spatial position that coordinate information of the corresponding limbs in space coordinates includes within a preset range, for example, the AR equipment
The coordinate information got is (10,20,30), this can be centered on the coordinate information with reference to spatial position, prolongs around
Spatial position after stretching 5 unit lengths.Wherein, which can be existed in advance by application developer or operation maintenance personnel
It is arranged in program.This can have multiple with reference to spatial position and corresponding first location information, that is, required movement can be with
Have 1 or more reference spatial position and corresponding first location information.
Optionally, in the dummy model which can get this according to the model of creation, required movement pair
Coordinate information of the limbs answered in space coordinates is stored in the memory of the AR equipment, in order to the inquiry of AR equipment.
In a kind of mode in the cards, which can also store position from other equipment without the creation of the dummy model
It sets and directly acquires in the dummy model, coordinate information of the corresponding limbs of required movement in space coordinates.
In a kind of mode in the cards, which can also obtain shape parameter, which is used to indicate
The figure of the wearer of the AR equipment;And it is zoomed in and out according to dummy model of the shape parameter to above-mentioned acquisition;According to scaling
Location information of the dummy model afterwards in space coordinates obtains in the dummy model, the corresponding limbs of required movement are in sky
Between coordinate information in coordinate system.
For example, the AR equipment can be according to the shape parameter that user selects come the progress of the dummy model to the above-mentioned creation
Scaling, optionally, zooming in and out to the dummy model can be first to obtaining the reference center and operating point of the dummy model
Positional relationship between information zooms in and out, then obtains the reference center of the dummy model after the scaling and operating point information exists
Coordinate information in space coordinates;It is also possible to first obtain the reference center and operating point information in space coordinates
After coordinate information, the coordinate information of the reference center and operating point information in space coordinates is zoomed in and out.Finally will
The coordinate information of the reference center and operating point information in space coordinates after the scaling is retrieved as in the dummy model,
Coordinate information of the corresponding limbs of required movement in space coordinates.
Referring to FIG. 6, a kind of interface schematic diagram for the creation user shape parameter being related to it illustrates the embodiment of the present disclosure,
As shown in fig. 6, wherein contain height input frame 601, weight input frame 602, brachium input frame 603, waistline input frame 604,
Shoulder breadth input frame 605, hip circumference input frame 606, bust input frame 607 etc., user can be in the shape parameter input interfaces
Input corresponding data, and to establishing the corresponding shape parameter of a user accordingly, when using the AR equipment, user can be with
By selecting corresponding user, so that the AR equipment is available to arrive the corresponding shape parameter of the user, in order to above-mentioned to void
Analog model zooms in and out the execution of step.Optionally, the shape parameter of the user is also possible to the associated terminal hair of the AR equipment
It send, is communicated to connect for example, the AR equipment can be established with smart phone, by the body for obtaining the user that associated terminal is sent
Shape parameter zooms in and out above-mentioned dummy model etc..Optionally, it can be smart phone with the associated terminal of AR equipment, put down
Plate computer, E-book reader, smartwatch, Intelligent bracelet, pocket computer on knee etc..
In step 402, second location information and Viewing-angle information are obtained.
Wherein, which is used to indicate spatial position of the AR equipment in space coordinates, the Viewing-angle information
It is used to indicate the view directions of AR equipment.
In a kind of mode in the cards, according to AR equipment in previous moment to the moving parameter between current time,
And AR equipment obtains second location information in the historical position information of previous moment.Wherein, which is from first
Beginning location information starts, and successively the moving parameter according to AR equipment between each adjacent moment determines.Optionally, which joins
Number can be the acceleration parameter that the acceleration transducer of the AR equipment obtains, the AR equipment can according in previous moment to working as
The corresponding historical position information of acceleration parameter and previous moment between the preceding moment calculates second location information, and obtains
The second location information being calculated.For example, the AR equipment be in the corresponding historical position information of previous moment (10,10,
10) location information at the current time that, which is calculated by acceleration parameter is (11,12,10), which can
The location information at current time is retrieved as second location information.
According to AR equipment in previous moment to the attitudes vibration parameter between current time, and, AR equipment is when previous
The historical angel acquisition of information Viewing-angle information at quarter.Wherein, which is the successively basis since initial Viewing-angle information
Attitudes vibration parameter of the AR equipment between each adjacent moment determines.Optionally, which can be the AR and sets
The angle change parameter of standby view directions, which, which can be, is used to indicate the AR equipment in unit time interior angle
The pace of change and variable quantity of degree, the AR equipment can get the AR equipment by gyro sensor etc. and be carved into for the moment in the past
Angle change parameter between current time, according to the view directions meter of the angle change parameter and previous moment AR equipment
The current time AR equipment view directions are calculated, the view directions that this is calculated are retrieved as current time corresponding visual angle letter
Breath.Wherein, the previous moment and time interval between current time can be preparatory by application developer or operation maintenance personnel
It is arranged in AR equipment.
In step 403, according to second location information and Viewing-angle information, the field range for obtaining AR equipment is corresponding in space
Space coordinate boundary in coordinate system.
Correspondingly, the AR equipment has also determined that after determining second location information and Viewing-angle information by second
The space coordinate boundary of field range corresponding to confidence breath and Viewing-angle information, it is corresponding in sky which obtains the field range
Between space coordinate boundary in coordinate system.
In step 404, it when the corresponding space coordinate of first location information is in the space coordinate border inner, determines
This is in AR equipment within sweep of the eye with reference to spatial position.
Optionally, which can judge whether the corresponding space coordinate of first location information obtained is in this constantly
The corresponding space coordinate boundary in space coordinates of the field range of AR equipment, when the corresponding space coordinate of first location information
When in the space coordinate border inner, determine the first location information it is corresponding with reference to spatial position in the AR equipment
Within sweep of the eye.
In step 405, the limbs image within the vision of AR equipment is acquired by image collection assembly.
In a kind of mode in the cards, when the AR equipment determines that this refers to the visual field model that spatial position is in AR equipment
When enclosing interior, the limbs image within the vision of AR equipment can be acquired by the camera in image collection assembly.Optionally,
The camera of the acquisition limbs image and above-mentioned identification specify the camera of indicia patterns to can be the same camera.It please refers to
Fig. 7, it illustrates a kind of schematic diagram of a scenario for acquisition limbs image that the embodiment of the present disclosure is related to, as shown in fig. 7, wherein including
Limbs 701, with reference to spatial position 702, field range 703, when Viewing-angle information is changed to field range 703 by user, and
There is with reference to spatial position 702 determination in the field range 703, at this point, if the position of limbs 701 is moved to the visual field by user
When in range 703, which can acquire the limbs image of the limbs 701 within the vision by camera.
In a step 406, according to collected limbs image, determine that including in limbs image, required movement is corresponding
The real space position of limbs.
In a kind of mode in the cards, which can identify including in the limbs image of acquisition, specified dynamic
Make the real space position of corresponding limbs, that is, the real space position of the limbs.Optionally, which can be with
It is the positional relationship between the limbs and the AR equipment, which can pass through ranging by the camera of the AR equipment
Technology determines, for example, monocular cam ranging, binocular camera ranging and other more mesh camera rangings.Further really
Coordinate information of the limbs profile of the limbs in space coordinates is made, by the coordinate information of limbs profile in space coordinates
In include space be retrieved as the real space positions of the limbs.
In step 407, it obtains with reference to the matching relationship between spatial position and real space position.
Wherein, which includes matching or mismatches.Optionally, the matching or mismatch can serve to indicate that
Whether the first location information for referring to spatial position is in the real space position.
In a kind of mode in the cards, which carries out the reference spatial position of acquisition and real space position
When matching, it can be carried out by the location information corresponding with the real space position of the first location information with reference to spatial position
Matching, that is, the coordinate information of first position whether there is in the real space position, if the coordinate information of first position is present in
In the real space position, then obtaining with reference to the relationship between spatial position and real space position is matching relationship, if first
The coordinate information of position in the real space position, is not then obtained with reference to the relationship between spatial position and real space position
To mismatch relationship.
In a step 408, prompt information is shown according to matching relationship.
Wherein, the prompt information is for prompting whether required movement is executed correctly.
A kind of pass in mode in the cards, between the reference spatial position and real space position of above-mentioned acquisition
It is corresponding that the first prompt information, the first prompt information are shown at the first location information of reference spatial position when system is mismatches
For prompting required movement not to be executed correctly;Optionally, when the matching relationship is matching, the AR equipment is available to aobvious
Show the display signal of the first prompt information, which includes the type and corresponding first location information for showing image, is made
First prompt information can be shown at the corresponding first location information in reference spatial position by obtaining the AR equipment, to prompt user
Actions taken is incorrect.
It is corresponding in reference sky when the relationship between the reference spatial position of above-mentioned acquisition and real space position is matching
Between the second prompt information is shown at position, the second prompt information is for prompting required movement to be executed correctly.When the matching relationship
When to match, the available display signal to the second prompt information of display of the AR equipment, the display signal includes display image
Type and corresponding first location information, allow the AR equipment at the corresponding first location information in reference spatial position
It shows second prompt information, is incorrect to prompt user's actions taken.
Optionally, first prompt information and the second prompt information are at least a kind of shape style, for example, circle, number,
Chinese character, box, to hook, cross etc..Optionally, first prompt information and the second prompt information can be that shape is identical and face
The different pattern of color, alternatively, first prompt information and the second prompt information can be the identical pattern of shape different colours, or
Person, first prompt information and the second prompt information can be all different pattern of shape and color.Referring to FIG. 8, it shows
The schematic diagram of a scenario for a kind of displaying prompt information that the embodiment of the present disclosure is related to, as shown in figure 8, wherein contain the first prompt
Information 801, the second prompt information 802, the first prompt information 801 can be red circle, for prompting the finger at the position
Fixed movement is not executed correctly, and the second prompt information 802 can be the circle of green, for prompting the required movement at the position
It is executed correctly.
Optionally, Viewing-angle information of the displaying of first prompt information and the second prompt information also all with the AR equipment has
Close, for example, if position coordinates at the corresponding position of the first prompt information not the AR equipment within sweep of the eye, the AR equipment
Display screen in there is no first prompt information, correspondingly, if the AR equipment within sweep of the eye comprising the first prompt letter
When ceasing the reference spatial position at corresponding position, which be may be displayed in the display screen of the AR equipment.
In a kind of mode in the cards, if not collecting limbs image in above-mentioned steps 405, which can be with
Directly the first prompt information is shown at corresponding first location information, to prompt user to need the movement done being moved to this
At corresponding position, effect is instructed to user action to play.Referring to FIG. 9, it illustrates the embodiment of the present disclosure to be related to
A kind of displaying prompt information schematic diagram of a scenario, as shown in figure 9, wherein containing the first prompt information 901, field range
902, with reference to spatial position 903, which does not collect limbs image in the field range 902, which can be with
The first prompt information 901 is shown at the corresponding first position of reference spatial position 903.
In a kind of mode in the cards, which goes back when showing first prompt information and the second prompt information
The broadcasting of suggestion voice can be opened simultaneously, which can be used for prompting whether required movement is performed correctly.It is optional
, in the space coordinates, which can also have display user to make to act the reference spatial position of appraisal result, should
With reference to spatial position location information can be it is pre-set by application developer or operation maintenance personnel.Referring to FIG. 10,
It illustrates a kind of schematic diagram of a scenario for displaying prompt information that the embodiment of the present disclosure is related to, as shown in Figure 10, wherein containing
Scoring refers to spatial position 1001, which can show the appraisal result of the user action accordingly with reference to space bit
It sets.
In conclusion the disclosure can prompt whether required movement is executed correctly by the prompt information, so that user
The requirement point that when individually taking exercise using the AR helmet, can more accurately can grasp required movement, makes itself
Made athletic performance more standard, and then obtain more accurate exercise guidance.
Figure 11 is please referred to, a kind of carrying out based on AR equipment device that information is shown it illustrates the disclosure one is illustrative
Block diagram, as shown in figure 11, being somebody's turn to do the device shown based on AR equipment progress information can be by the combination of software, hardware or both
Realize some or all of of the AR equipment in implementation environment shown in Fig. 1.With execute in Fig. 2 or embodiment illustrated in fig. 4 by comprising
The step of AR equipment of image collection assembly executes.This carries out the device that information is shown based on AR equipment
First obtains module 1101, for obtaining the first location information being used to indicate with reference to spatial position;The reference
When spatial position is that required movement is executed correctly, the spatial position of the corresponding limbs of the required movement;
Image capture module 1102, for when it is described with reference to spatial position be in the AR equipment within sweep of the eye when,
The limbs image within the vision of the AR equipment is acquired by described image acquisition component;
Position determination module 1103 is determined in the limbs image and includes for according to the collected limbs image
, the real space positions of the corresponding limbs of the required movement;
Second obtains module 1104, for obtaining the matching with reference between spatial position and the real space position
Relationship, the matching relationship include matching or mismatch;
Information display module 1105, for showing prompt information according to the matching relationship, the prompt information is for mentioning
Show whether the required movement is executed correctly.
Optionally, described first module 1101 is obtained, is specifically used for,
Inquire it is pre-stored, when the required movement is executed correctly, the corresponding limbs of the required movement with refer to
Determine the coordinate information in the space coordinates that indicia patterns position is reference;
The coordinate information inquired is retrieved as the first location information.
Optionally, described device further include:
Third obtains module, for obtaining the first position being used to indicate with reference to spatial position in the first acquisition module
Before information, the modeling information of the required movement is obtained, the modeling information includes the action message of the required movement, with
And location information of the corresponding dummy model of the required movement in the space coordinates;
Creation module, for creating the dummy model according to the action message of required movement;
4th obtains module, for the location information according to the dummy model in the space coordinates, obtains institute
It states in dummy model, coordinate information of the corresponding limbs of the required movement in the space coordinates;
Memory module, for storing in the dummy model, the corresponding limbs of the required movement are in the space coordinate
Coordinate information in system.
Optionally, described device further include:
5th obtains module, is used in the 4th acquisition module according to the dummy model in the space coordinates
Location information, obtain in the dummy model, coordinate of the corresponding limbs of the required movement in the space coordinates
Before information, shape parameter is obtained, the shape parameter is used to indicate the figure of the wearer of the AR equipment;
Zoom module, in the location information according to the dummy model in the space coordinates, described in acquisition
In dummy model, the corresponding limbs of the required movement before the coordinate information in the space coordinates, according to the body
Shape parameter zooms in and out the dummy model;
It is described 4th obtain module, specifically for according to the dummy model after scaling in the space coordinates
Location information obtains in the dummy model, coordinate letter of the corresponding limbs of the required movement in the space coordinates
Breath.
Optionally, described device further include:
6th obtains module, for being in the AR equipment with reference to spatial position when described in described image acquisition module
When within sweep of the eye, before the limbs image within the vision that the AR equipment is acquired by described image acquisition component, obtain
Second location information and Viewing-angle information are taken, the second location information is used to indicate the AR equipment in the space coordinates
Spatial position, the Viewing-angle information is used to indicate the view directions of the AR equipment;
7th obtains module, for obtaining the AR equipment according to the second location information and the Viewing-angle information
The corresponding space coordinate boundary in the space coordinates of field range;
First determining module, for being in the space coordinate boundary when the corresponding space coordinate of the first location information
When internal, determine and described be in the AR equipment within sweep of the eye with reference to spatial position.
Optionally, the described 6th module is obtained, is specifically used for,
According to the AR equipment in previous moment to the moving parameter between current time, and, the AR equipment is preceding
The historical position information at one moment obtains the second location information;The historical position information is opened from initial position message
Begin, successively the moving parameter according to the AR equipment between each adjacent moment determines;
According to the AR equipment in previous moment to the attitudes vibration parameter between current time, and, the AR equipment
The Viewing-angle information described in the historical angel acquisition of information of previous moment;The historical angel information is opened from initial Viewing-angle information
Begin, successively the attitudes vibration parameter according to the AR equipment between each adjacent moment determines.
Optionally, described device further include:
Acquisition module, for passing through institute before the described 6th obtains module acquisition second location information and Viewing-angle information
State the tag image that image collection assembly acquisition includes the specified indicia patterns;
Second determining module, for according to the collected tag image, determine acquire the tag image when
It carves, the relative positional relationship between the AR equipment and the specified indicia patterns;
8th obtains module, for basis at the time of acquiring the tag image, the AR equipment and the specified mark
Remember that the relative positional relationship between pattern obtains the initial position message and the initial Viewing-angle information.
Optionally, the specified indicia patterns are arranged on Yoga mat surface.
Optionally, the information display module 1105, is specifically used for,
It is corresponding at the reference spatial position when the reference spatial position and the real space position mismatch
Show the first prompt information, first prompt information is for prompting the required movement not to be executed correctly;
It is corresponding to be shown at the reference spatial position when reference spatial position and the real space location matches
Show the second prompt information, second prompt information is for prompting the required movement to be executed correctly.
Optionally, first prompt information and second prompt information are the patterns that shape is identical and color is different.
It should be noted is that device provided by the above embodiment is when realizing its function, only with above-mentioned each function
The division progress of module, can be according to actual needs and by above-mentioned function distribution by different function for example, in practical application
Energy module is completed, i.e., the content structure of equipment is divided into different functional modules, to complete whole described above or portion
Divide function.
About the device in above-described embodiment, wherein modules execute the concrete mode of operation in related this method
Embodiment in be described in detail, no detailed explanation will be given here.
One exemplary embodiment of the disclosure additionally provides a kind of device shown based on AR equipment progress information, can be realized
The whole that is executed in the above-mentioned Fig. 2 of the disclosure or embodiment illustrated in fig. 4 by the augmented reality AR equipment comprising image collection assembly or
Person's part steps, should carry out the device that information is shown based on AR equipment includes:
Processor;
For storing the memory of the executable instruction of the processor;
Wherein, the processor is configured to:
Obtain the first location information being used to indicate with reference to spatial position;The reference spatial position is required movement by just
Really execute when, the corresponding limbs of the required movement spatial position;
When it is described with reference to spatial position be in the AR equipment within sweep of the eye when, adopted by described image acquisition component
Collect the limbs image within the vision of the AR equipment;
According to the collected limbs image, determine that including in the limbs image, the required movement is corresponding
The real space position of limbs;
The matching relationship with reference between spatial position and the real space position is obtained, the matching relationship includes
Matching mismatches;
Prompt information is shown according to the matching relationship, and the prompt information is for prompting the required movement whether by just
Really execute.
Optionally, the acquisition is used to indicate the first location information with reference to spatial position, comprising:
Inquire it is pre-stored, when the required movement is executed correctly, the corresponding limbs of the required movement with refer to
Determine the coordinate information in the space coordinates that indicia patterns position is reference;
The coordinate information inquired is retrieved as the first location information.
Optionally, the processor is also configured to
Before acquisition is used to indicate with reference to the first location information of spatial position,
The modeling information of the required movement is obtained, the modeling information includes the action message of the required movement, with
And location information of the corresponding dummy model of the required movement in the space coordinates;
The dummy model is created according to the action message of required movement;
According to location information of the dummy model in the space coordinates, obtains in the dummy model, is described
Coordinate information of the corresponding limbs of required movement in the space coordinates;
It stores in the dummy model, coordinate letter of the corresponding limbs of the required movement in the space coordinates
Breath.
Optionally, the processor is also configured to
In the location information according to the dummy model in the space coordinates, obtain in the dummy model, institute
The corresponding limbs of required movement are stated before the coordinate information in the space coordinates,
Shape parameter is obtained, the shape parameter is used to indicate the figure of the wearer of the AR equipment;
The dummy model is zoomed in and out according to the shape parameter;
The location information according to the dummy model in the space coordinates, obtain in the dummy model,
Coordinate information of the corresponding limbs of the required movement in the space coordinates, comprising:
According to location information of the dummy model in the space coordinates after scaling, the dummy model is obtained
In, coordinate information of the corresponding limbs of the required movement in the space coordinates.
Optionally, the processor is also configured to
When it is described with reference to spatial position be in the AR equipment within sweep of the eye when, pass through described image acquisition component
Before the limbs image within the vision for acquiring the AR equipment,
It obtains second location information and Viewing-angle information, the second location information is used to indicate the AR equipment in the sky
Between spatial position in coordinate system, the Viewing-angle information is used to indicate the view directions of the AR equipment;
According to the second location information and the Viewing-angle information, the field range for obtaining the AR equipment is corresponding described
Space coordinate boundary in space coordinates;
When the corresponding space coordinate of the first location information is in the space coordinate border inner, the ginseng is determined
It examines spatial position and is in the AR equipment within sweep of the eye.
Optionally, the acquisition second location information and Viewing-angle information, comprising:
According to the AR equipment in previous moment to the moving parameter between current time, and, the AR equipment is preceding
The historical position information at one moment obtains the second location information;The historical position information is opened from initial position message
Begin, successively the moving parameter according to the AR equipment between each adjacent moment determines;
According to the AR equipment in previous moment to the attitudes vibration parameter between current time, and, the AR equipment
The Viewing-angle information described in the historical angel acquisition of information of previous moment;The historical angel information is opened from initial Viewing-angle information
Begin, successively the attitudes vibration parameter according to the AR equipment between each adjacent moment determines.
Optionally, the processor is also configured to
Before obtaining second location information and Viewing-angle information,
It include the tag image of the specified indicia patterns by the acquisition of described image acquisition component;
According to the collected tag image, the AR equipment and institute at the time of acquiring the tag image are determined
State the relative positional relationship between specified indicia patterns;
According to opposite position at the time of acquiring the tag image, between the AR equipment and the specified indicia patterns
Set initial position message described in Relation acquisition and the initial Viewing-angle information.
Optionally, the specified indicia patterns are arranged on Yoga mat surface.
It is optionally, described that prompt information is shown according to the matching relationship, comprising:
It is corresponding at the reference spatial position when the reference spatial position and the real space position mismatch
Show the first prompt information, first prompt information is for prompting the required movement not to be executed correctly;
It is corresponding to be shown at the reference spatial position when reference spatial position and the real space location matches
Show the second prompt information, second prompt information is for prompting the required movement to be executed correctly.
Optionally, first prompt information and second prompt information are the patterns that shape is identical and color is different.
It is above-mentioned mainly by taking terminal and server as an example, the scheme provided the embodiment of the present disclosure is described.It can manage
Solution, in order to realize the above functions, it comprises execute the corresponding hardware configuration of each function and/or soft for terminal and server
Part module.Each exemplary module and algorithm steps of embodiment description in conjunction with disclosed in the disclosure, embodiment of the present disclosure energy
The combining form with hardware or hardware and computer software is reached to realize.Some function is driven actually with hardware or computer software
The mode of dynamic hardware executes, the specific application and design constraint depending on technical solution.Those skilled in the art can be with
Described function is realized using different methods to each specific application, but this realization is it is not considered that beyond this
The range of the technical solution of open embodiment.
Figure 12 is shown according to an exemplary embodiment a kind of for carrying out the device that information is shown based on AR equipment
1200 block diagram.For example, device 1200 may be provided as terminal or AR equipment.Referring to Fig.1 2, device 1200 includes processing
Component 1222 further comprises one or more processors, and the memory resource as representated by memory 1232, is used for
Store the instruction that can be executed by processing component 1222, such as application program.The application program stored in memory 1232 can wrap
Include it is one or more each correspond to one group of instruction module.Refer in addition, processing component 1222 is configured as executing
It enables, to execute in the above-mentioned method shown based on AR equipment progress information, the whole executed by terminal or AR equipment or portion
Step by step.
Device 1200 can also include that a power supply module 1226 be configured as the power management of executive device 1200, and one
Wired or wireless network interface 1250 is configured as device 1200 being connected to network and input and output (I/O) interface
1238.Device 1200 can be operated based on the operating system for being stored in memory 1232, such as Windows ServerTM, Mac
OS XTM, UnixTM, LinuxTM, FreeBSDTM or similar.
Those skilled in the art it will be appreciated that in said one or multiple examples, retouched by the embodiment of the present disclosure
The function of stating can be realized with hardware, software, firmware or their any combination.It when implemented in software, can be by this
A little functions storages in computer-readable medium or as on computer-readable medium one or more instructions or code into
Row transmission.Computer-readable medium includes computer storage media and communication media, and wherein communication media includes convenient for from one
Any medium of the place to another place transmission computer program.Storage medium can be general or specialized computer and can deposit
Any usable medium taken.
The embodiment of the present disclosure additionally provides a kind of computer storage medium, sets for being stored as above-mentioned terminal or smart home
Standby computer software instructions used, it includes for executing designed by the above-mentioned method shown based on AR equipment progress information
Program.
Those skilled in the art after considering the specification and implementing the invention disclosed here, will readily occur to its of the disclosure
Its embodiment.The disclosure is intended to cover any variations, uses, or adaptations of the disclosure, these modifications, purposes or
Person's adaptive change follows the general principles of this disclosure and including the undocumented common knowledge in the art of the disclosure
Or conventional techniques.The description and examples are only to be considered as illustrative, and the true scope and spirit of the disclosure are wanted by right
It asks and points out.
It should be understood that the present disclosure is not limited to the precise structures that have been described above and shown in the drawings, and
And various modifications and changes may be made without departing from the scope thereof.The scope of the present disclosure is only limited by the accompanying claims.
Claims (18)
1. a kind of carry out the method that information is shown based on AR equipment, which is characterized in that the method is by including image collection assembly
Augmented reality AR equipment execute, which comprises
Obtain the first location information being used to indicate with reference to spatial position;The reference spatial position is that required movement is correctly held
When row, the spatial position of the corresponding limbs of the required movement;
When it is described with reference to spatial position be in the AR equipment within sweep of the eye when, pass through described image acquisition component acquire institute
State the limbs image within the vision of AR equipment;
According to the collected limbs image, the corresponding limbs of the required movement that include in the limbs image, described are determined
Real space position;
The matching relationship with reference between spatial position and the real space position is obtained, the matching relationship includes matching
Or it mismatches;
Prompt information is shown according to the matching relationship, and the prompt information is for prompting whether the required movement is correctly held
Row.
2. the method according to claim 1, wherein the acquisition is used to indicate first with reference to spatial position
Confidence breath, comprising:
Inquire pre-stored, when the required movement is executed correctly, the corresponding limbs of the required movement are with specified mark
Note pattern position is the coordinate information in the space coordinates of reference;
The coordinate information inquired is retrieved as the first location information.
3. method according to claim 1 or 2, which is characterized in that be used to indicate first with reference to spatial position obtaining
Before location information, further includes:
The modeling information of the required movement is obtained, the modeling information includes the action message of the required movement, Yi Jisuo
State location information of the corresponding dummy model of required movement in the space coordinates;
The dummy model is created according to the action message of required movement;
According to location information of the dummy model in the space coordinates, obtains in the dummy model, is described specified
Act coordinate information of the corresponding limbs in the space coordinates;
It stores in the dummy model, coordinate information of the corresponding limbs of the required movement in the space coordinates.
4. according to the method described in claim 3, it is characterized in that, according to the dummy model in the space coordinates
Location information, obtain in the dummy model, coordinate of the corresponding limbs of the required movement in the space coordinates
Before information, further includes:
Shape parameter is obtained, the shape parameter is used to indicate the figure of the wearer of the AR equipment;
The dummy model is zoomed in and out according to the shape parameter;
The location information according to the dummy model in the space coordinates obtains in the dummy model, is described
Coordinate information of the corresponding limbs of required movement in the space coordinates, comprising:
According to location information of the dummy model in the space coordinates after scaling, obtain in the dummy model,
Coordinate information of the corresponding limbs of the required movement in the space coordinates.
5. according to the method described in claim 2, it is characterized in that, described be in the AR equipment with reference to spatial position when described
Within sweep of the eye when, before the limbs image within the vision that the AR equipment is acquired by described image acquisition component,
Further include:
It obtains second location information and Viewing-angle information, the second location information is used to indicate the AR equipment and sits in the space
Spatial position in mark system, the Viewing-angle information are used to indicate the view directions of the AR equipment;
According to the second location information and the Viewing-angle information, the field range for obtaining the AR equipment is corresponding in the space
Space coordinate boundary in coordinate system;
When the corresponding space coordinate of the first location information is in the space coordinate border inner, determine described with reference to empty
Between position be in the AR equipment within sweep of the eye.
6. according to the method described in claim 5, it is characterized in that, the acquisition second location information and Viewing-angle information, comprising:
According to the AR equipment in previous moment to the moving parameter between current time, and, the AR equipment is when previous
The historical position information at quarter obtains the second location information;The historical position information be since initial position message, according to
What the secondary moving parameter according to the AR equipment between each adjacent moment determined;
According to the AR equipment in previous moment to the attitudes vibration parameter between current time, and, the AR equipment is preceding
Viewing-angle information described in the historical angel acquisition of information at one moment;The historical angel information be since initial Viewing-angle information, according to
What the secondary attitudes vibration parameter according to the AR equipment between each adjacent moment determined.
7. according to the method described in claim 6, it is characterized in that, before the acquisition second location information and Viewing-angle information,
Further include:
It include the tag image of the specified indicia patterns by the acquisition of described image acquisition component;
It according to the collected tag image, determines at the time of acquiring the tag image, the AR equipment and the finger
Determine the relative positional relationship between indicia patterns;
According at the time of acquiring the tag image, the relative position between the AR equipment and the specified indicia patterns is closed
System obtains the initial position message and the initial Viewing-angle information.
8. the method according to claim 1, wherein described show prompt information, packet according to the matching relationship
It includes:
It is corresponding to be shown at the reference spatial position when the reference spatial position and the real space position mismatch
First prompt information, first prompt information is for prompting the required movement not to be executed correctly;
When it is described with reference to spatial position and the real space location matches when, it is corresponding described with reference to display at spatial position the
Two prompt informations, second prompt information is for prompting the required movement to be executed correctly.
9. a kind of carry out the device that information is shown based on AR equipment, which is characterized in that described device is to contain Image Acquisition group
In the augmented reality AR equipment of part, described device includes:
First obtains module, for obtaining the first location information being used to indicate with reference to spatial position;It is described to refer to spatial position
When being that required movement is executed correctly, the spatial position of the corresponding limbs of the required movement;
Image capture module, for when it is described with reference to spatial position be in the AR equipment within sweep of the eye when, by described
Image collection assembly acquires the limbs image within the vision of the AR equipment;
Position determination module, for determining finger that include in the limbs image, described according to the collected limbs image
Surely the real space position of corresponding limbs is acted;
Second obtains module, for obtaining the matching relationship with reference between spatial position and the real space position, institute
Matching relationship is stated to include matching or mismatch;
Information display module, for showing prompt information according to the matching relationship, the prompt information is for prompting the finger
Whether fixed movement is executed correctly.
10. device according to claim 11, which is characterized in that described first obtains module, is specifically used for,
Inquire pre-stored, when the required movement is executed correctly, the corresponding limbs of the required movement are with specified mark
Note pattern position is the coordinate information in the space coordinates of reference;
The coordinate information inquired is retrieved as the first location information.
11. device according to claim 9 or 10, which is characterized in that described device further include:
Third obtains module, for obtaining the first location information being used to indicate with reference to spatial position in the first acquisition module
Before, the modeling information of the required movement is obtained, the modeling information includes the action message of the required movement, Yi Jisuo
State location information of the corresponding dummy model of required movement in the space coordinates;
Creation module, for creating the dummy model according to the action message of required movement;
4th acquisition module obtains the void for the location information according to the dummy model in the space coordinates
In analog model, coordinate information of the corresponding limbs of the required movement in the space coordinates;
Memory module, for storing in the dummy model, the corresponding limbs of the required movement are in the space coordinates
Coordinate information.
12. device according to claim 11, which is characterized in that described device further include:
5th obtains module, for obtaining position of the module according to the dummy model in the space coordinates the described 4th
Confidence breath, obtains in the dummy model, coordinate information of the corresponding limbs of the required movement in the space coordinates
Before, shape parameter is obtained, the shape parameter is used to indicate the figure of the wearer of the AR equipment;
Zoom module, for obtaining described virtual in the location information according to the dummy model in the space coordinates
In model, the corresponding limbs of the required movement before the coordinate information in the space coordinates, according to the figure join
It is several that the dummy model is zoomed in and out;
Described 4th obtains module, specifically for according to position of the dummy model in the space coordinates after scaling
Information obtains in the dummy model, coordinate information of the corresponding limbs of the required movement in the space coordinates.
13. device according to claim 10, which is characterized in that described device further include:
6th obtains module, for acquiring the view of the AR equipment by described image acquisition component in described image acquisition module
Before limbs image in wild range, obtains second location information and Viewing-angle information, the second location information are used to indicate institute
Spatial position of the AR equipment in the space coordinates is stated, the Viewing-angle information is used to indicate the visual angle side of the AR equipment
To;
7th obtains module, for obtaining the visual field of the AR equipment according to the second location information and the Viewing-angle information
The corresponding space coordinate boundary in the space coordinates of range;
First determining module, for being in the space coordinate border inner when the corresponding space coordinate of the first location information
When, it determines and described is in the AR equipment within sweep of the eye with reference to spatial position.
14. device according to claim 13, which is characterized in that the described 6th obtains module, is specifically used for,
According to the AR equipment in previous moment to the moving parameter between current time, and, the AR equipment is when previous
The historical position information at quarter obtains the second location information;The historical position information be since initial position message, according to
What the secondary moving parameter according to the AR equipment between each adjacent moment determined;
According to the AR equipment in previous moment to the attitudes vibration parameter between current time, and, the AR equipment is preceding
Viewing-angle information described in the historical angel acquisition of information at one moment;The historical angel information be since initial Viewing-angle information, according to
What the secondary attitudes vibration parameter according to the AR equipment between each adjacent moment determined.
15. device according to claim 14, which is characterized in that described device further include:
Acquisition module, for passing through the figure before the described 6th obtains module acquisition second location information and Viewing-angle information
As acquisition component acquisition includes the tag image of the specified indicia patterns;
Second determining module, for determining institute at the time of acquiring the tag image according to the collected tag image
State the relative positional relationship between AR equipment and the specified indicia patterns;
8th obtains module, for according at the time of acquiring the tag image, the AR equipment and the specified label to be schemed
Relative positional relationship between case obtains the initial position message and the initial Viewing-angle information.
16. device according to claim 9, which is characterized in that the information display module is specifically used for,
It is corresponding to be shown at the reference spatial position when the reference spatial position and the real space position mismatch
First prompt information, first prompt information is for prompting the required movement not to be executed correctly;
When it is described with reference to spatial position and the real space location matches when, it is corresponding described with reference to display at spatial position the
Two prompt informations, second prompt information is for prompting the required movement to be executed correctly.
17. a kind of carry out the device that information is shown based on AR equipment, which is characterized in that described device is to contain Image Acquisition group
In the augmented reality AR equipment of part, described device includes:
Processor;
For storing the memory of the executable instruction of the processor;
Wherein, the processor is configured to:
Obtain the first location information being used to indicate with reference to spatial position;The reference spatial position is that required movement is correctly held
When row, the spatial position of the corresponding limbs of the required movement;
When it is described with reference to spatial position be in the AR equipment within sweep of the eye when, pass through described image acquisition component acquire institute
State the limbs image within the vision of AR equipment;
According to the collected limbs image, the corresponding limbs of the required movement that include in the limbs image, described are determined
Real space position;
The matching relationship with reference between spatial position and the real space position is obtained, the matching relationship includes matching
Or it mismatches;
Prompt information is shown according to the matching relationship, and the prompt information is for prompting whether the required movement is correctly held
Row.
18. a kind of computer readable storage medium, which is characterized in that refer in the computer readable storage medium comprising executable
It enables, processor calls the executable instruction to realize that the claims 1 to 8 are any and described carry out information based on AR equipment
The method of display.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910363552.1A CN110110647A (en) | 2019-04-30 | 2019-04-30 | The method, apparatus and storage medium that information is shown are carried out based on AR equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910363552.1A CN110110647A (en) | 2019-04-30 | 2019-04-30 | The method, apparatus and storage medium that information is shown are carried out based on AR equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN110110647A true CN110110647A (en) | 2019-08-09 |
Family
ID=67487935
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910363552.1A Pending CN110110647A (en) | 2019-04-30 | 2019-04-30 | The method, apparatus and storage medium that information is shown are carried out based on AR equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110110647A (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112184920A (en) * | 2020-10-12 | 2021-01-05 | 中国联合网络通信集团有限公司 | AR-based skiing blind area display method and device and storage medium |
CN112348942A (en) * | 2020-09-18 | 2021-02-09 | 当趣网络科技(杭州)有限公司 | Body-building interaction method and system |
WO2021036954A1 (en) * | 2019-08-30 | 2021-03-04 | 华为技术有限公司 | Intelligent speech playing method and device |
CN113288087A (en) * | 2021-06-25 | 2021-08-24 | 成都泰盟软件有限公司 | Virtual-real linkage experimental system based on physiological signals |
US11880956B2 (en) | 2019-08-28 | 2024-01-23 | Shenzhen Sensetime Technology Co., Ltd. | Image processing method and apparatus, and computer storage medium |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102667881A (en) * | 2010-03-30 | 2012-09-12 | 新日铁系统集成株式会社 | Information processing apparatus, information processing method, and program |
CN103732297A (en) * | 2011-06-30 | 2014-04-16 | 奥林奇实验室 | Augmented-reality range-of-motion therapy system and method of operation thereof |
CN105027190A (en) * | 2013-01-03 | 2015-11-04 | 美达公司 | Extramissive spatial imaging digital eye glass for virtual or augmediated vision |
CN106110627A (en) * | 2016-06-20 | 2016-11-16 | 曲大方 | Physical culture and Wushu action correction equipment and method |
CN106233227A (en) * | 2014-03-14 | 2016-12-14 | 索尼互动娱乐股份有限公司 | There is the game device of volume sensing |
CN107564089A (en) * | 2017-08-10 | 2018-01-09 | 腾讯科技(深圳)有限公司 | Three dimensional image processing method, device, storage medium and computer equipment |
CN108434664A (en) * | 2018-04-08 | 2018-08-24 | 上海应用技术大学 | A kind of treadmill intelligent safe protector and guard method |
CN108579020A (en) * | 2018-03-12 | 2018-09-28 | 广州大象健康科技有限公司 | A kind of intelligence science movement technique and kinematic system |
CN108960002A (en) * | 2017-05-17 | 2018-12-07 | 中兴通讯股份有限公司 | A kind of movement adjustment information reminding method and device |
CN108984076A (en) * | 2017-06-01 | 2018-12-11 | 中兴通讯股份有限公司 | A kind of display methods and device |
CN109191588A (en) * | 2018-08-27 | 2019-01-11 | 百度在线网络技术(北京)有限公司 | Move teaching method, device, storage medium and electronic equipment |
-
2019
- 2019-04-30 CN CN201910363552.1A patent/CN110110647A/en active Pending
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102667881A (en) * | 2010-03-30 | 2012-09-12 | 新日铁系统集成株式会社 | Information processing apparatus, information processing method, and program |
CN103732297A (en) * | 2011-06-30 | 2014-04-16 | 奥林奇实验室 | Augmented-reality range-of-motion therapy system and method of operation thereof |
CN105027190A (en) * | 2013-01-03 | 2015-11-04 | 美达公司 | Extramissive spatial imaging digital eye glass for virtual or augmediated vision |
CN106233227A (en) * | 2014-03-14 | 2016-12-14 | 索尼互动娱乐股份有限公司 | There is the game device of volume sensing |
CN106110627A (en) * | 2016-06-20 | 2016-11-16 | 曲大方 | Physical culture and Wushu action correction equipment and method |
CN108960002A (en) * | 2017-05-17 | 2018-12-07 | 中兴通讯股份有限公司 | A kind of movement adjustment information reminding method and device |
CN108984076A (en) * | 2017-06-01 | 2018-12-11 | 中兴通讯股份有限公司 | A kind of display methods and device |
CN107564089A (en) * | 2017-08-10 | 2018-01-09 | 腾讯科技(深圳)有限公司 | Three dimensional image processing method, device, storage medium and computer equipment |
CN108579020A (en) * | 2018-03-12 | 2018-09-28 | 广州大象健康科技有限公司 | A kind of intelligence science movement technique and kinematic system |
CN108434664A (en) * | 2018-04-08 | 2018-08-24 | 上海应用技术大学 | A kind of treadmill intelligent safe protector and guard method |
CN109191588A (en) * | 2018-08-27 | 2019-01-11 | 百度在线网络技术(北京)有限公司 | Move teaching method, device, storage medium and electronic equipment |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11880956B2 (en) | 2019-08-28 | 2024-01-23 | Shenzhen Sensetime Technology Co., Ltd. | Image processing method and apparatus, and computer storage medium |
WO2021036954A1 (en) * | 2019-08-30 | 2021-03-04 | 华为技术有限公司 | Intelligent speech playing method and device |
CN112348942A (en) * | 2020-09-18 | 2021-02-09 | 当趣网络科技(杭州)有限公司 | Body-building interaction method and system |
CN112348942B (en) * | 2020-09-18 | 2024-03-19 | 当趣网络科技(杭州)有限公司 | Body-building interaction method and system |
CN112184920A (en) * | 2020-10-12 | 2021-01-05 | 中国联合网络通信集团有限公司 | AR-based skiing blind area display method and device and storage medium |
CN112184920B (en) * | 2020-10-12 | 2023-06-06 | 中国联合网络通信集团有限公司 | AR-based skiing blind area display method, device and storage medium |
CN113288087A (en) * | 2021-06-25 | 2021-08-24 | 成都泰盟软件有限公司 | Virtual-real linkage experimental system based on physiological signals |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110110647A (en) | The method, apparatus and storage medium that information is shown are carried out based on AR equipment | |
CN111460872B (en) | Image processing method and device, image equipment and storage medium | |
CN107349594B (en) | A kind of action evaluation method of virtual Dance System | |
CN109011508A (en) | A kind of intelligent coach system and method | |
CN109087379B (en) | Facial expression migration method and facial expression migration device | |
CN103400119B (en) | Face recognition technology-based mixed reality spectacle interactive display method | |
CN109815776B (en) | Action prompting method and device, storage medium and electronic device | |
CN107016704A (en) | A kind of virtual reality implementation method based on augmented reality | |
CN107243141A (en) | A kind of action auxiliary training system based on motion identification | |
CN106139564A (en) | Image processing method and device | |
CN109432753A (en) | Act antidote, device, storage medium and electronic equipment | |
CN109086798A (en) | A kind of data mask method and annotation equipment | |
CN108920490A (en) | Assist implementation method, device, electronic equipment and the storage medium of makeup | |
CN106166376A (en) | Simplify taijiquan in 24 forms comprehensive training system | |
CN107281710A (en) | A kind of method of remedial action error | |
CN105763829A (en) | Image processing method and electronic device | |
CN107961524A (en) | Body-building game and training system based on AR | |
KR102320960B1 (en) | Personalized home training behavior guidance and correction system | |
CN111383735A (en) | Unmanned body-building analysis method based on artificial intelligence | |
CN109091380A (en) | Traditional Chinese medicine system and method for realizing acupoint visualization by AR technology | |
CN108038601A (en) | Teaching methods of physical education, device and computer-readable recording medium | |
CN113542378A (en) | Remote rehabilitation service-oriented interactive exercise training method and device, computer equipment and storage medium | |
CN111967407B (en) | Action evaluation method, electronic device, and computer-readable storage medium | |
CN115131879B (en) | Action evaluation method and device | |
Xie et al. | Visual feedback for core training with 3d human shape and pose |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |