CN109840948A - The put-on method and device of target object based on augmented reality - Google Patents
The put-on method and device of target object based on augmented reality Download PDFInfo
- Publication number
- CN109840948A CN109840948A CN201711226533.1A CN201711226533A CN109840948A CN 109840948 A CN109840948 A CN 109840948A CN 201711226533 A CN201711226533 A CN 201711226533A CN 109840948 A CN109840948 A CN 109840948A
- Authority
- CN
- China
- Prior art keywords
- target object
- described image
- threedimensional model
- data
- module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Landscapes
- Processing Or Creating Images (AREA)
Abstract
The embodiment of the invention discloses the put-on methods and device of a kind of target object based on augmented reality.The put-on method and device of the target object are by carrying out three-dimensional reconstruction to target object, obtain the three-dimensional information of target object, then the normal direction of model any position is obtained by three-dimensional data model again, followed by the correlation data calculation direction of illumination of obtained normal data and described image, light source and target object are launched with corresponding position in the scene, to obtain the put-on method and device of a kind of target object of simple and effective augmented reality.
Description
Technical field
The present invention relates to the direction of illuminations of target object in technical field of virtual reality more particularly to a kind of augmented reality
Determine method and apparatus.
Background technique
AR (Augmented Reality, augmented reality) is a kind of by real world information and virtual world information " nothing
The integrated new technology of seam ", be script is difficult to experience in the certain time spatial dimension of real world entity information (depending on
Feel information, sound, taste, tactile etc.) by science and technology such as computers, after analog simulation, true environment and virtual object
Body has been added to the same picture in real time or space exists simultaneously, and virtual Information application to real world is felt by the mankind
Official is perceived, to reach the sensory experience of exceeding reality.Augmented reality not only presents the information of real world, and
And show virtual information simultaneously, two kinds of information are complementary to one another, are superimposed.In the augmented reality of visualization, Yong Huli
With Helmet Mounted Display, real world with computer graphic is multiple is synthesized together, the true world can be seen around it.
AR is a kind of new way that people pass through that computer carries out visualized operation to complex data and interacts, with traditional man-machine boundary
The window-operating of face and prevalence is compared, and AR is haveed a qualitative leap in technical idea.
The simulation that problems faced is exactly illumination, traditional illuminant direction estimation method are frequently necessary in augmented reality device
It is the direction based on image analysis illumination, normal direction of this method due to that cannot obtain object from image gives illumination side
To analysis bring difficulty.Therefore, a kind of determination side of the direction of illumination of target object based on augmented reality is needed now
Method.
Summary of the invention
The embodiment of the invention provides a kind of determination methods and dress based on the direction of illumination of target object in augmented reality
It sets, can simply and effectively determine the direction of illumination of target object in augmented reality.
According to an aspect of the present invention, the present invention provides a kind of dispensing sides of target object based on augmented reality
Method, comprising:
S101, the threedimensional model for obtaining the target object, obtain the phase of the threedimensional model according to the threedimensional model
Close data;
S102, image is obtained, and described image is obtained according to described image and the related data of the threedimensional model
Related data;
S103, the target object is determined by the related data of the threedimensional model and the related data of described image
Direction of illumination;
S104, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, and according to the figure
The related data of picture launches target object in the scene.
The present invention provides the put-on method and device of a kind of target object based on augmented reality, the target object
Put-on method and device obtain the three-dimensional information of target object, then pass through three again by carrying out three-dimensional reconstruction to target object
D Data Model obtains the normal direction of model any position, followed by the dependency number of obtained normal data and described image
According to direction of illumination is calculated, light source and target object are launched with corresponding position in the scene, to obtain a kind of simply and effectively increasing
The put-on method and device of the target object of strong reality.
Detailed description of the invention
To describe the technical solutions in the embodiments of the present invention more clearly, make required in being described below to embodiment
Attached drawing is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the invention, for
For those skilled in the art, without creative efforts, it can also be obtained according to these attached drawings other attached
Figure.
Fig. 1 is the flow diagram provided in an embodiment of the present invention based on the put-on method of target object in augmented reality.
Fig. 2 is the application scenarios signal provided in an embodiment of the present invention based on the delivery device of target object in augmented reality
Figure.
Fig. 3 is a kind of structural representation provided in an embodiment of the present invention based on the delivery device of target object in augmented reality
Figure.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete
Site preparation description.Obviously, described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.It is based on
Embodiment in the present invention, those skilled in the art's every other implementation obtained without creative efforts
Example, shall fall within the protection scope of the present invention.
Description and claims of this specification and term " first " in above-mentioned attached drawing, " second ", " third " etc.
(if present) is to be used to distinguish similar objects, without being used to describe a particular order or precedence order.It should be appreciated that this
The object of sample description is interchangeable under appropriate circumstances.In addition, term " includes " and " having " and their any deformation, meaning
Figure, which is to cover, non-exclusive includes.
In patent document, the attached drawing that is discussed herein below and for describing each embodiment of principle disclosed by the invention only
For illustrating, and should not be construed as limiting the scope of the present disclosure.Those skilled in the art will appreciate that original of the invention
Reason can be implemented in any device suitably arranged.It will be explained in illustrative embodiments, these realities be shown in the attached drawings
The example for applying mode.In addition, terminal accoding to exemplary embodiment will be described in detail with reference to the attached drawings.Identical attached drawing mark in attached drawing
Number refer to identical element.
Term used in description of the invention is only used to describe particular implementation, and is not intended to show of the invention
Concept.Unless have clearly different meanings in context, it is otherwise, used in the singular to express the table for covering plural form
It reaches.In the description of the present invention, it should be appreciated that there are this hairs for the terms meant for illustration such as " comprising ", " having " and " containing "
A possibility that feature for being disclosed in bright specification, number, step, movement or combinations thereof, and be not intended to exclude may be present or can
A possibility that adding other one or more features, number, step, movement or combinations thereof.Same reference numerals in attached drawing refer to
For same section.
The embodiment of the invention provides a kind of put-on method and device based on target object in augmented reality.It will divide below
It is not described in detail.
In a preferred embodiment, a kind of put-on method and device based on target object in augmented reality, such as Fig. 1 are provided
Shown, process can be such that
S101, the threedimensional model for obtaining the target object, obtain the phase of the threedimensional model according to the threedimensional model
Close data.
Specifically, the object can be obtained using spatial digitizer, binocular solid and AUTODESK 123D technology
The threedimensional model of body;The method for obtaining target object threedimensional model is all based on the three-dimensional reconstruction side of target object picture
Method.
For example, three-dimensional reconstruction is carried out to target object using AUTODESK 123D technology, if need to only shoot to target object
The three-dimensional reconstruction to target object can be realized in dry photo.
The normal of the target object surface point is obtained after the threedimensional model for obtaining target object, and institute is calculated
State the first derivative of normal.
S102, image is obtained, and described image is obtained according to described image and the related data of the threedimensional model
Related data.
In embodiments of the present invention, the step of related data of acquisition image and described image includes:
Image is obtained by camera;
The position orientation relation of the camera Yu the target object is obtained according to the flag information in described image;
According to the phase of position orientation relation and the target object threedimensional model between the camera and the target object
It closes data to be projected, obtains the coordinate of described image corresponding with the threedimensional model.
This step is in order to which determination need to project the specific coordinate of target object in reality scene, in next step in reality
Transmission goal object provides basis in scene.
S103, the target object is determined by the related data of the threedimensional model and the related data of described image
Direction of illumination.
In embodiments of the present invention, institute is determined by the related data of the threedimensional model and the related data of described image
The step of stating the direction of illumination of target object include:
Described image is filtered;
According to the coordinate of described image, the brightness data of described image is obtained from described image;
The target object is calculated using the brightness data of described image and the first derivative of the target object normal
Direction of illumination.
Common illuminant direction estimation method is the analysis carried out to continuous image, and such words operand is huge.For
Meet the needs of augmented reality target object device is to real-time, in the present invention, the method analyzed using discrete point,
Reduce operand, is tested by majority, it is final to determine on the threedimensional model of target object in embodiments of the present invention uniformly
40~60 sampled points are chosen, and the sampling can directly be extracted from described image by the image coordinate of the sampled point
The brightness value of point, and then the direction of illumination of the target object is calculated.
S104, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, and according to the figure
The related data of picture launches target object in the scene.
In embodiments of the present invention, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene,
And the step of launching target object in the scene according to the related data of described image, includes:
According to the direction of illumination of the target object, light source is launched in corresponding position in the scene;
According to the position orientation relation of the camera and target object drop target object in the scene.
Wherein, the position orientation relation according to the camera and the target object launch in the scene target object it
Before, the target object direction of illumination standard can also be checked to reach by detecting to the pre- image for launching target object
Purpose whether really.
The specific steps in the check stage are as follows:
Obtain the pre- image for launching target object;
Obtain the pre- brightness data launched in target object image;
The pre- image for launching target object is subjected to brightness contrast with the image obtained by camera, if described
Within the brightness value difference 5% of two images, then launched;If the brightness value of two image differs by more than 5%, then return
S101 step.
From the foregoing, it will be observed that the present invention provides a kind of put-on method of target object based on augmented reality, the object
The put-on method of body obtains the three-dimensional information of target object, then passes through three-dimensional again by carrying out three-dimensional reconstruction to target object
Data model obtains the normal direction of model any position, followed by the related data of obtained normal data and described image
Direction of illumination is calculated, light source and target object are launched with corresponding position in the scene, to obtain a kind of simply and effectively enhancing
The put-on method and device of the target object of reality.
With reference to Fig. 2, further embodiment of this invention provides a kind of video capture device based on virtual reality.Such as figure, the view
Frequency filming apparatus includes: photographic equipment 33, server 34 and control display equipment 36.
Wherein, photographic equipment 33 can be video camera, camera or the electronic equipment with camera function, can be used for
Acquire image information.For example, the photographic equipment can be used for limb action, expression and language message of user etc..
Server 34 is specifically as follows the network equipments such as data server, network server.The server 34 can be used for mentioning
For prop model 341 and the 3-D image of virtual scene 342.
Display control apparatus 36 may include the intelligence that computer, smart phone, tablet computer etc. have operation processing function
It can equipment.
In another embodiment in embodiment, a kind of one kind of delivery device based on target object in strong reality is provided and is answered
Use schematic diagram of a scenario.Below by the device based on above-mentioned target object, with the prop model 22 for target object physical model 22
For the target object put-on method is described in detail.
In the present embodiment, target object entity is exchanged for by scanning means 21 (such as spatial digitizer, binocular solid)
22 threedimensional model, and the related data of the threedimensional model according to the obtaining three-dimensional model.Then, pass through camera 23
Obtain the related data of 22 image of image and the target object of target object 22;Pass through the related data of the threedimensional model
Go out the direction of illumination of the target object 22 with the correlation data calculation of 22 image of target object, finally target in the scene
The relevant position of 22 direction of illumination of object places light source, and corresponding position dispensing image 25 in the scene.
As shown in figure 3, user 31 does the movement of weapon delivery to strange beast model 341, server 34 will be according to user's 31
Limb action generates corresponding light wave (i.e. virtual scene 342).Strange beast model 341 by the movement based on 31 weapon delivery of user,
Show the state gone down.Photographic equipment 33 captures the information such as limb action, expression, the language of user 31, real-time Transmission
It is shown to display control terminal 36, limb action of the server 34 based on user 31, in real time by virtual scene 342 and monster
Beast model 341 is transmitted to display control terminal 36 and is shown, to obtain fused image 35.
In practical application, in order to promote the authenticity of scene, a fusing device can be increased in augmented reality device,
For being rendered and being merged the target object image of dispensing with display scene.
From the foregoing, it will be observed that the put-on method provided in an embodiment of the present invention based on the target object in augmented reality, by right
The target object that need to be launched establishes threedimensional model to obtain the related data of the threedimensional model, to improve target in augmented reality
Effect is launched in the accuracy of object direction of illumination, enhancing.
In still another embodiment of the process, a kind of delivery device of target object based on augmented reality is also provided.Such as Fig. 3
Shown, being somebody's turn to do the video capture device based on virtual reality may include object module 31, scan module 32, Target Acquisition module
33, direction of illumination computing module 34 and putting module 35, in which:
Object module 31, for providing target object physical model;
Scan module 32, for scanning the object module and providing the three-dimensional data of the object module;
Target Acquisition module 33, for obtaining the image information data of object module;
Direction of illumination computing module 34, for being obtained according to the three-dimensional data of described image information data and the object module
Take the light source direction of the target object.
Putting module 35, for placing light source according to the corresponding position of the light source direction of the target object in the scene,
And target object is launched according to described image information data and the three-dimensional data in the scene
In some embodiments, the scan module 32 may include data establish unit 321, entity scanning element 322,
Data processing unit 323, in which:
Data establish unit 321, for establishing the information database of the object module;
Entity scanning element 322, for obtaining the threedimensional model of object module and three bit scan model datas and being stored in
In the information database;
Data processing unit 323, for calculated according to the threedimensional model threedimensional model normal and the method
The first derivative of line.
In some embodiments, the Target Acquisition module 33 may include image acquisition unit 331, image processing unit
332 and coordinate calculating unit 333.
Image acquisition unit 331, for obtaining the image information of object module;
Image processing unit 332, for calculating described image acquiring unit and institute according to the flag information in described image
State the position orientation relation of object module;
Coordinate calculating unit 333, for according to described image acquiring unit and the object module position orientation relation and institute
The scan data for stating threedimensional model is projected, and processing obtains the corresponding image coordinate of threedimensional model.
In some embodiments, the direction calculating unit 34 may include that brightness data unit 341 and direction of illumination are true
Order member 342.
Brightness data unit 341, for extracting the brightness data of described image from described image;
Direction of illumination determination unit 342, for reciprocal really according to the brightness data of described image and the single order of the normal
The direction of illumination for the module that sets the goal.
In some embodiments, the putting module 35 includes integrated unit 351 and dispensing unit 352.
Integrated unit 351, for merging the image for passing through rendering with the reality scene;
Unit 352 is launched, launches described image and the light source for the corresponding position in display scene
From the foregoing, it will be observed that the embodiment of the present invention obtains a kind of delivery device of target object based on augmented reality, pass through
Three-dimensional reconstruction is carried out to object module, the three-dimensional information of target object is obtained, model is then obtained by three-dimensional data model again
The normal direction of any position, followed by the correlation data calculation direction of illumination of obtained normal data and described image, with
Corresponding position launches light source and target object in the scene, thus obtain a kind of target object of simple and effective augmented reality
Put-on method and device.
Term " one " and " described " and similar word have been used during describing idea of the invention (especially
In the appended claims), it should be construed to not only cover odd number by these terms but also cover plural number.In addition, unless herein
In be otherwise noted, otherwise herein narration numberical range when referred to merely by quick method and belong to the every of relevant range
A independent value, and each independent value is incorporated into this specification, just as these values have individually carried out statement one herein
Sample.In addition, unless otherwise stated herein or context has specific opposite prompt, otherwise institute described herein is methodical
Step can be executed by any appropriate order.Change of the invention is not limited to the step of description sequence.Unless in addition
Advocate, is otherwise all only using any and all example or exemplary language presented herein (for example, " such as ")
Idea of the invention is better described, and not the range of idea of the invention limited.Spirit and model are not being departed from
In the case where enclosing, those skilled in the art becomes readily apparent that a variety of modifications and adaptation.
It is provided for the embodiments of the invention a kind of put-on method and device based on target object in augmented reality above
It is described in detail.It should be understood that illustrative embodiments as described herein should be to be considered only as descriptive, it is used to help manage
Method and its core concept of the invention are solved, and are not intended to restrict the invention.To feature in each illustrative embodiments
Or the description of aspect should usually be considered the similar features or aspects suitable for other exemplary embodiments.Although reference example
Property embodiment describe the present invention, but can suggest that those skilled in the art carries out various change and change.The invention is intended to
Cover these variations and the change in the scope of the appended claims.
Claims (10)
1. a kind of put-on method of the target object based on augmented reality characterized by comprising
S101, the threedimensional model for obtaining the target object, obtain the dependency number of the threedimensional model according to the threedimensional model
According to;
S102, image is obtained, and obtains the correlation of described image according to described image and the related data of the threedimensional model
Data;
S103, the illumination that the target object is determined by the related data of the threedimensional model and the related data of described image
Direction;
S104, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, and according to described image
Related data launches target object in the scene.
2. the put-on method of the target object according to claim 1 based on augmented reality, which is characterized in that described in acquisition
The threedimensional model of target object, the step of obtaining the related data of the threedimensional model according to the threedimensional model include:
The target object is scanned, the threedimensional model of the target object is obtained;
The normal of the target object and the single order of the target object normal are calculated according to the threedimensional model of the target object
Derivative.
3. the put-on method of the target object according to claim 2 based on augmented reality, which is characterized in that obtain figure
Picture, and the step of obtaining the related data of described image according to described image and the related data of the threedimensional model includes:
Image is obtained by camera;
The position orientation relation of the camera Yu the target object is obtained according to the flag information in described image;
According to the dependency number of position orientation relation and the target object threedimensional model between the camera and the target object
According to being projected, the coordinate of described image corresponding with the threedimensional model is obtained.
4. the put-on method of the target object according to claim 3 based on augmented reality, which is characterized in that by described
The step of related data of threedimensional model and the related data of described image determine the direction of illumination of the target object include:
Described image is filtered;
According to the coordinate of described image, the brightness data of described image is obtained from described image;
The illumination of the target object is calculated using the brightness data of described image and the first derivative of the target object normal
Direction.
5. the put-on method of the target object according to claim 4 based on augmented reality, which is characterized in that in the field
Light source is placed in the corresponding position of the target object direction of illumination in scape, and in the scene according to the related data of described image
Launch target object the step of include:
According to the direction of illumination of the target object, light source is launched in corresponding position in the scene;
According to the position orientation relation of the camera and target object drop target object in the scene.
6. a kind of delivery device of the target object based on augmented reality characterized by comprising
Object module, for providing the physical model for launching target;
Scan module, for scanning the object module and providing the three-dimensional data of the object module;
Target Acquisition module, for obtaining the image information data of object module;
Directions calculation module, for obtaining the target according to the three-dimensional data of described image information data and the object module
The light source direction of object;
Putting module, for placing light source, Yi Jigen according to the corresponding position of the light source direction of the target object in the scene
Target object is launched in the scene according to described image information data and the three-dimensional data.
7. a kind of delivery device of target object based on augmented reality according to claim 6, which is characterized in that described
Scan module includes:
Data establish unit, for establishing the information database of the object module;
Entity scanning element, for obtaining the threedimensional model of object module and three bit scan model datas and being stored in the information
In database;
Data processing unit, for calculating the normal of the threedimensional model and the single order of the normal according to the threedimensional model
Derivative.
8. a kind of delivery device of target object based on augmented reality according to claim 7, which is characterized in that target
Obtaining module includes:
Image acquisition unit, for obtaining the image information of object module;
Image processing unit, for calculating described image acquiring unit and the target mould according to the flag information in described image
The position orientation relation of block;
Coordinate calculating unit, for according to the position orientation relation of described image acquiring unit and the object module and the three-dimensional mould
The scan data of type is projected, and processing obtains the corresponding image coordinate of threedimensional model.
9. the delivery device of the target object based on augmented reality according to claim 8, which is characterized in that the telegoniometer
Calculating unit includes:
Brightness data unit, for extracting the brightness data of described image from described image;
Direction of illumination determination unit, for determining target mould according to the brightness data of described image and the single order inverse of the normal
The direction of illumination of block.
10. the delivery device of 6 a kind of target object based on augmented reality according to claim, which is characterized in that institute
Stating putting module includes:
Integrated unit, for merging the image for passing through rendering with the reality scene;
Unit is launched, launches described image and the light source for the corresponding position in display scene.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711226533.1A CN109840948B (en) | 2017-11-29 | 2017-11-29 | Target object throwing method and device based on augmented reality |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711226533.1A CN109840948B (en) | 2017-11-29 | 2017-11-29 | Target object throwing method and device based on augmented reality |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109840948A true CN109840948A (en) | 2019-06-04 |
CN109840948B CN109840948B (en) | 2023-08-15 |
Family
ID=66882043
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711226533.1A Active CN109840948B (en) | 2017-11-29 | 2017-11-29 | Target object throwing method and device based on augmented reality |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109840948B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112991556A (en) * | 2021-05-12 | 2021-06-18 | 航天宏图信息技术股份有限公司 | AR data display method and device, electronic equipment and storage medium |
CN113763090A (en) * | 2020-11-06 | 2021-12-07 | 北京沃东天骏信息技术有限公司 | Information processing method and device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100962557B1 (en) * | 2009-02-05 | 2010-06-11 | 한국과학기술원 | Augmented reality implementation apparatus and method of the same |
CN103218854A (en) * | 2013-04-01 | 2013-07-24 | 成都理想境界科技有限公司 | Method for realizing component marking during augmented reality process and augmented reality system |
CN106981087A (en) * | 2017-04-05 | 2017-07-25 | 杭州乐见科技有限公司 | Lighting effect rendering intent and device |
CN107071388A (en) * | 2016-12-26 | 2017-08-18 | 深圳增强现实技术有限公司 | A kind of three-dimensional augmented reality display methods and device |
-
2017
- 2017-11-29 CN CN201711226533.1A patent/CN109840948B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100962557B1 (en) * | 2009-02-05 | 2010-06-11 | 한국과학기술원 | Augmented reality implementation apparatus and method of the same |
CN103218854A (en) * | 2013-04-01 | 2013-07-24 | 成都理想境界科技有限公司 | Method for realizing component marking during augmented reality process and augmented reality system |
CN107071388A (en) * | 2016-12-26 | 2017-08-18 | 深圳增强现实技术有限公司 | A kind of three-dimensional augmented reality display methods and device |
CN106981087A (en) * | 2017-04-05 | 2017-07-25 | 杭州乐见科技有限公司 | Lighting effect rendering intent and device |
Non-Patent Citations (1)
Title |
---|
杜书侠等: "增强现实虚拟人中的光照方向估计", 《光学技术》 * |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113763090A (en) * | 2020-11-06 | 2021-12-07 | 北京沃东天骏信息技术有限公司 | Information processing method and device |
CN113763090B (en) * | 2020-11-06 | 2024-05-21 | 北京沃东天骏信息技术有限公司 | Information processing method and device |
CN112991556A (en) * | 2021-05-12 | 2021-06-18 | 航天宏图信息技术股份有限公司 | AR data display method and device, electronic equipment and storage medium |
CN112991556B (en) * | 2021-05-12 | 2022-05-27 | 航天宏图信息技术股份有限公司 | AR data display method and device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN109840948B (en) | 2023-08-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Oufqir et al. | ARKit and ARCore in serve to augmented reality | |
CN108447043B (en) | Image synthesis method, equipment and computer readable medium | |
CN107066082B (en) | Display methods and device | |
RU2586566C1 (en) | Method of displaying object | |
CN110866977B (en) | Augmented reality processing method, device, system, storage medium and electronic equipment | |
US20150279044A1 (en) | Method and apparatus for obtaining 3d face model using portable camera | |
CN106325509A (en) | Three-dimensional gesture recognition method and system | |
CN109815776B (en) | Action prompting method and device, storage medium and electronic device | |
CN106355153A (en) | Virtual object display method, device and system based on augmented reality | |
CN113822977A (en) | Image rendering method, device, equipment and storage medium | |
KR101885090B1 (en) | Image processing apparatus, apparatus and method for lighting processing | |
CN110458924B (en) | Three-dimensional face model establishing method and device and electronic equipment | |
CN110147737B (en) | Method, apparatus, device and storage medium for generating video | |
CN110059624B (en) | Method and apparatus for detecting living body | |
US20220358662A1 (en) | Image generation method and device | |
CN111833457A (en) | Image processing method, apparatus and storage medium | |
Bergig et al. | In-place augmented reality | |
EP3141985A1 (en) | A gazed virtual object identification module, a system for implementing gaze translucency, and a related method | |
CN109840948A (en) | The put-on method and device of target object based on augmented reality | |
KR20140126529A (en) | Physical Movement of Object on Reality-Augmented Reality Interaction System and Implementation Method for Electronic book | |
CN114067085A (en) | Virtual object display method and device, electronic equipment and storage medium | |
CN109816791B (en) | Method and apparatus for generating information | |
CN111899349A (en) | Model presentation method and device, electronic equipment and computer storage medium | |
CN114898447B (en) | Personalized fixation point detection method and device based on self-attention mechanism | |
CN111047632A (en) | Method and device for processing picture color of nail image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |