CN109840948A - The put-on method and device of target object based on augmented reality - Google Patents

The put-on method and device of target object based on augmented reality Download PDF

Info

Publication number
CN109840948A
CN109840948A CN201711226533.1A CN201711226533A CN109840948A CN 109840948 A CN109840948 A CN 109840948A CN 201711226533 A CN201711226533 A CN 201711226533A CN 109840948 A CN109840948 A CN 109840948A
Authority
CN
China
Prior art keywords
target object
described image
threedimensional model
data
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201711226533.1A
Other languages
Chinese (zh)
Other versions
CN109840948B (en
Inventor
李炜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Inlife Handnet Co Ltd
Original Assignee
Inlife Handnet Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Inlife Handnet Co Ltd filed Critical Inlife Handnet Co Ltd
Priority to CN201711226533.1A priority Critical patent/CN109840948B/en
Publication of CN109840948A publication Critical patent/CN109840948A/en
Application granted granted Critical
Publication of CN109840948B publication Critical patent/CN109840948B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Processing Or Creating Images (AREA)

Abstract

The embodiment of the invention discloses the put-on methods and device of a kind of target object based on augmented reality.The put-on method and device of the target object are by carrying out three-dimensional reconstruction to target object, obtain the three-dimensional information of target object, then the normal direction of model any position is obtained by three-dimensional data model again, followed by the correlation data calculation direction of illumination of obtained normal data and described image, light source and target object are launched with corresponding position in the scene, to obtain the put-on method and device of a kind of target object of simple and effective augmented reality.

Description

The put-on method and device of target object based on augmented reality
Technical field
The present invention relates to the direction of illuminations of target object in technical field of virtual reality more particularly to a kind of augmented reality Determine method and apparatus.
Background technique
AR (Augmented Reality, augmented reality) is a kind of by real world information and virtual world information " nothing The integrated new technology of seam ", be script is difficult to experience in the certain time spatial dimension of real world entity information (depending on Feel information, sound, taste, tactile etc.) by science and technology such as computers, after analog simulation, true environment and virtual object Body has been added to the same picture in real time or space exists simultaneously, and virtual Information application to real world is felt by the mankind Official is perceived, to reach the sensory experience of exceeding reality.Augmented reality not only presents the information of real world, and And show virtual information simultaneously, two kinds of information are complementary to one another, are superimposed.In the augmented reality of visualization, Yong Huli With Helmet Mounted Display, real world with computer graphic is multiple is synthesized together, the true world can be seen around it. AR is a kind of new way that people pass through that computer carries out visualized operation to complex data and interacts, with traditional man-machine boundary The window-operating of face and prevalence is compared, and AR is haveed a qualitative leap in technical idea.
The simulation that problems faced is exactly illumination, traditional illuminant direction estimation method are frequently necessary in augmented reality device It is the direction based on image analysis illumination, normal direction of this method due to that cannot obtain object from image gives illumination side To analysis bring difficulty.Therefore, a kind of determination side of the direction of illumination of target object based on augmented reality is needed now Method.
Summary of the invention
The embodiment of the invention provides a kind of determination methods and dress based on the direction of illumination of target object in augmented reality It sets, can simply and effectively determine the direction of illumination of target object in augmented reality.
According to an aspect of the present invention, the present invention provides a kind of dispensing sides of target object based on augmented reality Method, comprising:
S101, the threedimensional model for obtaining the target object, obtain the phase of the threedimensional model according to the threedimensional model Close data;
S102, image is obtained, and described image is obtained according to described image and the related data of the threedimensional model Related data;
S103, the target object is determined by the related data of the threedimensional model and the related data of described image Direction of illumination;
S104, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, and according to the figure The related data of picture launches target object in the scene.
The present invention provides the put-on method and device of a kind of target object based on augmented reality, the target object Put-on method and device obtain the three-dimensional information of target object, then pass through three again by carrying out three-dimensional reconstruction to target object D Data Model obtains the normal direction of model any position, followed by the dependency number of obtained normal data and described image According to direction of illumination is calculated, light source and target object are launched with corresponding position in the scene, to obtain a kind of simply and effectively increasing The put-on method and device of the target object of strong reality.
Detailed description of the invention
To describe the technical solutions in the embodiments of the present invention more clearly, make required in being described below to embodiment Attached drawing is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the invention, for For those skilled in the art, without creative efforts, it can also be obtained according to these attached drawings other attached Figure.
Fig. 1 is the flow diagram provided in an embodiment of the present invention based on the put-on method of target object in augmented reality.
Fig. 2 is the application scenarios signal provided in an embodiment of the present invention based on the delivery device of target object in augmented reality Figure.
Fig. 3 is a kind of structural representation provided in an embodiment of the present invention based on the delivery device of target object in augmented reality Figure.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description.Obviously, described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.It is based on Embodiment in the present invention, those skilled in the art's every other implementation obtained without creative efforts Example, shall fall within the protection scope of the present invention.
Description and claims of this specification and term " first " in above-mentioned attached drawing, " second ", " third " etc. (if present) is to be used to distinguish similar objects, without being used to describe a particular order or precedence order.It should be appreciated that this The object of sample description is interchangeable under appropriate circumstances.In addition, term " includes " and " having " and their any deformation, meaning Figure, which is to cover, non-exclusive includes.
In patent document, the attached drawing that is discussed herein below and for describing each embodiment of principle disclosed by the invention only For illustrating, and should not be construed as limiting the scope of the present disclosure.Those skilled in the art will appreciate that original of the invention Reason can be implemented in any device suitably arranged.It will be explained in illustrative embodiments, these realities be shown in the attached drawings The example for applying mode.In addition, terminal accoding to exemplary embodiment will be described in detail with reference to the attached drawings.Identical attached drawing mark in attached drawing Number refer to identical element.
Term used in description of the invention is only used to describe particular implementation, and is not intended to show of the invention Concept.Unless have clearly different meanings in context, it is otherwise, used in the singular to express the table for covering plural form It reaches.In the description of the present invention, it should be appreciated that there are this hairs for the terms meant for illustration such as " comprising ", " having " and " containing " A possibility that feature for being disclosed in bright specification, number, step, movement or combinations thereof, and be not intended to exclude may be present or can A possibility that adding other one or more features, number, step, movement or combinations thereof.Same reference numerals in attached drawing refer to For same section.
The embodiment of the invention provides a kind of put-on method and device based on target object in augmented reality.It will divide below It is not described in detail.
In a preferred embodiment, a kind of put-on method and device based on target object in augmented reality, such as Fig. 1 are provided Shown, process can be such that
S101, the threedimensional model for obtaining the target object, obtain the phase of the threedimensional model according to the threedimensional model Close data.
Specifically, the object can be obtained using spatial digitizer, binocular solid and AUTODESK 123D technology The threedimensional model of body;The method for obtaining target object threedimensional model is all based on the three-dimensional reconstruction side of target object picture Method.
For example, three-dimensional reconstruction is carried out to target object using AUTODESK 123D technology, if need to only shoot to target object The three-dimensional reconstruction to target object can be realized in dry photo.
The normal of the target object surface point is obtained after the threedimensional model for obtaining target object, and institute is calculated State the first derivative of normal.
S102, image is obtained, and described image is obtained according to described image and the related data of the threedimensional model Related data.
In embodiments of the present invention, the step of related data of acquisition image and described image includes:
Image is obtained by camera;
The position orientation relation of the camera Yu the target object is obtained according to the flag information in described image;
According to the phase of position orientation relation and the target object threedimensional model between the camera and the target object It closes data to be projected, obtains the coordinate of described image corresponding with the threedimensional model.
This step is in order to which determination need to project the specific coordinate of target object in reality scene, in next step in reality Transmission goal object provides basis in scene.
S103, the target object is determined by the related data of the threedimensional model and the related data of described image Direction of illumination.
In embodiments of the present invention, institute is determined by the related data of the threedimensional model and the related data of described image The step of stating the direction of illumination of target object include:
Described image is filtered;
According to the coordinate of described image, the brightness data of described image is obtained from described image;
The target object is calculated using the brightness data of described image and the first derivative of the target object normal Direction of illumination.
Common illuminant direction estimation method is the analysis carried out to continuous image, and such words operand is huge.For Meet the needs of augmented reality target object device is to real-time, in the present invention, the method analyzed using discrete point, Reduce operand, is tested by majority, it is final to determine on the threedimensional model of target object in embodiments of the present invention uniformly 40~60 sampled points are chosen, and the sampling can directly be extracted from described image by the image coordinate of the sampled point The brightness value of point, and then the direction of illumination of the target object is calculated.
S104, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, and according to the figure The related data of picture launches target object in the scene.
In embodiments of the present invention, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, And the step of launching target object in the scene according to the related data of described image, includes:
According to the direction of illumination of the target object, light source is launched in corresponding position in the scene;
According to the position orientation relation of the camera and target object drop target object in the scene.
Wherein, the position orientation relation according to the camera and the target object launch in the scene target object it Before, the target object direction of illumination standard can also be checked to reach by detecting to the pre- image for launching target object Purpose whether really.
The specific steps in the check stage are as follows:
Obtain the pre- image for launching target object;
Obtain the pre- brightness data launched in target object image;
The pre- image for launching target object is subjected to brightness contrast with the image obtained by camera, if described Within the brightness value difference 5% of two images, then launched;If the brightness value of two image differs by more than 5%, then return S101 step.
From the foregoing, it will be observed that the present invention provides a kind of put-on method of target object based on augmented reality, the object The put-on method of body obtains the three-dimensional information of target object, then passes through three-dimensional again by carrying out three-dimensional reconstruction to target object Data model obtains the normal direction of model any position, followed by the related data of obtained normal data and described image Direction of illumination is calculated, light source and target object are launched with corresponding position in the scene, to obtain a kind of simply and effectively enhancing The put-on method and device of the target object of reality.
With reference to Fig. 2, further embodiment of this invention provides a kind of video capture device based on virtual reality.Such as figure, the view Frequency filming apparatus includes: photographic equipment 33, server 34 and control display equipment 36.
Wherein, photographic equipment 33 can be video camera, camera or the electronic equipment with camera function, can be used for Acquire image information.For example, the photographic equipment can be used for limb action, expression and language message of user etc..
Server 34 is specifically as follows the network equipments such as data server, network server.The server 34 can be used for mentioning For prop model 341 and the 3-D image of virtual scene 342.
Display control apparatus 36 may include the intelligence that computer, smart phone, tablet computer etc. have operation processing function It can equipment.
In another embodiment in embodiment, a kind of one kind of delivery device based on target object in strong reality is provided and is answered Use schematic diagram of a scenario.Below by the device based on above-mentioned target object, with the prop model 22 for target object physical model 22 For the target object put-on method is described in detail.
In the present embodiment, target object entity is exchanged for by scanning means 21 (such as spatial digitizer, binocular solid) 22 threedimensional model, and the related data of the threedimensional model according to the obtaining three-dimensional model.Then, pass through camera 23 Obtain the related data of 22 image of image and the target object of target object 22;Pass through the related data of the threedimensional model Go out the direction of illumination of the target object 22 with the correlation data calculation of 22 image of target object, finally target in the scene The relevant position of 22 direction of illumination of object places light source, and corresponding position dispensing image 25 in the scene.
As shown in figure 3, user 31 does the movement of weapon delivery to strange beast model 341, server 34 will be according to user's 31 Limb action generates corresponding light wave (i.e. virtual scene 342).Strange beast model 341 by the movement based on 31 weapon delivery of user, Show the state gone down.Photographic equipment 33 captures the information such as limb action, expression, the language of user 31, real-time Transmission It is shown to display control terminal 36, limb action of the server 34 based on user 31, in real time by virtual scene 342 and monster Beast model 341 is transmitted to display control terminal 36 and is shown, to obtain fused image 35.
In practical application, in order to promote the authenticity of scene, a fusing device can be increased in augmented reality device, For being rendered and being merged the target object image of dispensing with display scene.
From the foregoing, it will be observed that the put-on method provided in an embodiment of the present invention based on the target object in augmented reality, by right The target object that need to be launched establishes threedimensional model to obtain the related data of the threedimensional model, to improve target in augmented reality Effect is launched in the accuracy of object direction of illumination, enhancing.
In still another embodiment of the process, a kind of delivery device of target object based on augmented reality is also provided.Such as Fig. 3 Shown, being somebody's turn to do the video capture device based on virtual reality may include object module 31, scan module 32, Target Acquisition module 33, direction of illumination computing module 34 and putting module 35, in which:
Object module 31, for providing target object physical model;
Scan module 32, for scanning the object module and providing the three-dimensional data of the object module;
Target Acquisition module 33, for obtaining the image information data of object module;
Direction of illumination computing module 34, for being obtained according to the three-dimensional data of described image information data and the object module Take the light source direction of the target object.
Putting module 35, for placing light source according to the corresponding position of the light source direction of the target object in the scene, And target object is launched according to described image information data and the three-dimensional data in the scene
In some embodiments, the scan module 32 may include data establish unit 321, entity scanning element 322, Data processing unit 323, in which:
Data establish unit 321, for establishing the information database of the object module;
Entity scanning element 322, for obtaining the threedimensional model of object module and three bit scan model datas and being stored in In the information database;
Data processing unit 323, for calculated according to the threedimensional model threedimensional model normal and the method The first derivative of line.
In some embodiments, the Target Acquisition module 33 may include image acquisition unit 331, image processing unit 332 and coordinate calculating unit 333.
Image acquisition unit 331, for obtaining the image information of object module;
Image processing unit 332, for calculating described image acquiring unit and institute according to the flag information in described image State the position orientation relation of object module;
Coordinate calculating unit 333, for according to described image acquiring unit and the object module position orientation relation and institute The scan data for stating threedimensional model is projected, and processing obtains the corresponding image coordinate of threedimensional model.
In some embodiments, the direction calculating unit 34 may include that brightness data unit 341 and direction of illumination are true Order member 342.
Brightness data unit 341, for extracting the brightness data of described image from described image;
Direction of illumination determination unit 342, for reciprocal really according to the brightness data of described image and the single order of the normal The direction of illumination for the module that sets the goal.
In some embodiments, the putting module 35 includes integrated unit 351 and dispensing unit 352.
Integrated unit 351, for merging the image for passing through rendering with the reality scene;
Unit 352 is launched, launches described image and the light source for the corresponding position in display scene
From the foregoing, it will be observed that the embodiment of the present invention obtains a kind of delivery device of target object based on augmented reality, pass through Three-dimensional reconstruction is carried out to object module, the three-dimensional information of target object is obtained, model is then obtained by three-dimensional data model again The normal direction of any position, followed by the correlation data calculation direction of illumination of obtained normal data and described image, with Corresponding position launches light source and target object in the scene, thus obtain a kind of target object of simple and effective augmented reality Put-on method and device.
Term " one " and " described " and similar word have been used during describing idea of the invention (especially In the appended claims), it should be construed to not only cover odd number by these terms but also cover plural number.In addition, unless herein In be otherwise noted, otherwise herein narration numberical range when referred to merely by quick method and belong to the every of relevant range A independent value, and each independent value is incorporated into this specification, just as these values have individually carried out statement one herein Sample.In addition, unless otherwise stated herein or context has specific opposite prompt, otherwise institute described herein is methodical Step can be executed by any appropriate order.Change of the invention is not limited to the step of description sequence.Unless in addition Advocate, is otherwise all only using any and all example or exemplary language presented herein (for example, " such as ") Idea of the invention is better described, and not the range of idea of the invention limited.Spirit and model are not being departed from In the case where enclosing, those skilled in the art becomes readily apparent that a variety of modifications and adaptation.
It is provided for the embodiments of the invention a kind of put-on method and device based on target object in augmented reality above It is described in detail.It should be understood that illustrative embodiments as described herein should be to be considered only as descriptive, it is used to help manage Method and its core concept of the invention are solved, and are not intended to restrict the invention.To feature in each illustrative embodiments Or the description of aspect should usually be considered the similar features or aspects suitable for other exemplary embodiments.Although reference example Property embodiment describe the present invention, but can suggest that those skilled in the art carries out various change and change.The invention is intended to Cover these variations and the change in the scope of the appended claims.

Claims (10)

1. a kind of put-on method of the target object based on augmented reality characterized by comprising
S101, the threedimensional model for obtaining the target object, obtain the dependency number of the threedimensional model according to the threedimensional model According to;
S102, image is obtained, and obtains the correlation of described image according to described image and the related data of the threedimensional model Data;
S103, the illumination that the target object is determined by the related data of the threedimensional model and the related data of described image Direction;
S104, according to the direction of illumination of the target object, light source is launched in corresponding position in the scene, and according to described image Related data launches target object in the scene.
2. the put-on method of the target object according to claim 1 based on augmented reality, which is characterized in that described in acquisition The threedimensional model of target object, the step of obtaining the related data of the threedimensional model according to the threedimensional model include:
The target object is scanned, the threedimensional model of the target object is obtained;
The normal of the target object and the single order of the target object normal are calculated according to the threedimensional model of the target object Derivative.
3. the put-on method of the target object according to claim 2 based on augmented reality, which is characterized in that obtain figure Picture, and the step of obtaining the related data of described image according to described image and the related data of the threedimensional model includes:
Image is obtained by camera;
The position orientation relation of the camera Yu the target object is obtained according to the flag information in described image;
According to the dependency number of position orientation relation and the target object threedimensional model between the camera and the target object According to being projected, the coordinate of described image corresponding with the threedimensional model is obtained.
4. the put-on method of the target object according to claim 3 based on augmented reality, which is characterized in that by described The step of related data of threedimensional model and the related data of described image determine the direction of illumination of the target object include:
Described image is filtered;
According to the coordinate of described image, the brightness data of described image is obtained from described image;
The illumination of the target object is calculated using the brightness data of described image and the first derivative of the target object normal Direction.
5. the put-on method of the target object according to claim 4 based on augmented reality, which is characterized in that in the field Light source is placed in the corresponding position of the target object direction of illumination in scape, and in the scene according to the related data of described image Launch target object the step of include:
According to the direction of illumination of the target object, light source is launched in corresponding position in the scene;
According to the position orientation relation of the camera and target object drop target object in the scene.
6. a kind of delivery device of the target object based on augmented reality characterized by comprising
Object module, for providing the physical model for launching target;
Scan module, for scanning the object module and providing the three-dimensional data of the object module;
Target Acquisition module, for obtaining the image information data of object module;
Directions calculation module, for obtaining the target according to the three-dimensional data of described image information data and the object module The light source direction of object;
Putting module, for placing light source, Yi Jigen according to the corresponding position of the light source direction of the target object in the scene Target object is launched in the scene according to described image information data and the three-dimensional data.
7. a kind of delivery device of target object based on augmented reality according to claim 6, which is characterized in that described Scan module includes:
Data establish unit, for establishing the information database of the object module;
Entity scanning element, for obtaining the threedimensional model of object module and three bit scan model datas and being stored in the information In database;
Data processing unit, for calculating the normal of the threedimensional model and the single order of the normal according to the threedimensional model Derivative.
8. a kind of delivery device of target object based on augmented reality according to claim 7, which is characterized in that target Obtaining module includes:
Image acquisition unit, for obtaining the image information of object module;
Image processing unit, for calculating described image acquiring unit and the target mould according to the flag information in described image The position orientation relation of block;
Coordinate calculating unit, for according to the position orientation relation of described image acquiring unit and the object module and the three-dimensional mould The scan data of type is projected, and processing obtains the corresponding image coordinate of threedimensional model.
9. the delivery device of the target object based on augmented reality according to claim 8, which is characterized in that the telegoniometer Calculating unit includes:
Brightness data unit, for extracting the brightness data of described image from described image;
Direction of illumination determination unit, for determining target mould according to the brightness data of described image and the single order inverse of the normal The direction of illumination of block.
10. the delivery device of 6 a kind of target object based on augmented reality according to claim, which is characterized in that institute Stating putting module includes:
Integrated unit, for merging the image for passing through rendering with the reality scene;
Unit is launched, launches described image and the light source for the corresponding position in display scene.
CN201711226533.1A 2017-11-29 2017-11-29 Target object throwing method and device based on augmented reality Active CN109840948B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711226533.1A CN109840948B (en) 2017-11-29 2017-11-29 Target object throwing method and device based on augmented reality

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711226533.1A CN109840948B (en) 2017-11-29 2017-11-29 Target object throwing method and device based on augmented reality

Publications (2)

Publication Number Publication Date
CN109840948A true CN109840948A (en) 2019-06-04
CN109840948B CN109840948B (en) 2023-08-15

Family

ID=66882043

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711226533.1A Active CN109840948B (en) 2017-11-29 2017-11-29 Target object throwing method and device based on augmented reality

Country Status (1)

Country Link
CN (1) CN109840948B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112991556A (en) * 2021-05-12 2021-06-18 航天宏图信息技术股份有限公司 AR data display method and device, electronic equipment and storage medium
CN113763090A (en) * 2020-11-06 2021-12-07 北京沃东天骏信息技术有限公司 Information processing method and device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100962557B1 (en) * 2009-02-05 2010-06-11 한국과학기술원 Augmented reality implementation apparatus and method of the same
CN103218854A (en) * 2013-04-01 2013-07-24 成都理想境界科技有限公司 Method for realizing component marking during augmented reality process and augmented reality system
CN106981087A (en) * 2017-04-05 2017-07-25 杭州乐见科技有限公司 Lighting effect rendering intent and device
CN107071388A (en) * 2016-12-26 2017-08-18 深圳增强现实技术有限公司 A kind of three-dimensional augmented reality display methods and device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100962557B1 (en) * 2009-02-05 2010-06-11 한국과학기술원 Augmented reality implementation apparatus and method of the same
CN103218854A (en) * 2013-04-01 2013-07-24 成都理想境界科技有限公司 Method for realizing component marking during augmented reality process and augmented reality system
CN107071388A (en) * 2016-12-26 2017-08-18 深圳增强现实技术有限公司 A kind of three-dimensional augmented reality display methods and device
CN106981087A (en) * 2017-04-05 2017-07-25 杭州乐见科技有限公司 Lighting effect rendering intent and device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
杜书侠等: "增强现实虚拟人中的光照方向估计", 《光学技术》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113763090A (en) * 2020-11-06 2021-12-07 北京沃东天骏信息技术有限公司 Information processing method and device
CN113763090B (en) * 2020-11-06 2024-05-21 北京沃东天骏信息技术有限公司 Information processing method and device
CN112991556A (en) * 2021-05-12 2021-06-18 航天宏图信息技术股份有限公司 AR data display method and device, electronic equipment and storage medium
CN112991556B (en) * 2021-05-12 2022-05-27 航天宏图信息技术股份有限公司 AR data display method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN109840948B (en) 2023-08-15

Similar Documents

Publication Publication Date Title
Oufqir et al. ARKit and ARCore in serve to augmented reality
CN108447043B (en) Image synthesis method, equipment and computer readable medium
CN107066082B (en) Display methods and device
RU2586566C1 (en) Method of displaying object
CN110866977B (en) Augmented reality processing method, device, system, storage medium and electronic equipment
US20150279044A1 (en) Method and apparatus for obtaining 3d face model using portable camera
CN106325509A (en) Three-dimensional gesture recognition method and system
CN109815776B (en) Action prompting method and device, storage medium and electronic device
CN106355153A (en) Virtual object display method, device and system based on augmented reality
CN113822977A (en) Image rendering method, device, equipment and storage medium
KR101885090B1 (en) Image processing apparatus, apparatus and method for lighting processing
CN110458924B (en) Three-dimensional face model establishing method and device and electronic equipment
CN110147737B (en) Method, apparatus, device and storage medium for generating video
CN110059624B (en) Method and apparatus for detecting living body
US20220358662A1 (en) Image generation method and device
CN111833457A (en) Image processing method, apparatus and storage medium
Bergig et al. In-place augmented reality
EP3141985A1 (en) A gazed virtual object identification module, a system for implementing gaze translucency, and a related method
CN109840948A (en) The put-on method and device of target object based on augmented reality
KR20140126529A (en) Physical Movement of Object on Reality-Augmented Reality Interaction System and Implementation Method for Electronic book
CN114067085A (en) Virtual object display method and device, electronic equipment and storage medium
CN109816791B (en) Method and apparatus for generating information
CN111899349A (en) Model presentation method and device, electronic equipment and computer storage medium
CN114898447B (en) Personalized fixation point detection method and device based on self-attention mechanism
CN111047632A (en) Method and device for processing picture color of nail image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant