WO2021227916A1 - 面部形象生成方法、装置、电子设备及可读存储介质 - Google Patents
面部形象生成方法、装置、电子设备及可读存储介质 Download PDFInfo
- Publication number
- WO2021227916A1 WO2021227916A1 PCT/CN2021/091859 CN2021091859W WO2021227916A1 WO 2021227916 A1 WO2021227916 A1 WO 2021227916A1 CN 2021091859 W CN2021091859 W CN 2021091859W WO 2021227916 A1 WO2021227916 A1 WO 2021227916A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- facial
- target user
- target
- image
- expression
- Prior art date
Links
- 230000001815 facial effect Effects 0.000 title claims abstract description 193
- 238000000034 method Methods 0.000 title claims abstract description 61
- 230000008921 facial expression Effects 0.000 claims abstract description 83
- 238000004891 communication Methods 0.000 claims abstract description 8
- 230000014509 gene expression Effects 0.000 claims description 23
- 230000004044 response Effects 0.000 claims description 5
- 230000003190 augmentative effect Effects 0.000 claims description 3
- 230000006870 function Effects 0.000 description 14
- 230000008569 process Effects 0.000 description 11
- 230000003993 interaction Effects 0.000 description 10
- 230000002452 interceptive effect Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 230000002996 emotional effect Effects 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- KLDZYURQCUYZBL-UHFFFAOYSA-N 2-[3-[(2-hydroxyphenyl)methylideneamino]propyliminomethyl]phenol Chemical compound OC1=CC=CC=C1C=NCCCN=CC1=CC=CC=C1O KLDZYURQCUYZBL-UHFFFAOYSA-N 0.000 description 1
- 241000282472 Canis lupus familiaris Species 0.000 description 1
- 241000282326 Felis catus Species 0.000 description 1
- 241001465754 Metazoa Species 0.000 description 1
- 229910000831 Steel Inorganic materials 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 201000001098 delayed sleep phase syndrome Diseases 0.000 description 1
- 208000033921 delayed sleep phase type circadian rhythm sleep disease Diseases 0.000 description 1
- 230000000994 depressogenic effect Effects 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 239000004744 fabric Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 239000010959 steel Substances 0.000 description 1
- 239000004575 stone Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/90—Pitch determination of speech signals
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/63—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for estimating an emotional state
Definitions
- the embodiments of the application relate to the field of communications, and in particular to a method, device, electronic device, and readable storage medium for generating a facial image.
- AR Augmented Reality
- users need to wear an AR device, which will occlude the user's face, making it impossible to see the user's face, and it is difficult to distinguish each other's identities.
- the current processing method is to replace the user image through the pre-recorded avatar and the capture of user actions, but the preset avatar is relatively fixed, and the face only has simple speaking actions, unable to make expressions, resulting in the user's AR
- the interactive experience is poor.
- the purpose of the embodiments of the present application is to provide a facial image generation method, device, electronic device, and readable storage medium, which can solve the problem that users cannot display facial images in existing AR interactive scenes, resulting in poor AR interactive experience.
- an embodiment of the present application provides a method for generating a facial image, which is applied to an electronic device, and the method includes:
- the facial image of the target user is generated.
- an embodiment of the present application provides a facial image generation device, which is applied to an electronic device, and includes:
- the first acquiring module is used to acquire the facial contour of the target user wearing the AR device
- a determining module configured to determine a facial angle of the target user according to the facial contour, the facial angle including the angle between the facial orientation of the target user and the electronic device;
- the second acquisition module is used to acquire the target facial expression
- the generating module is used to generate the facial image of the target user according to the target facial expression and the facial angle.
- an embodiment of the present application provides an electronic device, including a processor, a memory, and a program or instruction stored on the memory and capable of running on the processor, and the program or instruction is executed by the processor. When executed, the steps of the facial image generation method as described in the first aspect are realized.
- an embodiment of the present application provides a readable storage medium on which a program or instruction is stored, and when the program or instruction is executed by a processor, the facial image generation method as described in the first aspect is implemented A step of.
- an embodiment of the present application provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run a program or an instruction to implement the chip as in the first aspect The method described.
- embodiments of the present application provide a computer software product, the computer software product is stored in a non-volatile storage medium, and the software product is configured to be executed by at least one processor to implement the computer software product as described in the first aspect.
- the steps of the facial image generation method are described in the first aspect.
- an embodiment of the present application provides a facial image generating device configured to execute the facial image generating method as described in the first aspect.
- the facial angle of the target user is determined according to the facial contour of the target user wearing the AR device, and the facial image of the target user is generated based on the facial angle of the target user and the acquired target facial expressions.
- the user's facial image can be displayed, and according to different target facial expressions, the user can express through different expressions to improve the AR interactive experience.
- FIG. 1 is a schematic flowchart of a facial image generation method provided by an embodiment of the application
- Figure 2a is one of the schematic diagrams of application scenarios provided by an embodiment of the application.
- Figure 2b is the second schematic diagram of the application scenario provided by the embodiment of the application.
- FIG. 3 is a schematic structural diagram of a facial image generation device provided by an embodiment of the application.
- FIG. 4 is a schematic structural diagram of an electronic device provided by an embodiment of the application.
- the AR interactive scene can be an AR-based conference, that is, participants wear AR devices to discuss and interact in the meeting, or the AR interactive scene can also be a remote role-playing game or a lecture to children in a remote place.
- Casual scenes such as stories.
- An embodiment of the present application provides a method for generating a facial image, and the method is applied to an electronic device.
- the above-mentioned electronic device is a device with a camera, based on the image obtained by the camera, combined with AR technology for AR interaction, the electronic device can be an AR device, such as AR glasses, etc., the electronic device can also be a mobile phone, a computer, etc., the embodiment of the present application
- the specific type of electronic equipment is not limited.
- the method includes:
- Step 101 Obtain the facial contour of the target user wearing the AR device
- the target user refers to an object that performs AR interaction.
- the target user may be a speaker wearing an AR device.
- the facial contour of the target user can be obtained through an AR device with a camera, or it can be obtained by a mobile phone, a computer and other devices with a camera, or it can be obtained through a separate setting.
- a camera captures the facial contours of the target user, and then transmits the captured images to the device for display.
- acquiring the facial contour of the target user wearing the AR device specifically includes:
- the closest point of the face refers to the point where the face of the target user is closest to the camera when the face image of the target user is acquired through the camera.
- the closest point of the face corresponds to the target The position of the tip of the user's nose.
- the aforementioned preset distance corresponds to a distance of one head width of the target user, and the distance of one head width can be set based on the average head width of a person.
- the facial contour of the target user can reflect the facial shape and the angle of the face.
- Step 102 Determine the face angle of the target user according to the facial contour
- the face angle includes the angle between the face orientation of the target user and the electronic device.
- the facial angle is used to reflect the facial posture of the target user's face relative to the electronic device. For example, according to the facial contour of the target user, it is determined that the target user's face is 30 degrees upwards relative to the direct counterpart of the electronic device. This can be reflected When the target user's facial posture is 30 degrees upside down. Further, for more complex facial angles, for example, the target user’s face is inclined obliquely, the deflection angle of the target user’s face relative to the electronic device in multiple directions (for example, horizontal and vertical) can be determined based on the current situation. Some spatial angle algorithms calculate the face angle of the target user.
- determining the facial angle of the target user according to the facial contour specifically includes: matching the facial angle corresponding to the facial contour in the facial contour library; wherein the facial contour library includes pre-recorded target users from multiple facial angles Facial contours.
- the facial contours of the target user can be recorded from multiple angles in advance, and a facial contour library is generated according to the recorded facial contours.
- a facial contour library different facial contours correspond to different facial angles. After obtaining the facial contour of the target user, the corresponding facial angle can be directly matched in the facial contour library.
- Step 103 Obtain the target facial expression
- the target facial expression refers to the facial expression used to generate the facial image of the target user.
- the target facial expression can be the expression of the target user, the expression of other users, or it can be obtained from the network. Some cartoon expressions.
- Step 104 Generate a facial image of the target user according to the target facial expression and facial angle.
- the selected target facial expressions are generated according to the determined facial angles to generate the facial image of the target user, so that the angle of the generated facial image can be consistent with the current facial angle of the target user, making the facial image closer The face of the target user.
- the facial angle of the target user is determined according to the facial contour of the target user wearing the AR device, and the facial image of the target user is generated based on the facial angle of the target user and the acquired target facial expressions.
- the user's facial image can be displayed, and according to different target facial expressions, the user can express through different expressions to improve the AR interactive experience.
- acquiring the target facial expression described above specifically includes: acquiring the facial expression of the target user recorded by the AR device;
- the AR device worn by the target user has an image acquisition function, for example: the AR device is AR glasses with a camera
- the AR device worn by the target user can directly perform facial expressions on the target user. Record and scan in real time, and use the recorded facial expressions as target facial expressions to generate the facial image of the target user.
- the above-mentioned acquiring target facial expressions specifically includes: acquiring input information of the target user, and according to the context of the input information, matching facial expressions corresponding to the context from the facial expression database; wherein, facial expressions
- the expression library includes a variety of facial expressions at multiple facial angles.
- various expressions of the target user from various angles can be recorded in advance, such as expressionless, angry, happy, depressed, Sad, serious, questioning, etc., set a corresponding label for each expression accordingly.
- the input information of the target user is obtained.
- the input information can be the text input by the target user or the voice input by the target user.
- the context recognition is performed based on the input information, and then the facial expression database is displayed in the facial expression database according to the context. Select the corresponding tags and match the corresponding facial expressions. For example, if the target user says "I am very happy today", it will be recognized according to the context. At this time, the context should correspond to the happy tag, and then select the band from the facial expression library. Facial expressions with happy tags.
- the facial expression library can also store a variety of facial expressions of some other objects at multiple facial angles, such as cats, dogs and other animals, superman, steel and other cartoon characters, which can be used in remote role-playing games. Or in casual scenes such as telling a story to a child in a different place, the facial image of the target user can be replaced with a cartoon image to improve the AR interaction effect.
- the method further includes at least one of the following:
- the target user’s mouth state is used to reflect the target user’s mouth actions, such as: opening mouth, closing mouth, pursing mouth, mouth tilted to one side, etc.
- the adjustment of the mouth state enables the target facial expressions to accurately and synchronously show the subtle expressions of the target user, making the facial expressions richer.
- the voice of the target user is called to reflect the emotional state of the target user, and can then be used to determine the magnitude of the target expression. For example, it can be divided into 1-10 degrees to correspond to the recorded expression from no expression to happy expression. In the 10 moments, as the pitch of the target user’s voice becomes higher, the emoji used is more backward. In this way, the expression amplitude of the target facial expression is adjusted through the voice tone, so that the target facial expression can accurately show the emotional state of the target user, and the facial expressions are richer.
- the target user wearing the AR device has multiple people, and facial image generation is required. Then these target users can bind the facial image generation function through preset gesture actions. For example, the AR device will randomly generate a set of gestures, such as stone, cloth, and scissors. Those who need to bind the facial image generation function need to do so. Raise your hands and make gestures in the specified order to confirm. This can make the facial image generation in the AR interaction process more flexible.
- the target user needs to temporarily handle other affairs.
- the target user can lock his current facial image, so as to avoid processing other affairs.
- the change of facial image affects other people. For example, in the scene of AR meeting, one of the participants needs to deal with other affairs temporarily, then he can lock his facial image to prevent facial image from happening when dealing with other affairs. The change made other participants mistakenly express his opinions.
- the method further includes: receiving a first input of the target user; in response to the first input, replacing the facial image of the target user with a preset image; the preset image includes at least one of the following : Cartoon emoticon image; an image corresponding to the input information of the target user.
- the image corresponding to the input information can be in the form of an icon.
- the target user can replace the facial image with a preset image.
- a preset image For example, referring to Figure 2a, replace the facial image with a cartoon emoticon image, such as emoji, a popular emoticon package on the Internet, etc.
- a cartoon emoticon image such as emoji, a popular emoticon package on the Internet, etc.
- a "check" or "fork” icon is used to replace the facial image.
- the execution subject may be a facial image generation device, or a control module in the facial image generation device for executing the method of loading the facial image generation.
- the facial image generation apparatus executes the method for generating a loaded facial image as an example to illustrate the facial image generation provided by the embodiment of the present application.
- an embodiment of the present application provides a facial image generating apparatus 300, which is applied to electronic equipment, and includes:
- the first acquiring module 301 is configured to acquire the facial contour of the target user wearing the AR device;
- the determining module 302 is configured to determine a facial angle of the target user according to the facial contour, where the facial angle includes the angle between the facial orientation of the target user and the electronic device;
- the second acquiring module 303 is used to acquire target facial expressions
- the generating module 304 is configured to generate the facial image of the target user according to the target facial expression and the facial angle.
- the first obtaining module 301 includes:
- the first matching unit is configured to match the facial angle corresponding to the facial contour in the facial contour library
- the facial contour library includes the facial contours of the target user recorded in advance from multiple facial angles.
- the second obtaining module 303 includes:
- the first acquiring unit is configured to acquire the facial expression of the target user recorded by the AR device
- the second acquiring unit is used to acquire the input information of the target user
- the second matching unit is configured to match facial expressions corresponding to the context from the facial expression database according to the context of the input information
- the facial expression library includes multiple facial expressions at multiple facial angles.
- the device 300 further includes:
- the third obtaining module is used to obtain the mouth state of the target user
- the first adjustment module is configured to adjust the mouth state of the target facial expression according to the mouth state of the target user
- the fourth acquiring module is used to acquire the voice pitch of the target user
- the first adjustment module is configured to adjust the expression amplitude of the target facial expression according to the voice pitch.
- the device 300 further includes:
- a receiving module for receiving the first input of the target user
- a replacement module configured to replace the facial image of the target user with a preset image in response to the first input
- the preset image includes at least one of the following:
- the facial angle of the target user is determined according to the facial contour of the target user wearing the AR device, and the facial image of the target user is generated based on the facial angle of the target user and the acquired target facial expressions.
- the user's facial image can be displayed, and according to different target facial expressions, the user can express through different expressions to improve the AR interactive experience.
- the facial image generation device in the embodiment of the present application may be a device, or a component, integrated circuit, or chip in a terminal.
- the device can be a mobile electronic device or a non-mobile electronic device.
- the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a handheld computer, a vehicle electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook, or a personal digital assistant (personal digital assistant).
- UMPC ultra-mobile personal computer
- netbook or a personal digital assistant (personal digital assistant).
- non-mobile electronic devices can be servers, network attached storage (NAS), personal computers (PC), televisions (television, TV), teller machines or self-service machines, etc., this application
- NAS network attached storage
- PC personal computers
- TV televisions
- teller machines or self-service machines etc.
- the facial image generation device in the embodiment of the present application may be a device with an operating system.
- the operating system may be an Android operating system, an ios operating system, or other possible operating systems, which are not specifically limited in the embodiment of the present application.
- the facial image generating device provided by the embodiment of the present application can implement the various processes implemented by the facial image generating device in the method embodiments of FIGS. 1 to 2b. In order to avoid repetition, details are not described here
- an embodiment of the present application further provides an electronic device, including a processor 410, a memory 409, a program or instruction that is stored on the memory 409 and can run on the processor 410, and the program or instruction is executed by the processor 410.
- an electronic device including a processor 410, a memory 409, a program or instruction that is stored on the memory 409 and can run on the processor 410, and the program or instruction is executed by the processor 410.
- the electronic devices in the embodiments of the present application include the above-mentioned mobile electronic devices and non-mobile electronic devices.
- FIG. 4 is a schematic diagram of the hardware structure of an electronic device that implements an embodiment of the present application.
- the electronic device 400 includes but is not limited to: a radio frequency unit 401, a network module 402, an audio output unit 403, an input unit 404, a sensor 405, a display unit 406, a user input unit 407, an interface unit 408, a memory 409, a processor 410, etc. part.
- the electronic device 400 may also include a power source (such as a battery) for supplying power to various components.
- the power source may be logically connected to the processor 410 through a power management system, so that the power management system can manage charging, discharging, and power management. Consumption management and other functions.
- the structure of the electronic device shown in FIG. 4 does not constitute a limitation on the electronic device.
- the electronic device may include more or less components than those shown in the figure, or some components may be combined, or different component arrangements, which will not be repeated here. .
- the input unit 404 may be a camera, which is used to obtain the facial contour of the target user wearing the AR device;
- the processor 410 is configured to determine a facial angle of the target user according to the facial contour, where the facial angle includes the angle between the facial orientation of the target user and the electronic device;
- the input unit 404 is also used to obtain target facial expressions
- the processor 410 is further configured to generate a facial image of the target user according to the target facial expression and the facial angle.
- the facial angle of the target user is determined according to the facial contour of the target user wearing the AR device, and the facial image of the target user is generated based on the facial angle of the target user and the acquired target facial expressions.
- the user's facial image can be displayed, and according to different target facial expressions, the user can express through different expressions to improve the AR interactive experience.
- the processor 410 is further configured to match a facial angle corresponding to the facial contour in a facial contour library;
- the facial contour library includes the facial contours of the target user recorded in advance from multiple facial angles.
- the input unit 404 is further configured to obtain the facial expression of the target user recorded by the AR device; or, obtain the input information of the target user, and obtain the input information from the facial expression according to the context of the input information. Match facial expressions corresponding to the context in the library;
- the facial expression library includes multiple facial expressions at multiple facial angles.
- the input unit 404 is further configured to obtain the mouth state of the target user, and adjust the mouth state of the target facial expression according to the mouth state of the target user;
- the input unit 404 is also used to obtain the voice pitch of the target user, and adjust the expression amplitude of the target facial expression according to the voice pitch.
- the user input unit 407 is configured to receive the first input of the target user
- the processor 410 is further configured to, in response to the first input, replace the facial image of the target user with a preset image; the preset image includes at least one of the following: a cartoon emoticon image; The image corresponding to the input information.
- the embodiment of the present application also provides a readable storage medium with a program or instruction stored on the readable storage medium.
- the program or instruction is executed by a processor, each process of the above-mentioned facial image generation method embodiment is realized, and can achieve The same technical effect, in order to avoid repetition, will not be repeated here.
- the processor is the processor in the electronic device described in the foregoing embodiment.
- the readable storage medium includes a computer readable storage medium, such as a computer read-only memory (Read-Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic disk, or optical disk.
- the embodiment of the present application further provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is used to run a program or an instruction to implement the above embodiment of the facial image generation method
- the chip includes a processor and a communication interface
- the communication interface is coupled to the processor
- the processor is used to run a program or an instruction to implement the above embodiment of the facial image generation method
- chips mentioned in the embodiments of the present application may also be referred to as system-level chips, system-on-chips, system-on-chips, or system-on-chips.
- the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of the embodiments.
- the functional units in the various embodiments of the present disclosure may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to make a terminal (which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) execute the methods described in the various embodiments of the present application.
- a terminal which may be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.
- the program can be stored in a computer readable storage medium. When executed, it may include the procedures of the above-mentioned method embodiments.
- the storage medium may be a magnetic disk, an optical disc, a read-only memory (Read-Only Memory, ROM), or a random access memory (Random Access Memory, RAM), etc.
- modules, units, and sub-units can be implemented in one or more application specific integrated circuits (ASIC), digital signal processors (Digital Signal Processor, DSP), and digital signal processing equipment (DSP Device, DSPD). ), programmable logic devices (Programmable Logic Device, PLD), Field-Programmable Gate Array (FPGA), general-purpose processors, controllers, microcontrollers, microprocessors, used to implement this disclosure Other electronic units or a combination of the functions described above.
- ASIC application specific integrated circuits
- DSP Digital Signal Processor
- DSP Device digital signal processing equipment
- DSPD digital signal processing equipment
- PLD programmable logic devices
- FPGA Field-Programmable Gate Array
- the technology described in the embodiments of the present disclosure can be implemented by modules (for example, procedures, functions, etc.) that perform the functions described in the embodiments of the present disclosure.
- the software codes can be stored in the memory and executed by the processor.
- the memory can be implemented in the processor or external to the processor.
Abstract
Description
Claims (15)
- 一种面部形象生成方法,应用于电子设备,包括:获取佩戴增强现实AR设备的目标用户的面部轮廓;根据所述面部轮廓确定所述目标用户的面部角度,所述面部角度包括所述目标用户的面部朝向与所述电子设备之间的角度;获取目标面部表情;根据所述目标面部表情和所述面部角度,生成所述目标用户的面部形象。
- 根据权利要求1所述的方法,其中,所述根据所述面部轮廓确定所述目标用户的面部角度,包括:在面部轮廓库中匹配与所述面部轮廓对应的面部角度;其中,所述面部轮廓库中包括预先从多个面部角度录制的所述目标用户的面部轮廓。
- 根据权利要求1所述的方法,其中,所述获取目标面部表情,包括:获取由所述AR设备录制的所述目标用户的面部表情;或者,获取所述目标用户的输入信息,根据所述输入信息的语境,从面部表情库中匹配与所述语境对应的面部表情;其中,所述面部表情库中包括在多个面部角度下的多种面部表情。
- 根据权利要求1所述的方法,其中,在根据所述目标面部表情和所述面部角度,生成所述目标用户的面部形象之前,所述方法还包括以下至少一项:获取所述目标用户的嘴部状态,并根据所述目标用户的嘴部状态调整所述目标面部表情的嘴部状态;获取所述目标用户的语音音调,并根据所述语音音调调整所述目标面部表情的表情幅度。
- 根据权利要求1所述的方法,还包括:接收所述目标用户的第一输入;响应于所述第一输入,将所述目标用户的面部形象替换为预设图像;所述预设图像包括以下至少一项:卡通表情图像;与所述目标用户输入信息对应的图像。
- 一种面部形象生成装置,应用于电子设备,包括:第一获取模块,用于获取佩戴增强现实AR设备的目标用户的面部轮廓;确定模块,用于根据所述面部轮廓确定所述目标用户的面部角度,所述面部角度包括所述目标用户的面部朝向与所述电子设备之间的角度;第二获取模块,用于获取目标面部表情;生成模块,用于根据所述目标面部表情和所述面部角度,生成所述目标用户的面部形象。
- 根据权利要求6所述的装置,其中,所述第一获取模块,包括:第一匹配单元,用于在面部轮廓库中匹配与所述面部轮廓对应的面部角度;其中,所述面部轮廓库中包括预先从多个面部角度录制的所述目标用户的面部轮廓。
- 根据权利要求6所述的装置,其中,所述第二获取模块,包括:第一获取单元,用于获取由所述AR设备录制的所述目标用户的面部表情;第二获取单元,用于获取所述目标用户的输入信息;第二匹配单元,用于根据所述输入信息的语境,从面部表情库中匹配与所述语境对应的面部表情;其中,所述面部表情库中包括在多个面部角度下的多种面部表情。
- 根据权利要求6所述的装置,还包括:第三获取模块,用于获取所述目标用户的嘴部状态;第一调整模块,用于根据所述目标用户的嘴部状态调整所述目标面部表情的嘴部状态;第四获取模块,用于获取所述目标用户的语音音调;第一调整模块,用于根据所述语音音调调整所述目标面部表情的表情幅度。
- 根据权利要求6所述的装置,还包括:接收模块,用于接收所述目标用户的第一输入;替换模块,用于响应于所述第一输入,将所述目标用户的面部形象替换为预设图像;所述预设图像包括以下至少一项:卡通表情图像;与所述目标用户的输入信息对应的图像。
- 一种电子设备,包括处理器、存储器及存储在所述存储器上并可在所述处理器上运行的程序或指令,所述程序或指令被所述处理器执行时实现如权利要求1至5中任一项所述的面部形象生成方法的步骤。
- 一种可读存储介质,所述可读存储介质上存储程序或指令,所述程序或指令被处理器执行时实现如权利要求1至5中任一项所述的面部形象生成方法的步骤。
- 一种芯片,所述芯片包括处理器和通信接口,所述通信接口和所述处理器耦合,所述处理器用于运行程序或指令,实现如权利要求1至5中任一项所述的面部形象生成方法。
- 一种计算机软件产品,所述计算机软件产品被存储在非易失的存储介质中,所述软件产品被配置成被至少一个处理器执行以实现如权利要求1至5中任一项所述的面部形象生成方法的步骤。
- 一种面部形象生成装置,所述面部形象生成装置被配置成用于执行如权利要求1至5中任一项所述的面部形象生成方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020227042575A KR20230006009A (ko) | 2020-05-09 | 2021-05-06 | 안면 영상 생성 방법, 장치, 전자기기 및 판독가능 저장 매체 |
EP21804806.4A EP4148677A4 (en) | 2020-05-09 | 2021-05-06 | METHOD AND DEVICE FOR GENERATING FACIAL IMAGE, ELECTRONIC DEVICE AND READABLE STORAGE MEDIUM |
JP2022567137A JP2023524119A (ja) | 2020-05-09 | 2021-05-06 | 顔イメージ生成方法、装置、電子機器及び可読記憶媒体 |
US17/984,212 US20230085099A1 (en) | 2020-05-09 | 2022-11-09 | Facial image generation method and apparatus, electronic device, and readable storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010386661.8A CN111583355B (zh) | 2020-05-09 | 2020-05-09 | 面部形象生成方法、装置、电子设备及可读存储介质 |
CN202010386661.8 | 2020-05-09 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/984,212 Continuation US20230085099A1 (en) | 2020-05-09 | 2022-11-09 | Facial image generation method and apparatus, electronic device, and readable storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021227916A1 true WO2021227916A1 (zh) | 2021-11-18 |
Family
ID=72124806
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/091859 WO2021227916A1 (zh) | 2020-05-09 | 2021-05-06 | 面部形象生成方法、装置、电子设备及可读存储介质 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20230085099A1 (zh) |
EP (1) | EP4148677A4 (zh) |
JP (1) | JP2023524119A (zh) |
KR (1) | KR20230006009A (zh) |
CN (1) | CN111583355B (zh) |
WO (1) | WO2021227916A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115359166A (zh) * | 2022-10-20 | 2022-11-18 | 北京百度网讯科技有限公司 | 一种图像生成方法、装置、电子设备和介质 |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111583355B (zh) * | 2020-05-09 | 2024-01-23 | 维沃移动通信有限公司 | 面部形象生成方法、装置、电子设备及可读存储介质 |
US20230103125A1 (en) * | 2021-09-30 | 2023-03-30 | Snap Inc. | 3d upper garment tracking |
CN114779948B (zh) * | 2022-06-20 | 2022-10-11 | 广东咏声动漫股份有限公司 | 基于面部识别的动画人物即时交互控制方法、装置及设备 |
CN115946150B (zh) * | 2022-12-13 | 2023-11-14 | 深圳无芯科技有限公司 | 具有面部表情的头部结构及具有其的机器人 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107491165A (zh) * | 2016-06-12 | 2017-12-19 | 张翔宇 | 一种vr眼镜面部3d图像、平面图像捕获与手势捕获系统 |
US20180052841A1 (en) * | 2016-08-16 | 2018-02-22 | Shanghai Zhangmen Science And Technology Co., Ltd. | Method and Device thereof for User Interaction Based on Virtual Objects and Non-volatile Storage Medium |
CN108614638A (zh) * | 2018-04-23 | 2018-10-02 | 太平洋未来科技(深圳)有限公司 | Ar成像方法和装置 |
CN109101953A (zh) * | 2018-09-07 | 2018-12-28 | 大连东锐软件有限公司 | 基于人类面部表情的分区要素化的表情动作生成方法 |
US20200090392A1 (en) * | 2018-09-19 | 2020-03-19 | XRSpace CO., LTD. | Method of Facial Expression Generation with Data Fusion |
CN111583355A (zh) * | 2020-05-09 | 2020-08-25 | 维沃移动通信有限公司 | 面部形象生成方法、装置、电子设备及可读存储介质 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8488023B2 (en) * | 2009-05-20 | 2013-07-16 | DigitalOptics Corporation Europe Limited | Identifying facial expressions in acquired digital images |
TW201039251A (en) * | 2009-04-30 | 2010-11-01 | Novatek Microelectronics Corp | Facial expression recognition apparatus and facial expression recognition method thereof |
KR102507567B1 (ko) * | 2015-06-09 | 2023-03-09 | 삼성전자주식회사 | 이미지를 처리하는 전자 장치 및 그 제어 방법 |
KR101810190B1 (ko) * | 2016-07-14 | 2017-12-18 | 김용상 | 얼굴 인식을 이용한 사용자 인증 방법 및 그 장치 |
CN108241434B (zh) * | 2018-01-03 | 2020-01-14 | Oppo广东移动通信有限公司 | 基于景深信息的人机交互方法、装置、介质及移动终端 |
US10706577B2 (en) * | 2018-03-06 | 2020-07-07 | Fotonation Limited | Facial features tracker with advanced training for natural rendering of human faces in real-time |
EP4262193A3 (en) * | 2018-05-07 | 2023-11-29 | Google LLC | Puppeteering remote avatar by facial expressions |
CN108875633B (zh) * | 2018-06-19 | 2022-02-08 | 北京旷视科技有限公司 | 表情检测与表情驱动方法、装置和系统及存储介质 |
CN109166164B (zh) * | 2018-07-25 | 2023-04-07 | 维沃移动通信有限公司 | 一种表情图片的生成方法及终端 |
-
2020
- 2020-05-09 CN CN202010386661.8A patent/CN111583355B/zh active Active
-
2021
- 2021-05-06 EP EP21804806.4A patent/EP4148677A4/en active Pending
- 2021-05-06 JP JP2022567137A patent/JP2023524119A/ja active Pending
- 2021-05-06 WO PCT/CN2021/091859 patent/WO2021227916A1/zh unknown
- 2021-05-06 KR KR1020227042575A patent/KR20230006009A/ko unknown
-
2022
- 2022-11-09 US US17/984,212 patent/US20230085099A1/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107491165A (zh) * | 2016-06-12 | 2017-12-19 | 张翔宇 | 一种vr眼镜面部3d图像、平面图像捕获与手势捕获系统 |
US20180052841A1 (en) * | 2016-08-16 | 2018-02-22 | Shanghai Zhangmen Science And Technology Co., Ltd. | Method and Device thereof for User Interaction Based on Virtual Objects and Non-volatile Storage Medium |
CN108614638A (zh) * | 2018-04-23 | 2018-10-02 | 太平洋未来科技(深圳)有限公司 | Ar成像方法和装置 |
CN109101953A (zh) * | 2018-09-07 | 2018-12-28 | 大连东锐软件有限公司 | 基于人类面部表情的分区要素化的表情动作生成方法 |
US20200090392A1 (en) * | 2018-09-19 | 2020-03-19 | XRSpace CO., LTD. | Method of Facial Expression Generation with Data Fusion |
CN111583355A (zh) * | 2020-05-09 | 2020-08-25 | 维沃移动通信有限公司 | 面部形象生成方法、装置、电子设备及可读存储介质 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4148677A4 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115359166A (zh) * | 2022-10-20 | 2022-11-18 | 北京百度网讯科技有限公司 | 一种图像生成方法、装置、电子设备和介质 |
Also Published As
Publication number | Publication date |
---|---|
KR20230006009A (ko) | 2023-01-10 |
JP2023524119A (ja) | 2023-06-08 |
CN111583355A (zh) | 2020-08-25 |
EP4148677A4 (en) | 2023-11-15 |
US20230085099A1 (en) | 2023-03-16 |
EP4148677A1 (en) | 2023-03-15 |
CN111583355B (zh) | 2024-01-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021227916A1 (zh) | 面部形象生成方法、装置、电子设备及可读存储介质 | |
TWI751161B (zh) | 終端設備、智慧型手機、基於臉部識別的認證方法和系統 | |
JP2019145108A (ja) | 顔に対応する3次元アバターを用いて顔の動きが反映された3dアバターを含むイメージを生成する電子装置 | |
KR20230156408A (ko) | 전자 미러링 디바이스를 동작시키는 핸즈-프리 모드를 활성화함 | |
US8044989B2 (en) | Mute function for video applications | |
JP7268071B2 (ja) | バーチャルアバターの生成方法及び生成装置 | |
CN110418095B (zh) | 虚拟场景的处理方法、装置、电子设备及存储介质 | |
WO2014094199A1 (en) | Facial movement based avatar animation | |
CN110555507B (zh) | 虚拟机器人的交互方法、装置、电子设备及存储介质 | |
EP3550812B1 (en) | Electronic device and method for delivering message by same | |
CN110794964A (zh) | 虚拟机器人的交互方法、装置、电子设备及存储介质 | |
US20210192192A1 (en) | Method and apparatus for recognizing facial expression | |
JP2020064426A (ja) | コミュニケーションシステム及びプログラム | |
US20230097257A1 (en) | Electronic communication interface with haptic feedback response | |
US20220206581A1 (en) | Communication interface with haptic feedback response | |
US11562548B2 (en) | True size eyewear in real time | |
CN112449098B (zh) | 一种拍摄方法、装置、终端及存储介质 | |
US20230368472A1 (en) | Dynamic augmented reality experience | |
US20220206582A1 (en) | Media content items with haptic feedback augmentations | |
US20230289919A1 (en) | Video stream refinement for dynamic scenes | |
KR20230160905A (ko) | 햅틱 및 오디오 피드백 응답을 갖는 인터페이스 | |
KR20230124081A (ko) | 햅틱 피드백 응답을 갖는 통신 인터페이스 | |
KR20230160918A (ko) | 햅틱 및 오디오 피드백 응답을 갖는 인터페이스 | |
CN113989463A (zh) | 一种会议显示方法、装置、系统和电子设备 | |
CN115861575A (zh) | 商品虚拟试用效果展示方法及电子设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21804806 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2022567137 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20227042575 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2021804806 Country of ref document: EP Effective date: 20221209 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |