CN111435546A - Model action method and device, sound box with screen, electronic equipment and storage medium - Google Patents

Model action method and device, sound box with screen, electronic equipment and storage medium Download PDF

Info

Publication number
CN111435546A
CN111435546A CN201910037303.3A CN201910037303A CN111435546A CN 111435546 A CN111435546 A CN 111435546A CN 201910037303 A CN201910037303 A CN 201910037303A CN 111435546 A CN111435546 A CN 111435546A
Authority
CN
China
Prior art keywords
face
human face
action
image
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910037303.3A
Other languages
Chinese (zh)
Inventor
不公告发明人
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing ByteDance Network Technology Co Ltd
Original Assignee
Beijing ByteDance Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing ByteDance Network Technology Co Ltd filed Critical Beijing ByteDance Network Technology Co Ltd
Priority to CN201910037303.3A priority Critical patent/CN111435546A/en
Priority to PCT/CN2020/070375 priority patent/WO2020147598A1/en
Publication of CN111435546A publication Critical patent/CN111435546A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • G06T13/403D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The disclosure discloses a model action method, a model action device, a sound box with a screen, electronic equipment and a storage medium. The method comprises the following steps: acquiring two or more continuous human face images; determining corresponding human face action characteristic parameters according to human face changes in the human face images; and generating a model action instruction according to the human face action characteristic parameters, so that a preset 3D image executes corresponding human face action according to the human face action characteristic parameters. The embodiment of the disclosure solves the problems that the current model action method in the two-dimensional space cannot simulate the real figure image and the model action effect is poor, realizes the simulation of the human face action through the preset 3D image, improves the simulation effect of the human face action, enhances the authenticity of the human face action simulation, and improves the interaction experience.

Description

Model action method and device, sound box with screen, electronic equipment and storage medium
Technical Field
The embodiment of the disclosure relates to an image processing technology, and in particular, to a model action method and apparatus, a sound box with a screen, an electronic device, and a storage medium.
Background
With the development of electronic products, the user can meet the use requirements of the user by carrying out good interaction with the electronic products, and the experience of the user in using the electronic products is improved.
Currently, the interaction method of human and machine stays in two-dimensional space, for example, voice is used for interaction; or in the screen interaction, the interaction is realized by detecting the moving distance, the moving speed and the moving direction of the limb moving in the plane. The interaction in the two-dimensional space can not simulate the real figure image, and the interaction effect is poor.
Disclosure of Invention
The invention provides a model action method, a model action device, a sound box with a screen, electronic equipment and a storage medium, and aims to solve the problems that the current model action method in a two-dimensional space cannot simulate a real character image and the model action effect is poor.
In a first aspect, an embodiment of the present disclosure provides a model action method, including:
acquiring two or more continuous human face images;
determining corresponding human face action characteristic parameters according to human face changes in the human face images;
and generating a model action instruction according to the human face action characteristic parameters, so that a preset 3D image executes corresponding human face action according to the human face action characteristic parameters.
In the foregoing solution, optionally, before generating a model action instruction according to the face action feature parameter, so that a preset 3D image executes a corresponding face action according to the face action feature parameter, the method further includes:
acquiring a face image of at least one preset angle, and extracting face feature data of the face image;
constructing a human face 3D model corresponding to the human face image according to the human face feature data;
and applying the human face 3D model to a preset 3D image to obtain the preset 3D image with the human face characteristic data.
The human face action characteristic parameters at least comprise one of the following parameters: speed of movement, direction of movement, and distance of movement.
In the foregoing scheme, optionally, the determining, according to the face change in the face image, a corresponding face action characteristic parameter includes:
and determining the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face image.
In the foregoing solution, optionally, the generating a model action instruction according to the face action feature parameter to enable a preset 3D image to execute a corresponding face action according to the face action feature parameter includes:
and generating a model action instruction according to the moving speed, the moving direction and the moving distance of the human face, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face.
Wherein, predetermine the 3D image and be: a cartoon 3D character, a professional 3D character, or a gender 3D character.
In a second aspect, an embodiment of the present disclosure further provides a model action apparatus, including:
the face image acquisition module is used for acquiring two or more continuous face images;
the human face action characteristic parameter determining module is used for determining corresponding human face action characteristic parameters according to human face changes in the human face image;
and the preset 3D image action execution module is used for generating a model action instruction according to the human face action characteristic parameters so that the preset 3D image executes corresponding human face actions according to the human face action characteristic parameters.
In the foregoing solution, optionally, the apparatus further includes: the preset 3D image acquisition module is used for acquiring a face image of at least one preset angle before generating a model action instruction according to the face action characteristic parameters to enable the preset 3D image to execute corresponding face action according to the face action characteristic parameters, and extracting face characteristic data of the face image; constructing a human face 3D model corresponding to the human face image according to the human face feature data; and applying the human face 3D model to a preset 3D image to obtain the preset 3D image with the human face characteristic data.
The human face action characteristic parameters at least comprise one of the following parameters: speed of movement, direction of movement, and distance of movement.
In the foregoing scheme, optionally, the face motion characteristic parameter determining module is specifically configured to:
and determining the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face image.
In the foregoing scheme, optionally, the preset 3D image action execution module is specifically configured to:
and generating a model action instruction according to the moving speed, the moving direction and the moving distance of the human face, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face.
Wherein, predetermine the 3D image and be: a cartoon 3D character, a professional 3D character, or a gender 3D character.
In a third aspect, an embodiment of the present disclosure further provides a sound box with a screen, including a main body, a controller located in the main body, and at least two cameras located on the main body; the distance between the at least two cameras is larger than a distance threshold value, and the model action device provided by any embodiment of the disclosure is arranged in the controller.
In a fourth aspect, an embodiment of the present disclosure further provides an electronic device, where the electronic device includes:
one or more processors;
a memory for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement a model action method as described in any of the embodiments of the present disclosure.
In a fifth aspect, the embodiments of the present disclosure further provide a computer-readable storage medium, on which a computer program is stored, and the computer program, when executed by a processor, implements the model action method according to any one of the embodiments of the present disclosure.
The model action method provided by the embodiment of the disclosure obtains two or more continuous human face images; determining corresponding human face action characteristic parameters according to human face changes in the human face images; and generating a model action instruction according to the face action characteristic parameters, so that the preset 3D image executes corresponding face action according to the face action characteristic parameters, the simulation of face action through the preset 3D image can be realized, the face action simulation effect is improved, the authenticity of face action simulation is enhanced, and the interaction experience is improved.
Drawings
FIG. 1 is a flowchart of a method for model actions according to an embodiment of the present disclosure;
FIG. 2 is a flowchart of a model action method provided in the second embodiment of the disclosure;
FIG. 3 is a flowchart of a method for model actions provided by a third embodiment of the present disclosure;
fig. 4 is a schematic structural diagram of a model actuating device according to a fourth embodiment of the present disclosure;
fig. 5 is a schematic structural diagram of a sound box with a screen according to a fifth embodiment of the present disclosure;
fig. 6 is a schematic structural diagram of an action interface control of a speaker model with a screen according to a fifth embodiment of the present disclosure;
fig. 7 is a schematic structural diagram of an electronic device according to a sixth embodiment of the present disclosure.
Detailed Description
The present disclosure is described in further detail below with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the disclosure and are not limiting of the disclosure. It should be further noted that, for the convenience of description, only some of the structures relevant to the present disclosure are shown in the drawings, not all of them.
In the following embodiments, optional features and examples are provided in each embodiment, and various features described in the embodiments may be combined to form a plurality of alternatives, and each numbered embodiment should not be regarded as only one technical solution.
Example one
Fig. 1 is a flowchart of a model action method according to an embodiment of the present disclosure, where the present embodiment is applicable to a case of human face action interaction, and the method may be executed by a model action apparatus, and the apparatus may be composed of hardware and/or software, and may be generally integrated in an electronic device such as a mobile phone, a tablet, and a computer. The method specifically comprises the following steps:
and S110, acquiring two or more continuous face images.
Two or more continuous human face images are obtained through at least one camera, and the time interval between every two images is preset. Optionally, a face image may be obtained by the camera at preset time intervals, face video data may be recorded by the camera, and the face image may be captured from the face video data according to the preset time intervals.
And S120, determining corresponding human face action characteristic parameters according to human face changes in the human face images.
The change of the face in the face image can be detected by two or more face images, and specifically, when detecting a face change in the face images, the face change is determined by comparing pixel changes at corresponding positions in the two face images, and, illustratively, the face change is determined by the face change of a fixed part in the face, such as the eyes or the chin, taking eyes as an example, the change of the eyes in the human face in the two images is used for determining the change of the human face, the change of the eyes in the two images can have position change, the moving speed of the eyes can be confirmed according to the interval time of the two images, and through the position change of the eyes in the plurality of images, the movement track of the eyes can be obtained, and the movement speed information can be obtained by combining the movement time information, so that the human face action characteristic parameters can be determined.
Optionally, the face motion feature parameters at least include one of the following parameters: speed of movement, direction of movement, and distance of movement. The human face can move up and down, left and right or rotate and shake and the like. The face motion characteristic parameter corresponding to the face motion is at least one of a moving speed, a moving direction and a moving distance, and optionally, the face motion characteristic parameter includes the moving speed, the moving direction and the moving distance. The human face action can be restored through the human face action characteristic parameters, and the effect of interaction with the user is achieved.
And S130, generating a model action instruction according to the human face action characteristic parameters, and enabling a preset 3D image to execute corresponding human face actions according to the human face action characteristic parameters.
Wherein, predetermine the 3D image and be: a cartoon 3D character, a professional 3D character, or a gender 3D character. The cartoon 3D image may be an animal image such as a kitten, a puppy, or a monkey, or an animated image such as a piggy cookie or a cherry ball. The professional 3D figure may be, for example, a doctor, teacher, fireman or police figure. The gender 3D image can be, for example, a man or a woman, and age information can be further incorporated into the gender 3D image to specifically set the images as a boy, a girl, an adult man, an adult woman, an elderly man or an elderly woman.
In the foregoing scheme, optionally, the determining, according to the face change in the face image, a corresponding face action feature parameter includes: and determining the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face image. The predetermined region may be, for example, eyes or chin, or may be two cheeks, and the specific region is not specifically limited herein.
In the foregoing solution, optionally, the generating a model action instruction according to the face action feature parameter to enable a preset 3D image to execute a corresponding face action according to the face action feature parameter includes: and generating a model action instruction according to the moving speed, the moving direction and the moving distance of the human face, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face. The model action instruction comprises the moving speed, the moving direction and the moving distance of the human face, the preset 3D image simulates human face action according to the model action instruction, and human face action interaction with a user is achieved through the preset 3D image. The model action command comprises a face moving speed, a moving direction and a moving distance of a preset 3D image, the face moving speed, the moving direction and the moving distance of the preset 3D image can be the same as the moving speed, the moving direction and the moving distance of a face of a user, and can also be set according to a preset rule, for example, the moving speed of the face of the user is v, and the face moving speed of the preset 3D image is 2 v; the moving direction of the face of the user is left and right, and the moving direction of the face of the preset 3D image is left and right; the moving distance of the face of the user is D, and the moving distance of the face of the preset 3D image is 2D. The preset rule may be, for example, that the moving speed of the face of the user is v, and the moving speed of the face of the 3D avatar is v/2; the moving direction of the face of the user is left and right, and the moving direction of the face of the preset 3D image is left and right; the moving distance of the face of the user is D, and the moving distance of the face of the preset 3D image is D/2. The preset rules can be set at will, and the interaction interest is improved.
The model action method provided by the embodiment of the disclosure obtains two or more continuous human face images; determining corresponding human face action characteristic parameters according to human face changes in the human face images; and generating a model action instruction according to the face action characteristic parameters, so that the preset 3D image executes corresponding face action according to the face action characteristic parameters, the simulation of face action through the preset 3D image can be realized, the face action simulation effect is improved, the authenticity of face action simulation is enhanced, and the interaction experience is improved.
Example two
Fig. 2 is a schematic flow chart of a model action method according to a second embodiment of the present disclosure. The present embodiment is embodied on the basis of various alternatives in the above-described embodiments. The method specifically comprises the following steps:
s210, two or more continuous face images are obtained.
And S220, determining corresponding human face action characteristic parameters according to human face changes in the human face images.
And S230, obtaining a face image of at least one preset angle, and extracting face feature data of the face image.
The human face image of at least one preset angle is obtained through at least two cameras, the distance between the at least two cameras is larger than the distance threshold value, the human face image can be conveniently shot in multiple directions through multiple angles, the angle information obtained by the human face image is enriched, and then the human face feature data of the human face image is enriched, so that the effect of improving the precision of the human face 3D model building is achieved. And extracting the face feature data of the face image by a face feature extraction algorithm. The face feature data mainly represents the data of eyes, eyebrows, nose, mouth, ears and face contour, and the face can be uniquely represented by the face feature data.
S240, constructing a human face 3D model corresponding to the human face image according to the human face feature data.
And constructing a face 3D model corresponding to the face image according to the acquired face feature data, wherein the adopted 3D model can be a general face model or a three-dimensional deformation model when the face 3D model is constructed.
And S250, applying the human face 3D model to a preset 3D image to obtain the preset 3D image with the human face characteristic data.
The constructed human face 3D model is applied to the preset 3D image, the preset 3D image has the same human face characteristic data as the user, the purpose of simulating the user's growth is achieved, role playing of the user can be achieved through the preset 3D image, if the preset 3D image is a doctor, the doctor with the human face data of the user can be formed after the human face 3D model of the user is applied to the preset 3D image, the doctor with the same growth as the user can be obtained, the user playing of the doctor is achieved, and entertainment is improved.
And S260, generating a model action instruction according to the human face action characteristic parameters, so that the preset 3D image executes corresponding human face action according to the human face action characteristic parameters.
The preset 3D image is the 3D image with the face feature data of the user, the preset 3D image generates a model action instruction according to the face action feature parameters, corresponding face action is executed according to the model action instruction, the face action interaction between the user and the preset 3D image with the same face feature data of the user can be realized, the interestingness is improved, and the interaction experience is enhanced.
The technical scheme of the embodiment extracts face characteristic data of the face image by acquiring the face image of at least one preset angle; constructing a human face 3D model corresponding to the human face image according to the human face feature data; the human face 3D model is applied to the preset 3D image, the preset 3D image with the human face feature data is obtained, human face action interaction with the virtual image with the human face features of the user can be achieved, entertainment is improved, and interaction experience is enhanced.
It should be noted that in this embodiment, the execution sequence of steps S210 and S220 and steps S230, S240, and S250 is not limited, and may be executed according to the sequence of this embodiment, or may be executed first in steps S230, S240, and S250, and then in steps S210 and S220, or may be executed in steps S210 and S220 in synchronization with steps S230, S240, and S250.
EXAMPLE III
Fig. 3 is a schematic flow chart of a model action method according to a third embodiment of the present disclosure. The present embodiment is embodied on the basis of various alternatives in the above-described embodiments. The method specifically comprises the following steps:
s310, two or more continuous face images are obtained.
And S320, determining the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face image.
The preset part in the face image at least comprises: eyes, mouth, nose, eyebrows, chin, forehead and cheeks. In this embodiment, taking a preset portion in a face image as a mouth as an example, a position change of the mouth in the face image is detected. In the obtained multiple face images, the moving direction and the moving distance of the mouth can be determined by detecting the position of the mouth in the face images, and the time period corresponding to the movement of the position of the mouth can be determined by the time information of the face images, so that the moving speed of the mouth can be determined. The moving speed, moving direction and moving distance of the mouth are the moving speed, moving direction and moving distance of the face.
And S330, generating a model action instruction according to the moving speed, the moving direction and the moving distance of the human face, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face.
Generating a model action instruction according to the face movement speed, the movement direction and the movement distance of the user, wherein the model action instruction comprises the face movement speed, the movement direction and the movement distance of the preset 3D image, the face movement speed, the movement direction and the movement distance of the preset 3D image can be the same as or different from the face movement speed, the movement direction and the movement distance of the user, and can be set according to a preset rule, and the preset rule can refer to the disclosed embodiment.
The technical scheme of the embodiment obtains two or more continuous face images, determines the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face images, and improves the accuracy of determining the moving speed, the moving direction and the moving distance of the face; according to the moving speed, the moving direction and the moving distance of the human face, a model action instruction is generated, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face, human face action simulation effect is improved by simulating human face action through the preset 3D image, the authenticity of human face action simulation is enhanced, and interactive experience is improved.
Example four
Fig. 4 is a schematic structural diagram of a model actuating device according to a fourth embodiment of the present disclosure. Referring to fig. 4, the model action means includes: a face image obtaining module 410, a face motion characteristic parameter determining module 420, and a preset 3D image motion executing module 430, which are described in detail below.
A face image obtaining module 410, configured to obtain two or more consecutive face images;
the face motion characteristic parameter determining module 420 is configured to determine a corresponding face motion characteristic parameter according to a face change in the face image;
and the preset 3D image action execution module 430 is configured to generate a model action instruction according to the face action feature parameter, so that the preset 3D image executes a corresponding face action according to the face action feature parameter.
The model action device that this embodiment provided can realize improving the effect of people's face action simulation through predetermineeing 3D image simulation people's face action, strengthens the authenticity of people's face action simulation, improves interactive experience.
In the foregoing solution, optionally, the apparatus further includes: the preset 3D image acquisition module is used for acquiring a face image of at least one preset angle before generating a model action instruction according to the face action characteristic parameters to enable the preset 3D image to execute corresponding face action according to the face action characteristic parameters, and extracting face characteristic data of the face image; constructing a human face 3D model corresponding to the human face image according to the human face feature data; and applying the human face 3D model to a preset 3D image to obtain the preset 3D image with the human face characteristic data.
The human face action characteristic parameters at least comprise one of the following parameters: speed of movement, direction of movement, and distance of movement.
In the foregoing scheme, optionally, the face motion characteristic parameter determining module is specifically configured to:
and determining the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face image.
In the foregoing scheme, optionally, the preset 3D image action execution module is specifically configured to:
and generating a model action instruction according to the moving speed, the moving direction and the moving distance of the human face, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face.
Wherein, predetermine the 3D image and be: a cartoon 3D character, a professional 3D character, or a gender 3D character.
The model action device provided by the disclosure can execute the model action method provided by any embodiment of the disclosure, and has corresponding functional modules and beneficial effects for executing the model action method.
EXAMPLE five
Fig. 5 is a schematic structural diagram of a sound box with a screen according to a fifth embodiment of the present disclosure. Referring to fig. 5, the sound box with a screen includes: comprises a main body 51, a controller 52 positioned in the main body 51 and at least two cameras 53 positioned on the main body 51; the distance between the at least two cameras 53 is greater than a distance threshold setting, and any model action device provided by the embodiment of the disclosure is arranged in the controller 52.
Distance between two at least cameras is greater than apart from the threshold value setting, use two cameras as an example, a camera can be placed and take the upper portion position of screen audio amplifier main part, another camera can prevent to take the lower part position of screen audio amplifier main part, the distance is greater than apart from the threshold value setting, can make things convenient for the multi-angle diversified facial image of shooing, richen the angle information that facial image acquireed, and then richen facial image's facial feature data, in order to reach the effect that improves the precision that people's face 3D model found.
Fig. 6 is a schematic structural diagram of a model action interface control of a sound box with a screen according to an embodiment of the present disclosure, after a user starts a model action interface, the user selects a preset image type of a 3D image by setting a menu control 1, and presets rules of a face moving speed, a moving direction, and a moving distance of the 3D image and a face moving speed, a moving direction, and a moving distance of the user. In the preset 3D character type setting, a quick setting can be performed through the control 2. The method comprises the steps that firstly, a control 3 is clicked for multiple times, a camera obtains face image data of a user, face characteristic data are extracted, a face 3D model is constructed, the face 3D model is applied to a preset 3D image, the preset 3D image has the face characteristic data of the user, then the control 3 is pressed for a long time, the camera captures face actions of the user, the preset 3D image executes the face actions according to face moving speed, moving direction and moving distance of the user, and interaction with the user through the face actions is achieved.
For example, the sound box with the screen may be applied to a reading scene, and image data of a reading material is acquired through a camera of the sound box with the screen, where the reading material is, for example, a book, and the reading material may be placed on a desktop, for example. The image data of the point reading material is analyzed to obtain image text contents, the text contents are converted into voice data, and point reading is achieved through a loudspeaker of the sound box. The screen in the audio amplifier of area screen can show and predetermine the 3D image, click to read through predetermineeing the 3D image, realizes the interest of click to read, predetermine the 3D image for example and be the teacher, click to read through the teacher, realizes the simulation to real teaching, improves the study enjoyment, and then improves learning efficiency.
The user can also realize the human face action interaction with the preset 3D image by adopting the model action method in the disclosed embodiment through the preset 3D image displayed in the sound box with the screen, so that the interaction experience is improved.
The area screen audio amplifier that this embodiment provided can realize improving the effect of people's face action simulation through predetermineeing 3D image simulation people's face action, strengthens the authenticity of people's face action simulation, improves mutual experience.
EXAMPLE six
Referring now to FIG. 7, a block diagram of an electronic device 600 suitable for use in implementing embodiments of the present disclosure is shown. The electronic devices in the embodiments of the present disclosure may include, but are not limited to, mobile terminals such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g., car navigation terminals), and the like, and fixed terminals such as digital TVs, desktop computers, and the like. The electronic device shown in fig. 7 is only an example, and should not bring any limitation to the functions and the scope of use of the embodiments of the present disclosure.
As shown in fig. 7, electronic device 600 may include a processing means (e.g., central processing unit, graphics processor, etc.) 601 that may perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM)602 or a program loaded from a storage means 608 into a Random Access Memory (RAM) 603. In the RAM 603, various programs and data necessary for the operation of the electronic apparatus 600 are also stored. The processing device 601, the ROM 602, and the RAM 603 are connected to each other via a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
In general, input devices 606 including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc., output devices 607 including, for example, a liquid crystal display (L CD), speaker, vibrator, etc., storage devices 608 including, for example, magnetic tape, hard disk, etc., and communication devices 609.
In particular, according to an embodiment of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication means 609, or may be installed from the storage means 608, or may be installed from the ROM 602. The computer program, when executed by the processing device 601, performs the above-described functions defined in the methods of the embodiments of the present disclosure.
It should be noted that the computer readable medium in the present disclosure can be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the present disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In contrast, in the present disclosure, a computer readable signal medium may comprise a propagated data signal with computer readable program code embodied therein, either in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
The computer readable medium may be embodied in the electronic device; or may exist separately without being assembled into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: acquiring two or more continuous human face images; determining corresponding human face action characteristic parameters according to human face changes in the human face images; and generating a model action instruction according to the human face action characteristic parameters, so that a preset 3D image executes corresponding human face action according to the human face action characteristic parameters.
Alternatively, the computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: acquiring two or more continuous human face images; determining corresponding human face action characteristic parameters according to human face changes in the human face images; and generating a model action instruction according to the human face action characteristic parameters, so that a preset 3D image executes corresponding human face action according to the human face action characteristic parameters.
Computer program code for carrying out operations of the present disclosure may be written in any combination of one or more programming languages, including AN object oriented programming language such as Java, Smalltalk, C + +, and conventional procedural programming languages, such as the "C" programming language or similar programming languages.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units described in the embodiments of the present disclosure may be implemented by software or hardware. Wherein the name of the unit does not in some cases constitute a limitation of the unit itself, for example, the preset 3D character motion performing module may also be described as an "motion performing module".
EXAMPLE seven
The seventh embodiment of the present disclosure further provides a computer-readable storage medium, on which a computer program is stored, where the computer program, when executed by a model action apparatus, implements a model action method according to the first embodiment of the present disclosure, where the method includes: acquiring two or more continuous human face images; determining corresponding human face action characteristic parameters according to human face changes in the human face images; and generating a model action instruction according to the human face action characteristic parameters, so that a preset 3D image executes corresponding human face action according to the human face action characteristic parameters.
Of course, the computer program stored on the computer readable storage medium provided by the embodiments of the present disclosure is not limited to implement the method operations described above when being executed, and may also implement the relevant operations in the model action method provided by any embodiments of the present disclosure.
Based on the understanding that the technical solutions of the present disclosure can be embodied in the form of software products, which can be stored in a computer-readable storage medium, such as a floppy disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a flash Memory (F L ASH), a hard disk or an optical disk of a computer, and the like, and include instructions for enabling a computer device (which may be a personal computer, a server, or a network device, and the like) to execute the methods described in the embodiments of the present disclosure.
It should be noted that, in the embodiment of the model action apparatus, the included units and modules are merely divided according to functional logic, but are not limited to the above division as long as the corresponding functions can be realized; in addition, specific names of the functional units are only used for distinguishing one functional unit from another, and are not used for limiting the protection scope of the present disclosure.
It is to be noted that the foregoing is only illustrative of the preferred embodiments of the present disclosure and the technical principles employed. Those skilled in the art will appreciate that the present disclosure is not limited to the particular embodiments described herein, and that various obvious changes, adaptations, and substitutions are possible, without departing from the scope of the present disclosure. Therefore, although the present disclosure has been described in greater detail with reference to the above embodiments, the present disclosure is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present disclosure, the scope of which is determined by the scope of the appended claims.

Claims (10)

1. A method of model action, comprising:
acquiring two or more continuous human face images;
determining corresponding human face action characteristic parameters according to human face changes in the human face images;
and generating a model action instruction according to the human face action characteristic parameters, so that a preset 3D image executes corresponding human face action according to the human face action characteristic parameters.
2. The method according to claim 1, before generating a model action command according to the face action feature parameters to enable a preset 3D avatar to perform a corresponding face action according to the face action feature parameters, further comprising:
acquiring a face image of at least one preset angle, and extracting face feature data of the face image;
constructing a human face 3D model corresponding to the human face image according to the human face feature data;
and applying the human face 3D model to a preset 3D image to obtain the preset 3D image with the human face characteristic data.
3. The method of claim 1, wherein the face motion feature parameters comprise at least one of: speed of movement, direction of movement, and distance of movement.
4. The method of claim 3, wherein determining the corresponding facial motion feature parameters according to the facial changes in the facial image comprises:
and determining the moving speed, the moving direction and the moving distance of the face according to the position change of the preset part in the face image.
5. The method according to claim 4, wherein the generating a model action command according to the face action feature parameters to enable a preset 3D image to execute corresponding face actions according to the face action feature parameters comprises:
and generating a model action instruction according to the moving speed, the moving direction and the moving distance of the human face, so that the preset 3D image executes human face action according to the moving speed, the moving direction and the moving distance of the human face.
6. The method of claim 1, wherein the preset 3D avatar is: a cartoon 3D character, a professional 3D character, or a gender 3D character.
7. A model motion device, comprising:
the face image acquisition module is used for acquiring two or more continuous face images;
the human face action characteristic parameter determining module is used for determining corresponding human face action characteristic parameters according to human face changes in the human face image;
and the preset 3D image action execution module is used for generating a model action instruction according to the human face action characteristic parameters so that the preset 3D image executes corresponding human face actions according to the human face action characteristic parameters.
8. A sound box with a screen is characterized by comprising a main body, a controller positioned in the main body and at least two cameras positioned on the main body; the distance between the at least two cameras is greater than a distance threshold setting, the controller having disposed therein the apparatus of claim 7.
9. An electronic device, characterized in that the device comprises:
one or more processors;
a memory for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement the model action method of any of claims 1-6.
10. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the model action method according to any one of claims 1 to 6.
CN201910037303.3A 2019-01-15 2019-01-15 Model action method and device, sound box with screen, electronic equipment and storage medium Pending CN111435546A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201910037303.3A CN111435546A (en) 2019-01-15 2019-01-15 Model action method and device, sound box with screen, electronic equipment and storage medium
PCT/CN2020/070375 WO2020147598A1 (en) 2019-01-15 2020-01-06 Model action method and apparatus, speaker having screen, electronic device, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910037303.3A CN111435546A (en) 2019-01-15 2019-01-15 Model action method and device, sound box with screen, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN111435546A true CN111435546A (en) 2020-07-21

Family

ID=71580067

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910037303.3A Pending CN111435546A (en) 2019-01-15 2019-01-15 Model action method and device, sound box with screen, electronic equipment and storage medium

Country Status (2)

Country Link
CN (1) CN111435546A (en)
WO (1) WO2020147598A1 (en)

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1991817A (en) * 2005-12-29 2007-07-04 罗兆鑫 E-mail auxiliary and words-to-voice system
CN102169642A (en) * 2011-04-06 2011-08-31 李一波 Interactive virtual teacher system having intelligent error correction function
CN102222363A (en) * 2011-07-19 2011-10-19 杭州实时数码科技有限公司 Method for fast constructing high-accuracy personalized face model on basis of facial images
CN103414782A (en) * 2013-08-20 2013-11-27 苏州跨界软件科技有限公司 Parent-child system and method based on virtual character
CN103413468A (en) * 2013-08-20 2013-11-27 苏州跨界软件科技有限公司 Parent-child educational method based on a virtual character
CN105518582A (en) * 2015-06-30 2016-04-20 北京旷视科技有限公司 Vivo detection method and device, computer program product
CN105590486A (en) * 2014-10-21 2016-05-18 黄小曼 Machine vision-based pedestal-type finger reader, related system device and related method
CN106023692A (en) * 2016-05-13 2016-10-12 广东博士早教科技有限公司 AR interest learning system and method based on entertainment interaction
CN106910247A (en) * 2017-03-20 2017-06-30 厦门幻世网络科技有限公司 Method and apparatus for generating three-dimensional head portrait model
CN107333086A (en) * 2016-04-29 2017-11-07 掌赢信息科技(上海)有限公司 A kind of method and device that video communication is carried out in virtual scene
CN107705355A (en) * 2017-09-08 2018-02-16 郭睿 A kind of 3D human body modeling methods and device based on plurality of pictures
CN107831902A (en) * 2017-11-23 2018-03-23 腾讯科技(上海)有限公司 A kind of motion control method and its equipment, storage medium, terminal
CN108090463A (en) * 2017-12-29 2018-05-29 腾讯科技(深圳)有限公司 Object control method, apparatus, storage medium and computer equipment
CN108229239A (en) * 2016-12-09 2018-06-29 武汉斗鱼网络科技有限公司 A kind of method and device of image procossing
CN108615256A (en) * 2018-03-29 2018-10-02 西南民族大学 A kind of face three-dimensional rebuilding method and device
CN108806360A (en) * 2018-05-31 2018-11-13 北京智能管家科技有限公司 Reading partner method, apparatus, equipment and storage medium
CN109118562A (en) * 2018-08-31 2019-01-01 百度在线网络技术(北京)有限公司 Explanation video creating method, device and the terminal of virtual image

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103105924B (en) * 2011-11-15 2015-09-09 中国科学院深圳先进技术研究院 Man-machine interaction method and device
US20150178988A1 (en) * 2012-05-22 2015-06-25 Telefonica, S.A. Method and a system for generating a realistic 3d reconstruction model for an object or being
CN104616347A (en) * 2015-01-05 2015-05-13 掌赢信息科技(上海)有限公司 Expression migration method, electronic equipment and system
CN106447785A (en) * 2016-09-30 2017-02-22 北京奇虎科技有限公司 Method for driving virtual character and device thereof
CN107479693A (en) * 2017-07-07 2017-12-15 大圣科技股份有限公司 Real-time hand recognition methods based on RGB information, storage medium, electronic equipment
CN108875633B (en) * 2018-06-19 2022-02-08 北京旷视科技有限公司 Expression detection and expression driving method, device and system and storage medium

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1991817A (en) * 2005-12-29 2007-07-04 罗兆鑫 E-mail auxiliary and words-to-voice system
CN102169642A (en) * 2011-04-06 2011-08-31 李一波 Interactive virtual teacher system having intelligent error correction function
CN102222363A (en) * 2011-07-19 2011-10-19 杭州实时数码科技有限公司 Method for fast constructing high-accuracy personalized face model on basis of facial images
CN103414782A (en) * 2013-08-20 2013-11-27 苏州跨界软件科技有限公司 Parent-child system and method based on virtual character
CN103413468A (en) * 2013-08-20 2013-11-27 苏州跨界软件科技有限公司 Parent-child educational method based on a virtual character
CN105590486A (en) * 2014-10-21 2016-05-18 黄小曼 Machine vision-based pedestal-type finger reader, related system device and related method
CN105518582A (en) * 2015-06-30 2016-04-20 北京旷视科技有限公司 Vivo detection method and device, computer program product
CN107333086A (en) * 2016-04-29 2017-11-07 掌赢信息科技(上海)有限公司 A kind of method and device that video communication is carried out in virtual scene
CN106023692A (en) * 2016-05-13 2016-10-12 广东博士早教科技有限公司 AR interest learning system and method based on entertainment interaction
CN108229239A (en) * 2016-12-09 2018-06-29 武汉斗鱼网络科技有限公司 A kind of method and device of image procossing
CN106910247A (en) * 2017-03-20 2017-06-30 厦门幻世网络科技有限公司 Method and apparatus for generating three-dimensional head portrait model
CN107705355A (en) * 2017-09-08 2018-02-16 郭睿 A kind of 3D human body modeling methods and device based on plurality of pictures
CN107831902A (en) * 2017-11-23 2018-03-23 腾讯科技(上海)有限公司 A kind of motion control method and its equipment, storage medium, terminal
CN108090463A (en) * 2017-12-29 2018-05-29 腾讯科技(深圳)有限公司 Object control method, apparatus, storage medium and computer equipment
CN108615256A (en) * 2018-03-29 2018-10-02 西南民族大学 A kind of face three-dimensional rebuilding method and device
CN108806360A (en) * 2018-05-31 2018-11-13 北京智能管家科技有限公司 Reading partner method, apparatus, equipment and storage medium
CN109118562A (en) * 2018-08-31 2019-01-01 百度在线网络技术(北京)有限公司 Explanation video creating method, device and the terminal of virtual image

Also Published As

Publication number Publication date
WO2020147598A1 (en) 2020-07-23

Similar Documents

Publication Publication Date Title
CN109462776B (en) Video special effect adding method and device, terminal equipment and storage medium
US20210029305A1 (en) Method and apparatus for adding a video special effect, terminal device and storage medium
US11463611B2 (en) Interactive application adapted for use by multiple users via a distributed computer-based system
US20160110922A1 (en) Method and system for enhancing communication by using augmented reality
CN109474850B (en) Motion pixel video special effect adding method and device, terminal equipment and storage medium
CN113946211A (en) Method for interacting multiple objects based on metauniverse and related equipment
KR20120123330A (en) Camera navigation for presentations
WO2022170958A1 (en) Augmented reality-based display method and device, storage medium, and program product
CN109600559B (en) Video special effect adding method and device, terminal equipment and storage medium
EP4243398A1 (en) Video processing method and apparatus, electronic device, and storage medium
KR20220093342A (en) Method, device and related products for implementing split mirror effect
CN109348277B (en) Motion pixel video special effect adding method and device, terminal equipment and storage medium
WO2020151491A1 (en) Image deformation control method and device and hardware device
CN110047119B (en) Animation generation method and device comprising dynamic background and electronic equipment
CN111638791B (en) Virtual character generation method and device, electronic equipment and storage medium
CN109035415B (en) Virtual model processing method, device, equipment and computer readable storage medium
US20230182028A1 (en) Game live broadcast interaction method and apparatus
WO2023116653A1 (en) Element display method and apparatus, and electronic device and storage medium
CN110035271B (en) Fidelity image generation method and device and electronic equipment
KR20220148915A (en) Audio processing methods, apparatus, readable media and electronic devices
CN113806306B (en) Media file processing method, device, equipment, readable storage medium and product
JP2022500795A (en) Avatar animation
CN109271929B (en) Detection method and device
US20230267664A1 (en) Animation processing method and apparatus, electronic device and storage medium
CN114697568B (en) Special effect video determining method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination