WO2020186934A1 - Method, apparatus, and electronic device for generating animation containing dynamic background - Google Patents

Method, apparatus, and electronic device for generating animation containing dynamic background Download PDF

Info

Publication number
WO2020186934A1
WO2020186934A1 PCT/CN2020/074369 CN2020074369W WO2020186934A1 WO 2020186934 A1 WO2020186934 A1 WO 2020186934A1 CN 2020074369 W CN2020074369 W CN 2020074369W WO 2020186934 A1 WO2020186934 A1 WO 2020186934A1
Authority
WO
WIPO (PCT)
Prior art keywords
animation
specific area
target object
input information
area
Prior art date
Application number
PCT/CN2020/074369
Other languages
French (fr)
Chinese (zh)
Inventor
郭冠军
Original Assignee
北京字节跳动网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字节跳动网络技术有限公司 filed Critical 北京字节跳动网络技术有限公司
Publication of WO2020186934A1 publication Critical patent/WO2020186934A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations

Definitions

  • the present disclosure relates to the field of data processing technology, and in particular, to an animation generating method, device and electronic device including a dynamic background.
  • a virtual anchor is provided to anthropomorphically broadcast the live content and provide necessary guidance for the live broadcast. Thereby enhancing the presence and interactivity of the live broadcast and improving the effect of the web live broadcast.
  • Expression simulation for example, mouth movement simulation
  • the current realization of expression simulation is mainly based on text-driven, natural voice-driven and audio-video hybrid modeling methods to drive the facial expressions of characters.
  • the text-driven approach is usually the TTS (Text to Speech, text-to-speech) engine that converts the input text information into the corresponding phoneme sequence, phoneme duration, and corresponding voice waveform, and then select the corresponding model unit in the model library , Through smooth processing and corresponding synchronization algorithm, finally present the voice and facial expression actions corresponding to the input text content.
  • TTS Text to Speech, text-to-speech
  • the emoji simulation in the prior art has a single or even distorted emoji simulation, which is more like a "robot" performing a performance, and the fidelity of the emoji action is still far from the expression of a real person.
  • the embodiments of the present disclosure provide a method, device and electronic device for generating an animation including a dynamic background, which at least partially solves the problems existing in the prior art.
  • an embodiment of the present disclosure provides a method for generating an animation including a dynamic background, including: obtaining a reconstruction model related to a specific area, a first element, and a second element of a target object, the specific area belonging to the A part of the target object, the first element and the second element are located in the specific area;
  • While generating the first animation dynamically selecting an animation matching the first animation from a plurality of preset animations as the background animation of the first animation;
  • a final animation related to the input information is generated.
  • the method before obtaining the reconstruction model related to the specific area, the first element and the second element of the target object, the method further includes:
  • a plurality of images including the target object are acquired, and based on the plurality of images, a reconstruction model related to the specific area, the first element and the second element of the target object is trained.
  • the training of the reconstruction model related to the specific area, the first element and the second element of the target object includes:
  • the texture map of the specific area is determined.
  • the training of the reconstruction model related to the specific area, the first element and the second element of the target object includes:
  • the detected feature points into a first type feature point and a second type feature point, the first type feature point is used to form a first closed area, and the second type feature point is used to form a second closed area;
  • a first color is filled in the first closed area, and a second color is filled in the second closed area, and the first color is different from the second color.
  • the training of the reconstruction model related to the specific area, the first element and the second element of the target object includes:
  • a third color is filled in the third closed area.
  • the determining the texture feature of the specific area related to the input information, the action of the first element, and the action of the second element based on the reconstruction model includes :
  • the motion parameters obtained after parsing the input information are matched to the first element and the second element to form an action of the first element and the second element.
  • the dynamically selecting an animation matching the first animation from a plurality of preset animations as the background animation of the first animation includes:
  • Analyze the current scene of the input information dynamically select an animation that matches the current scene from a plurality of preset animations as the background animation.
  • the generating a final animation related to the input information based on the first animation and the background animation includes:
  • the specific area is a face area
  • the first element is an eye
  • the second element is a mouth
  • an animation generating device including a dynamic background, including:
  • the acquiring module is used to acquire a reconstruction model related to a specific area, a first element and a second element of a target object, the specific area belongs to a part of the target object, and the first element and the second element are located in the Within a specific area
  • the determining module is configured to determine, based on the reconstruction model, the texture feature of the specific area related to the input information, the action of the first element and the action of the second element, the texture feature of the specific area, the The actions of the first element and the actions of the second element form a first animation related to the input information;
  • the selection module is configured to dynamically select an animation matching the first animation from a plurality of preset animations as the background animation of the first animation while generating the first animation;
  • a generating module is used to generate a final animation related to the input information based on the first animation and the background animation.
  • an embodiment of the present disclosure also provides an electronic device, which includes:
  • At least one processor and,
  • a memory communicatively connected with the at least one processor; wherein,
  • the memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute any of the foregoing first aspect or any implementation of the first aspect Including dynamic background animation generation method.
  • embodiments of the present disclosure also provide a non-transitory computer-readable storage medium that stores computer instructions that are used to make the computer execute the first aspect or the first aspect described above.
  • An animation generating method including a dynamic background in any implementation manner of one aspect.
  • the embodiments of the present disclosure also provide a computer program product.
  • the computer program product includes a computing program stored on a non-transitory computer-readable storage medium.
  • the computer program includes program instructions. When executing, the computer is caused to execute the animation generating method including the dynamic background in the foregoing first aspect or any implementation manner of the first aspect.
  • the animation generation solution including dynamic background in the embodiment of the present disclosure includes obtaining a reconstruction model related to a specific area, a first element, and a second element of a target object.
  • the specific area belongs to a part of the target object.
  • the first element and the second element are located in the specific area; based on the reconstruction model, the texture feature of the specific area related to the input information, the action of the first element and the action of the second element are determined, so The texture feature of the specific area, the action of the first element, and the action of the second element form a first animation related to the input information; while the first animation is generated, the preset Among the animations, an animation matching the first animation is dynamically selected as a background animation of the first animation; based on the first animation and the background animation, a final animation related to the input information is generated.
  • FIG. 1 is a schematic diagram of an animation generation process including a dynamic background provided by an embodiment of the disclosure
  • FIG. 2 is a schematic diagram of another animation generation process including a dynamic background provided by an embodiment of the present disclosure
  • FIG. 3 is a schematic diagram of another animation generation process including a dynamic background provided by an embodiment of the present disclosure
  • FIG. 4 is a schematic diagram of another animation generation process including a dynamic background provided by an embodiment of the present disclosure
  • FIG. 5 is a schematic structural diagram of an animation generating device including a dynamic background provided by an embodiment of the disclosure
  • FIG. 6 is a schematic diagram of an electronic device provided by an embodiment of the disclosure.
  • the embodiment of the present disclosure provides a method for generating an animation including a dynamic background.
  • the method for generating an animation containing a dynamic background provided in this embodiment can be executed by a computing device, which can be implemented as software, or as a combination of software and hardware, and the computing device can be integrated in a server, terminal device, etc. .
  • an animation generating method including a dynamic background includes the following steps S101-S104:
  • the target object can be a real person who can broadcast on the Internet, or other objects with information dissemination functions, such as TV programs. Hosts, news program announcers, teachers who give lectures, etc.
  • the target object is usually a person with a communication function. Since this type of person usually has a certain reputation, when there is a large amount of content that the target object needs to broadcast including voice and/or video actions, it usually costs more. At the same time, for live broadcast programs, the target object usually cannot appear in multiple live broadcast rooms (or multiple live channels) at the same time. At this time, if you want to present an effect such as "Anchor Clone", it is usually difficult to achieve this effect through live broadcast by real people.
  • the reconstruction model which can predict the image and actions of the target object based on the input information, thereby generating actions and expressions that match the input information. For example, for a piece of news that needs to be broadcast, the target audience can broadcast the news in the form of a news broadcaster.
  • the solution of the present disclosure only selects the component elements that are most relevant to the identification of the target object: the specific area, the first element and the second element.
  • the specific area is the face area of the target object
  • the first element is the eyes of the target object
  • the second element is the mouth of the target object.
  • the various actions and expressions of the target object in the video can be predicted through video animation.
  • the fidelity image can be used to generate a video file containing the actions and expressions of the target object.
  • the fidelity image can be used as all frames or key frames of the video file, and the fidelity image contains information that matches the input information.
  • the input information can be in multiple ways.
  • the input information can be that the target object needs to play a piece of text or a piece of audio content at the same time during the animation display process, that is, the input information can be in the form of text or audio.
  • the target object can produce different animations.
  • the input information is transformed into parameters matching the texture map and the shape constraint map, and the reconstruction model obtained after training is used to finally complete the guarantee by calling the texture map and the shape constraint map Image generation.
  • the texture map of the specific area of the target object and the shape constraints of the first element and the second element can be given, and the trained reconstruction model can be used to predict the image of the two-dimensional anchor image.
  • the continuous first element The shape constraint of the second element and the fixed texture of the specific area are used as input to predict the continuous animation image of the target object.
  • the background animation can be a variety of different types of background animation, for example, it can be serious, lively, happy, sad, and other different styles of animation.
  • the background animation has a certain length so that it can be called easily.
  • the background animation When calling the background animation, it is necessary to analyze the current scene of the input information. Based on the results of the analysis, dynamically select the animation that matches the current scene from the preset multiple animations as the background animation. For example, if the analysis knows that the current input information is excited, then the excited background animation is called. When a change in the scene of the input information is detected, even if the currently called background animation has not been played, it will switch to call the background animation matching the changed scene. After the currently called background animation is played, if it is found that the current scene of the input information has not changed, the current background animation will continue to be played in a loop.
  • S104 Generate a final animation related to the input information based on the first animation and the background animation.
  • the final animation related to the input information is generated.
  • the final animation can be stored or disseminated in the form of video files.
  • the background animation is switched. For this, it can be determined whether the background animation is composed of multiple different types of animation splicing, and if so, smooth processing is performed on the different types of animation. Make the transition between different types of animation as background animation more natural.
  • the reconstruction model is obtained through neural network training. For this reason, before obtaining the reconstruction model related to the specific area, the first element and the second element of the target object, it is also necessary to collect multiple images containing the target object. Images, training a reconstruction model related to the specific area, the first element and the second element of the target object.
  • training a reconstruction model related to a specific region, first element, and second element of the target object can be performed through a specific implementation manner according to an embodiment of the present disclosure, which can include:
  • S201 Detect specific areas on the multiple images to obtain a target area.
  • the target area may be a face area.
  • face areas existing on multiple images can be detected by means of face detection.
  • S202 Perform 3D reconstruction on the target area to obtain a 3D area object.
  • the constituent objects on the target object can be selected to model the target object.
  • Various methods can be used for 3D modeling of the target area. For example, a depth value can be added on the basis of the two-dimensional pixel values of the multiple image planes, and the depth value can be obtained based on the brightness values of the multiple images.
  • S203 Acquire a three-dimensional grid of the 3D area object, where the three-dimensional grid includes preset coordinate values.
  • the 3D area object uses a three-dimensional grid to describe its specific location.
  • specific coordinate values are set for the three-dimensional grid.
  • the three-dimensional grid can be described by setting two-dimensional plane coordinates and spatial height coordinates.
  • S204 Determine a texture map of the specific region based on pixel values on different three-dimensional grid coordinates.
  • the pixel values on different three-dimensional grid coordinates can be connected together to form a grid plane, which forms a texture map of a specific area.
  • the training is related to the specific area of the target object, the first element and Refactored models related to the second element, including:
  • S301 Perform feature point detection on the first element on the multiple images.
  • the contour and pupil of the eye can be used as feature objects, and feature points related to the contour and pupil of the eye can be obtained through feature detection.
  • the first element is composed of at least two components (for example, pupil and sclera). Based on the different parts to which the feature points belong, the detected feature points are divided into first type feature points and second type feature points, for example, the first Type feature points are feature points related to the pupil, and the second type feature points are feature points related to the sclera.
  • the first type feature points and the second type feature points are feature points related to the pupil
  • the second type feature points are feature points related to the sclera.
  • S303 Fill a first color in the first closed area, and a second color in the second closed area, where the first color is different from the second color.
  • the training of a reconstruction model related to a specific area, a first element, and a second element of the target object includes:
  • S401 Perform feature point detection on the second element on the multiple images.
  • the feature setting of the second element can detect the feature points related to the second element.
  • the entire mouth can be used as a feature object, and the feature points related to the mouth can be obtained through feature detection.
  • S402 Form a third closed area based on all the detected feature points.
  • a closed area can be formed by connecting all the feature points together, and then a third closed area can be obtained.
  • the third closed area can be filled with a color matching the second element. Taking the mouth as an example, the third closed area can be filled with red that matches the mouth.
  • the contour of the specific area of the target object is predicted, and the texture map determined by the reconstruction model is filled in the predicted contour, thereby obtaining the contour of the specific area of the target object Figure (e.g., face contour map).
  • the motion parameters obtained after parsing the input information are matched to the first element and the second element, and the actions of the first element and the second element are formed.
  • the input information may be in multiple ways, for example, the input information may be in the form of text or audio.
  • the input information is transformed into a first analysis result after data analysis.
  • the first analysis result includes parameters matching the texture map and the shape constraint map of the first element and the second element.
  • the reconstruction model obtained after training is used to pass The texture map and the shape constraint map are called to finally complete the generation of the target object prediction action.
  • the first analysis result contains the motion amplitude parameters for the first element and the second element on the target object. Take the mouth as an example. When the mouth is fully opened, the motion amplitude can be quantified as 1. When the mouth is fully closed, the motion amplitude It can be quantified as 0. By quantifying the value between 0 and 1, the intermediate state of the mouth between fully opened and fully closed can be described.
  • an animation generating device 50 containing a dynamic background including:
  • the obtaining module 501 is configured to obtain a reconstruction model related to a specific area, a first element, and a second element of a target object, the specific area belongs to a part of the target object, and the first element and the second element are located in the target object. Said within a specific area.
  • the target object can be a real person who can broadcast on the Internet, or other objects with information dissemination functions, such as TV programs. Hosts, news program announcers, teachers who give lectures, etc.
  • the target object is usually a person with a communication function. Since this type of person usually has a certain reputation, when there is a large amount of content that the target object needs to broadcast including voice and/or video actions, it usually costs more. At the same time, for live broadcast programs, the target object usually cannot appear in multiple live broadcast rooms (or multiple live channels) at the same time. At this time, if you want to present an effect such as "Anchor Clone", it is usually difficult to achieve this effect through live broadcast by real people.
  • the reconstruction model which can predict the image and actions of the target object based on the input information, thereby generating actions and expressions that match the input information. For example, for a piece of news that needs to be broadcast, the target audience can broadcast the news in the form of a news broadcaster.
  • the solution of the present disclosure only selects the component elements that are most relevant to the identification of the target object: the specific area, the first element and the second element.
  • the specific area is the face area of the target object
  • the first element is the eyes of the target object
  • the second element is the mouth of the target object.
  • the determining module 502 is configured to determine, based on the reconstruction model, the texture feature of the specific area, the action of the first element and the action of the second element related to the input information, the texture feature of the specific area, the The action of the first element and the action of the second element form a first animation related to the input information.
  • the various actions and expressions of the target object in the video can be predicted through video animation.
  • the fidelity image can be used to generate a video file containing the actions and expressions of the target object.
  • the fidelity image can be used as all frames or key frames of the video file, and the fidelity image contains information that matches the input information.
  • the input information can be in a variety of ways.
  • the input information can be in the form of text or audio.
  • the input information is transformed into parameters matching the texture map and the shape constraint map, and the reconstruction model obtained after training is used to finally complete the guarantee by calling the texture map and the shape constraint map Image generation.
  • the texture map of the specific area of the target object and the shape constraints of the first element and the second element can be given, and the trained reconstruction model can be used to predict the image of the two-dimensional anchor image through the continuous first element,
  • the shape constraint of the second element and the fixed texture of the specific area are used as input to predict the continuous animation image of the target object.
  • the selection module 503 is configured to dynamically select an animation matching the first animation from a plurality of preset animations as the background animation of the first animation while generating the first animation.
  • the background animation can be a variety of different types of background animation, for example, it can be serious, lively, happy, sad, and other different styles of animation.
  • the background animation has a certain length so that it can be called easily.
  • the background animation When calling the background animation, it is necessary to analyze the current scene of the input information, and based on the analysis result, dynamically select an animation that matches the current scene from a plurality of preset animations as the background animation. For example, if the analysis knows that the current input information is excited, then the excited background animation is called. When a change in the scene of the input information is detected, even if the currently called background animation has not been played, it will switch to call the background animation matching the changed scene. After the currently called background animation is played, if it is found that the current scene of the input information has not changed, the current background animation will continue to be played in a loop.
  • the generating module 504 is configured to generate a final animation related to the input information based on the first animation and the background animation.
  • the final animation related to the input information is generated.
  • the final animation can be stored or disseminated in the form of video files.
  • the background animation is switched. For this, it can be determined whether the background animation is composed of multiple different types of animation splicing, and if so, smooth processing is performed on the different types of animation. Make the transition between different types of animation more natural.
  • the device shown in FIG. 5 can correspondingly execute the content in the foregoing method embodiment.
  • an electronic device 60 which includes:
  • At least one processor and,
  • a memory communicatively connected with the at least one processor; wherein,
  • the memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute the animation generation method including the dynamic background in the foregoing method embodiment.
  • the embodiments of the present disclosure also provide a non-transitory computer-readable storage medium that stores computer instructions, and the computer instructions are used to make the computer execute the foregoing method embodiments.
  • the embodiments of the present disclosure also provide a computer program product, the computer program product includes a calculation program stored on a non-transitory computer-readable storage medium, the computer program includes program instructions, when the program instructions are executed by a computer, The computer executes the animation generation method including the dynamic background in the foregoing method embodiment.
  • FIG. 6 shows a schematic structural diagram of an electronic device 60 suitable for implementing embodiments of the present disclosure.
  • Electronic devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablets), PMPs (portable multimedia players), vehicle-mounted terminals (for example, Mobile terminals such as car navigation terminals) and fixed terminals such as digital TVs, desktop computers, etc.
  • the electronic device shown in FIG. 6 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present disclosure.
  • the electronic device 60 may include a processing device (such as a central processing unit, a graphics processor, etc.) 601, which may be loaded into a random access device according to a program stored in a read-only memory (ROM) 602 or from a storage device 608.
  • the program in the memory (RAM) 603 executes various appropriate actions and processing.
  • the RAM 603 also stores various programs and data required for the operation of the electronic device 60.
  • the processing device 601, the ROM 602, and the RAM 603 are connected to each other through a bus 604.
  • An input/output (I/O) interface 605 is also connected to the bus 604.
  • the following devices can be connected to the I/O interface 605: including input devices 606 such as touch screen, touch panel, keyboard, mouse, image sensor, microphone, accelerometer, gyroscope, etc.; including, for example, liquid crystal display (LCD), speakers, An output device 607 such as a vibrator; a storage device 608 such as a magnetic tape, a hard disk, etc.; and a communication device 609.
  • the communication device 609 may allow the electronic device 60 to perform wireless or wired communication with other devices to exchange data.
  • the figure shows the electronic device 60 having various devices, it should be understood that it is not required to implement or have all the illustrated devices. It may alternatively be implemented or provided with more or fewer devices.
  • the process described above with reference to the flowchart can be implemented as a computer software program.
  • the embodiments of the present disclosure include a computer program product, which includes a computer program carried on a computer-readable medium, and the computer program contains program code for executing the method shown in the flowchart.
  • the computer program may be downloaded and installed from the network through the communication device 609, or installed from the storage device 608, or installed from the ROM602.
  • the processing device 601 the above-mentioned functions defined in the method of the embodiment of the present disclosure are executed.
  • the above-mentioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the two.
  • the computer-readable storage medium may be, for example, but not limited to, an electric, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the above. More specific examples of computer-readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program, and the program may be used by or in combination with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in a baseband or as a part of a carrier wave, and a computer-readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • the computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium.
  • the computer-readable signal medium may send, propagate, or transmit the program for use by or in combination with the instruction execution system, apparatus, or device .
  • the program code contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to: wire, optical cable, RF (Radio Frequency), etc., or any suitable combination of the above.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; or it may exist alone without being assembled into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device: obtains at least two Internet protocol addresses; and sends to the node evaluation device including the at least two A node evaluation request for an Internet Protocol address, wherein the node evaluation device selects an Internet Protocol address from the at least two Internet Protocol addresses and returns it; receives the Internet Protocol address returned by the node evaluation device; wherein, the obtained The Internet Protocol address indicates the edge node in the content distribution network.
  • the aforementioned computer-readable medium carries one or more programs, and when the aforementioned one or more programs are executed by the electronic device, the electronic device: receives a node evaluation request including at least two Internet Protocol addresses; Among the at least two Internet Protocol addresses, select an Internet Protocol address; return the selected Internet Protocol address; wherein, the received Internet Protocol address indicates an edge node in the content distribution network.
  • the computer program code used to perform the operations of the present disclosure may be written in one or more programming languages or a combination thereof.
  • the above-mentioned programming languages include object-oriented programming languages—such as Java, Smalltalk, C++, and also conventional Procedural programming language-such as "C" language or similar programming language.
  • the program code can be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, using an Internet service provider to pass Internet connection).
  • LAN local area network
  • WAN wide area network
  • each block in the flowchart or block diagram can represent a module, program segment, or part of code, and the module, program segment, or part of code contains one or more for realizing the specified logical function Executable instructions.
  • the functions marked in the block may also occur in a different order from the order marked in the drawings. For example, two blocks shown in succession can actually be executed substantially in parallel, or they can sometimes be executed in the reverse order, depending on the functions involved.
  • each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart can be implemented by a dedicated hardware-based system that performs the specified functions or operations Or it can be realized by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present disclosure may be implemented in a software manner, or may be implemented in a hardware manner.
  • the name of the unit does not constitute a limitation on the unit itself under certain circumstances.
  • the first obtaining unit can also be described as "a unit for obtaining at least two Internet Protocol addresses.”

Abstract

The embodiments of the present disclosure provide a method, apparatus, and electronic device for generating an animation containing a dynamic background, belonging to the technical field of data processing, said method comprising: obtaining a reconstruction model related to a specific region and first element and second element of a target object; on the basis of said reconstruction model, determining a texture feature of the specific region related to input information, and an action of the first element and an action of the second element; while generating a first animation, dynamically selecting, from a plurality of preset animations, an animation matching said first animation to be the background animation of the first animation; on the basis of the first animation and said background animation, generating a final animation related to the input information. The processing solution of the present disclosure increases the authenticity of the generated image.

Description

包含动态背景的动画生成方法、装置及电子设备Animation generating method, device and electronic equipment containing dynamic background
相关申请的交叉引用Cross references to related applications
本申请要求于2019年03月20日提交的,申请号为201910214896.6、发明名称为“包含动态背景的动画生成方法、装置及电子设备”的中国专利申请的优先权,该申请的全文通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed on March 20, 2019, with the application number 201910214896.6 and the title of the invention "animation generating method, device and electronic equipment containing dynamic background", the full text of which is incorporated by reference In this application.
技术领域Technical field
本公开涉及数据处理技术领域,尤其涉及一种包含动态背景的动画生成方法、装置及电子设备。The present disclosure relates to the field of data processing technology, and in particular, to an animation generating method, device and electronic device including a dynamic background.
背景技术Background technique
随着网络技术的发展,人工智能技术在网络场景中的应用得到了极大的提升。作为一个具体的应用需求,越来越多的网络环境中使用虚拟的人物进行交互,例如在网络直播中提供虚拟的主播对直播内容进行拟人化的播报,并为直播用于提供必要的引导,从而增强直播的临场感和交互性,提高了网络直播的效果。With the development of network technology, the application of artificial intelligence technology in network scenarios has been greatly improved. As a specific application requirement, more and more virtual characters are used in the network environment to interact. For example, in the web live broadcast, a virtual anchor is provided to anthropomorphically broadcast the live content and provide necessary guidance for the live broadcast. Thereby enhancing the presence and interactivity of the live broadcast and improving the effect of the web live broadcast.
表情模拟(例如,嘴型动作模拟)技术是人工智能技术的一种,目前实现表情模拟主要基于文本驱动、自然语音驱动和音视频混合建模方法来驱动人物的面部表情。例如,文本驱动的方式通常是TTS(Text to Speech,文-语转换)引擎将输入的文本信息转化为对应的音素序列、音素时长以及对应的语音波形,然后在模型库中选择对应的模型单元,通过平滑处理和相应的同步算法最终呈现出和输入文本内容相对应的语音和人脸表情动作。Expression simulation (for example, mouth movement simulation) technology is a kind of artificial intelligence technology. The current realization of expression simulation is mainly based on text-driven, natural voice-driven and audio-video hybrid modeling methods to drive the facial expressions of characters. For example, the text-driven approach is usually the TTS (Text to Speech, text-to-speech) engine that converts the input text information into the corresponding phoneme sequence, phoneme duration, and corresponding voice waveform, and then select the corresponding model unit in the model library , Through smooth processing and corresponding synchronization algorithm, finally present the voice and facial expression actions corresponding to the input text content.
现有技术中的表情模拟存在表情模拟单一、甚至失真的情况,更多的像是“机器人”在进行表演,表情动作的逼真度距离真实人物的表情还具有较大的差距。The emoji simulation in the prior art has a single or even distorted emoji simulation, which is more like a "robot" performing a performance, and the fidelity of the emoji action is still far from the expression of a real person.
发明内容Summary of the invention
有鉴于此,本公开实施例提供一种包含动态背景的动画生成方法、装置及电子设备,至少部分解决现有技术中存在的问题。In view of this, the embodiments of the present disclosure provide a method, device and electronic device for generating an animation including a dynamic background, which at least partially solves the problems existing in the prior art.
第一方面,本公开实施例提供了一种包含动态背景的动画生成方法,包括:获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内;In a first aspect, an embodiment of the present disclosure provides a method for generating an animation including a dynamic background, including: obtaining a reconstruction model related to a specific area, a first element, and a second element of a target object, the specific area belonging to the A part of the target object, the first element and the second element are located in the specific area;
基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入信息相关的第一动画;Based on the reconstruction model, determine the texture feature of the specific area related to the input information, the action of the first element and the action of the second element, the texture feature of the specific area, the action of the first element And the action of the second element forms a first animation related to the input information;
在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画;While generating the first animation, dynamically selecting an animation matching the first animation from a plurality of preset animations as the background animation of the first animation;
基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。Based on the first animation and the background animation, a final animation related to the input information is generated.
根据本公开实施例的一种具体实现方式,获取与目标对象的特定区域、第一元素及第二元素相关的重构模型之前,所述方法还包括:According to a specific implementation manner of the embodiments of the present disclosure, before obtaining the reconstruction model related to the specific area, the first element and the second element of the target object, the method further includes:
采集包含目标对象的多个图像,基于所述多个图像,训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型。A plurality of images including the target object are acquired, and based on the plurality of images, a reconstruction model related to the specific area, the first element and the second element of the target object is trained.
根据本公开实施例的一种具体实现方式,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:According to a specific implementation of the embodiment of the present disclosure, the training of the reconstruction model related to the specific area, the first element and the second element of the target object includes:
对所述多个图像上的特定区域进行检测,得到目标区域;Detecting specific areas on the multiple images to obtain a target area;
对所述目标区域进行3D重建,得到3D区域对象;Performing 3D reconstruction on the target area to obtain a 3D area object;
获取所述3D区域对象的三维网格,所述三维网格包含预设坐标值;Acquiring a three-dimensional grid of the 3D area object, where the three-dimensional grid includes preset coordinate values;
基于不同三维网格坐标上的像素值,确定所述特定区域的纹理图。Based on the pixel values on different three-dimensional grid coordinates, the texture map of the specific area is determined.
根据本公开实施例的一种具体实现方式,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:According to a specific implementation of the embodiment of the present disclosure, the training of the reconstruction model related to the specific area, the first element and the second element of the target object includes:
在所述多个图像上对所述第一元素进行特征点检测;Performing feature point detection on the first element on the multiple images;
将检测得到的特征点划分为第一类型特征点和第二类型特征点,所述第一类型特征点用于形成第一闭合区域,所述第二类型特征点用于形成第二闭合区域;Divide the detected feature points into a first type feature point and a second type feature point, the first type feature point is used to form a first closed area, and the second type feature point is used to form a second closed area;
在所述第一闭合区域内填充第一颜色,在所述第二闭合区域内填充第二颜色,所述第一颜色与所述第二颜色不同。A first color is filled in the first closed area, and a second color is filled in the second closed area, and the first color is different from the second color.
根据本公开实施例的一种具体实现方式,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:According to a specific implementation of the embodiment of the present disclosure, the training of the reconstruction model related to the specific area, the first element and the second element of the target object includes:
在所述多个图像上对所述第二元素进行特征点检测;Performing feature point detection on the second element on the multiple images;
基于检测到的所有特征点,形成第三闭合区域;Based on all the detected feature points, a third closed area is formed;
在所述第三闭合区域内填充第三颜色。A third color is filled in the third closed area.
根据本公开实施例的一种具体实现方式,所述基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,包括:According to a specific implementation manner of an embodiment of the present disclosure, the determining the texture feature of the specific area related to the input information, the action of the first element, and the action of the second element based on the reconstruction model includes :
将所述目标对象的特定区域的轮廓进行预测,在预测的轮廓内填充所述重构模型确定的纹理图;Predicting the contour of the specific area of the target object, and filling the predicted contour with the texture map determined by the reconstruction model;
将所述输入信息解析后的得到运动参数匹配到所述第一元素及第二元素,形成所述第一元素及所述第二元素的动作。The motion parameters obtained after parsing the input information are matched to the first element and the second element to form an action of the first element and the second element.
根据本公开实施例的一种具体实现方式,所述从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画,包括:According to a specific implementation of the embodiment of the present disclosure, the dynamically selecting an animation matching the first animation from a plurality of preset animations as the background animation of the first animation includes:
对输入信息的当前情景进行分析,从预先设置的多个动画中动态选择选择与当前情景匹配的动画作为背景动画。Analyze the current scene of the input information, dynamically select an animation that matches the current scene from a plurality of preset animations as the background animation.
根据本公开实施例的一种具体实现方式,所述基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画,包括:According to a specific implementation of the embodiment of the present disclosure, the generating a final animation related to the input information based on the first animation and the background animation includes:
判断所述背景动画是否有多个不同类型的动画拼接组成;Judging whether the background animation is composed of multiple animations of different types;
若是,则对不同类型的动画进行平滑处理。If it is, smooth the different types of animations.
根据本公开实施例的一种具体实现方式,所述特定区域为脸部区域,所述第一元素为眼睛,所述第二元素为嘴巴。According to a specific implementation of an embodiment of the present disclosure, the specific area is a face area, the first element is an eye, and the second element is a mouth.
第二方面,本公开实施例提供了一种包含动态背景的动画生成装置,包括:In the second aspect, the embodiments of the present disclosure provide an animation generating device including a dynamic background, including:
获取模块,用于获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内;The acquiring module is used to acquire a reconstruction model related to a specific area, a first element and a second element of a target object, the specific area belongs to a part of the target object, and the first element and the second element are located in the Within a specific area
确定模块,用于基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入 信息相关的第一动画;The determining module is configured to determine, based on the reconstruction model, the texture feature of the specific area related to the input information, the action of the first element and the action of the second element, the texture feature of the specific area, the The actions of the first element and the actions of the second element form a first animation related to the input information;
选择模块,用于在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画;The selection module is configured to dynamically select an animation matching the first animation from a plurality of preset animations as the background animation of the first animation while generating the first animation;
生成模块,用于基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。A generating module is used to generate a final animation related to the input information based on the first animation and the background animation.
第三方面,本公开实施例还提供了一种电子设备,该电子设备包括:In a third aspect, an embodiment of the present disclosure also provides an electronic device, which includes:
至少一个处理器;以及,At least one processor; and,
与该至少一个处理器通信连接的存储器;其中,A memory communicatively connected with the at least one processor; wherein,
该存储器存储有可被该至少一个处理器执行的指令,该指令被该至少一个处理器执行,以使该至少一个处理器能够执行前述任第一方面或第一方面的任一实现方式中的包含动态背景的动画生成方法。The memory stores instructions that can be executed by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute any of the foregoing first aspect or any implementation of the first aspect Including dynamic background animation generation method.
第四方面,本公开实施例还提供了一种非暂态计算机可读存储介质,该非暂态计算机可读存储介质存储计算机指令,该计算机指令用于使该计算机执行前述第一方面或第一方面的任一实现方式中的包含动态背景的动画生成方法。In a fourth aspect, embodiments of the present disclosure also provide a non-transitory computer-readable storage medium that stores computer instructions that are used to make the computer execute the first aspect or the first aspect described above. An animation generating method including a dynamic background in any implementation manner of one aspect.
第五方面,本公开实施例还提供了一种计算机程序产品,该计算机程序产品包括存储在非暂态计算机可读存储介质上的计算程序,该计算机程序包括程序指令,当该程序指令被计算机执行时,使该计算机执行前述第一方面或第一方面的任一实现方式中的包含动态背景的动画生成方法。In a fifth aspect, the embodiments of the present disclosure also provide a computer program product. The computer program product includes a computing program stored on a non-transitory computer-readable storage medium. The computer program includes program instructions. When executing, the computer is caused to execute the animation generating method including the dynamic background in the foregoing first aspect or any implementation manner of the first aspect.
本公开实施例中的包含动态背景的动画生成方案,包括获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内;基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入信息相关的第一动画;在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画;基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。通过本公开的处理方案,能够真实的模拟出于输入信息相匹配的动画图像,提高了用户体验。The animation generation solution including dynamic background in the embodiment of the present disclosure includes obtaining a reconstruction model related to a specific area, a first element, and a second element of a target object. The specific area belongs to a part of the target object. The first element and the second element are located in the specific area; based on the reconstruction model, the texture feature of the specific area related to the input information, the action of the first element and the action of the second element are determined, so The texture feature of the specific area, the action of the first element, and the action of the second element form a first animation related to the input information; while the first animation is generated, the preset Among the animations, an animation matching the first animation is dynamically selected as a background animation of the first animation; based on the first animation and the background animation, a final animation related to the input information is generated. Through the processing scheme of the present disclosure, it is possible to realistically simulate an animated image matching the input information, which improves the user experience.
附图说明Description of the drawings
为了更清楚地说明本公开实施例的技术方案,下面将对实施例中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to explain the technical solutions of the embodiments of the present disclosure more clearly, the following will briefly introduce the drawings that need to be used in the embodiments. Obviously, the drawings in the following description are only some embodiments of the present disclosure. For those of ordinary skill in the art, other drawings can be obtained from these drawings without creative work.
图1为本公开实施例提供的一种包含动态背景的动画生成流程示意图;FIG. 1 is a schematic diagram of an animation generation process including a dynamic background provided by an embodiment of the disclosure;
图2为本公开实施例提供的另一种包含动态背景的动画生成流程示意图;FIG. 2 is a schematic diagram of another animation generation process including a dynamic background provided by an embodiment of the present disclosure;
图3为本公开实施例提供的另一种包含动态背景的动画生成流程示意图;FIG. 3 is a schematic diagram of another animation generation process including a dynamic background provided by an embodiment of the present disclosure;
图4为本公开实施例提供的另一种包含动态背景的动画生成流程示意图;FIG. 4 is a schematic diagram of another animation generation process including a dynamic background provided by an embodiment of the present disclosure;
图5为本公开实施例提供的包含动态背景的动画生成装置结构示意图;FIG. 5 is a schematic structural diagram of an animation generating device including a dynamic background provided by an embodiment of the disclosure;
图6为本公开实施例提供的电子设备示意图。FIG. 6 is a schematic diagram of an electronic device provided by an embodiment of the disclosure.
具体实施方式detailed description
下面结合附图对本公开实施例进行详细描述。The embodiments of the present disclosure will be described in detail below with reference to the accompanying drawings.
以下通过特定的具体实例说明本公开的实施方式,本领域技术人员可由本说明书所揭露的内容轻易地了解本公开的其他优点与功效。显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。本公开还可以通过另外不同的具体实施方式加以实施或应用,本说明书中的各项细节也可以基于不同观点与应用,在没有背离本公开的精神下进行各种修饰或改变。需说明的是,在不冲突的情况下,以下实施例及实施例中的特征可以相互组合。基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本公开保护的范围。The following describes the implementation of the present disclosure through specific specific examples, and those skilled in the art can easily understand other advantages and effects of the present disclosure from the content disclosed in this specification. Obviously, the described embodiments are only a part of the embodiments of the present disclosure, rather than all the embodiments. The present disclosure can also be implemented or applied through other different specific embodiments, and various details in this specification can also be modified or changed based on different viewpoints and applications without departing from the spirit of the present disclosure. It should be noted that the following embodiments and the features in the embodiments can be combined with each other if there is no conflict. Based on the embodiments in the present disclosure, all other embodiments obtained by those of ordinary skill in the art without creative work shall fall within the protection scope of the present disclosure.
需要说明的是,下文描述在所附权利要求书的范围内的实施例的各种方面。应显而易见,本文中所描述的方面可体现于广泛多种形式中,且本文中所描述的任何特定结构及/或功能仅为说明性的。基于本公开,所属领域的技术人员应了解,本文中所描述的一个方面可与任何其它方面独立地实施,且可以各种方式组合这些方面中的两者或两者以上。举例来说,可使用本文中所阐述的任何数目个方面来实施设备及/或实践方法。另外,可使用除了本文中所阐述的方面中的一或多者之外的其它结构及/或功能性实施此设备及/或 实践此方法。It should be noted that various aspects of the embodiments within the scope of the appended claims are described below. It should be obvious that the aspects described herein can be embodied in a wide variety of forms, and any specific structure and/or function described herein are only illustrative. Based on the present disclosure, those skilled in the art should understand that one aspect described herein can be implemented independently of any other aspects, and two or more of these aspects can be combined in various ways. For example, any number of aspects set forth herein can be used to implement devices and/or methods of practice. In addition, structures and/or functionalities other than one or more of the aspects set forth herein may be used to implement this device and/or practice this method.
还需要说明的是,以下实施例中所提供的图示仅以示意方式说明本公开的基本构想,图式中仅显示与本公开中有关的组件而非按照实际实施时的组件数目、形状及尺寸绘制,其实际实施时各组件的型态、数量及比例可为一种随意的改变,且其组件布局型态也可能更为复杂。It should also be noted that the illustrations provided in the following embodiments merely illustrate the basic idea of the present disclosure in a schematic manner. The figures only show the components related to the present disclosure rather than the number, shape, and shape of the components in actual implementation. For size drawing, the type, quantity, and ratio of each component can be changed at will during actual implementation, and the component layout type may also be more complicated.
另外,在以下描述中,提供具体细节是为了便于透彻理解实例。然而,所属领域的技术人员将理解,可在没有这些特定细节的情况下实践所述方面。In addition, in the following description, specific details are provided to facilitate a thorough understanding of the examples. However, those skilled in the art will understand that the described aspects may be practiced without these specific details.
本公开实施例提供一种包含动态背景的动画生成方法。本实施例提供的包含动态背景的动画生成方法可以由一计算装置来执行,该计算装置可以实现为软件,或者实现为软件和硬件的组合,该计算装置可以集成设置在服务器、终端设备等中。The embodiment of the present disclosure provides a method for generating an animation including a dynamic background. The method for generating an animation containing a dynamic background provided in this embodiment can be executed by a computing device, which can be implemented as software, or as a combination of software and hardware, and the computing device can be integrated in a server, terminal device, etc. .
参见图1,本公开实施例提供的一种包含动态背景的动画生成方法,包括如下步骤S101-S104:Referring to FIG. 1, an animation generating method including a dynamic background provided by an embodiment of the present disclosure includes the following steps S101-S104:
S101,获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内。S101. Acquire a reconstruction model related to a specific area, a first element, and a second element of a target object, where the specific area belongs to a part of the target object, and the first element and the second element are located in the specific area .
目标对象的动作和表情是本公开的方案所要模拟和预测的内容,作为一个例子,目标对象可以是真实的能够进行网络播报的人,也可以是其他的具有信息传播功能的对象,例如电视节目主持人、新闻节目播音员、进行授课的教师等。The actions and expressions of the target object are the content to be simulated and predicted by the solution of the present disclosure. As an example, the target object can be a real person who can broadcast on the Internet, or other objects with information dissemination functions, such as TV programs. Hosts, news program announcers, teachers who give lectures, etc.
目标对象通常为具有传播功能的人,由于该类型的人通常具有一定的知名度,当有海量的内容需要目标对象进行包含语音和/或视频动作的播报时,通常需要花费较多的成本。同时,对于直播类的节目,目标对象通常无法在同一个时间内出现在多个直播间(或多个直播频道)。此时如果想呈现出例如“主播分身”这样的效果,通过真人进行现场播报通常难以达到这种效果。The target object is usually a person with a communication function. Since this type of person usually has a certain reputation, when there is a large amount of content that the target object needs to broadcast including voice and/or video actions, it usually costs more. At the same time, for live broadcast programs, the target object usually cannot appear in multiple live broadcast rooms (or multiple live channels) at the same time. At this time, if you want to present an effect such as "Anchor Clone", it is usually difficult to achieve this effect through live broadcast by real people.
为此,需要预先训练重构模型,重构模型能够基于输入信息预测目标对象的形象和动作,从而产生与输入信息匹配的动作和表情。例如,针对一段需要播报的新闻,目标对象能够以新闻播报员的形式播报该新闻。For this reason, it is necessary to pre-train the reconstruction model, which can predict the image and actions of the target object based on the input information, thereby generating actions and expressions that match the input information. For example, for a piece of news that needs to be broadcast, the target audience can broadcast the news in the form of a news broadcaster.
为了能够预测出目标对象的动作及表情,通常需要模仿目标对象上的较多组成要素(例如,眼睛,耳朵,眉毛,鼻子,嘴巴等)。但是,预测较多 的组成要素会影响预测的效率,消耗较多的资源。为此,本公开的方案仅选取与目标对象的辨识最相关的组成要素:特定区域、第一元素及第二元素。作为一个例子,特定区域为目标对象的脸部区域,第一元素为目标对象的眼睛,第二元素为目标对象的嘴巴。In order to be able to predict the actions and expressions of the target object, it is usually necessary to imitate more components on the target object (for example, eyes, ears, eyebrows, nose, mouth, etc.). However, components with more predictions will affect the efficiency of predictions and consume more resources. For this reason, the solution of the present disclosure only selects the component elements that are most relevant to the identification of the target object: the specific area, the first element and the second element. As an example, the specific area is the face area of the target object, the first element is the eyes of the target object, and the second element is the mouth of the target object.
S102,基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入信息相关的第一动画。S102. Based on the reconstruction model, determine the texture feature of the specific region, the action of the first element and the action of the second element related to the input information, the texture feature of the specific region, the first element The action of and the action of the second element form a first animation related to the input information.
基于重建模型,可以通过视频动画的方式来预测目标对象在视频中的各种动作和表情。具体的,可以通过产生保真图像的方式来生成包含目标对象动作和表情的视频文件,保真图像可以作为视频文件的全部帧或关键帧,保真图像包含了与所述输入信息相匹配的一个或多个预测动作的多个图像的集合。Based on the reconstruction model, the various actions and expressions of the target object in the video can be predicted through video animation. Specifically, the fidelity image can be used to generate a video file containing the actions and expressions of the target object. The fidelity image can be used as all frames or key frames of the video file, and the fidelity image contains information that matches the input information. A collection of multiple images for one or more predicted actions.
输入信息可以是多种方式,例如,输入信息可以是目标对象在动画展示的过程中需要同时播放一段文字或一段音频内容,即输入信息可以是文本或音频的形式。基于不同的输入信息,目标对象可以产生不同的动画。输入信息通过数据解析后转化为与所述纹理图及所述形状约束图相匹配的参数,利用训练后得到的所述重建模型,通过调用所述纹理图及所述形状约束图来最终完成保证图像的生成。The input information can be in multiple ways. For example, the input information can be that the target object needs to play a piece of text or a piece of audio content at the same time during the animation display process, that is, the input information can be in the form of text or audio. Based on different input information, the target object can produce different animations. After data analysis, the input information is transformed into parameters matching the texture map and the shape constraint map, and the reconstruction model obtained after training is used to finally complete the guarantee by calling the texture map and the shape constraint map Image generation.
在进行预测的阶段,可以给出目标对象的特定区域纹理图和第一元素、第二元素的形状约束,利用所训练的重建模型,预测二维主播形象的图像,通过连续的第一元素、第二元素的形状约束和固定的特定区域的纹理作为输入,预测出连续的目标对象动画形象。In the stage of prediction, the texture map of the specific area of the target object and the shape constraints of the first element and the second element can be given, and the trained reconstruction model can be used to predict the image of the two-dimensional anchor image. Through the continuous first element, The shape constraint of the second element and the fixed texture of the specific area are used as input to predict the continuous animation image of the target object.
S103,在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画。S103: While generating the first animation, dynamically select an animation matching the first animation from a plurality of preset animations as a background animation of the first animation.
为了提高生成画面的真实度,预先录制与目标对象相关的背景动画,背景动画可以是多种不同类型的背景动画,例如,可以是严肃、活泼、高兴、悲伤等多种不同风格的动画。背景动画具有一定的长度,以便于被调用。In order to improve the realism of the generated screen, the background animation related to the target object is recorded in advance. The background animation can be a variety of different types of background animation, for example, it can be serious, lively, happy, sad, and other different styles of animation. The background animation has a certain length so that it can be called easily.
在调用背景动画时,需要对输入信息的当前情景进行分析,基于分析的结果,从预先设置的多个动画中动态选择选择与当前情景匹配的动画作为背 景动画。例如,分析得知当前输入信息的情景是兴奋,则调用兴奋风格的背景动画。当侦测到输入信息的情景发生变化时,即便是当前调用的背景动画没有播放完毕,也会切换调用与变化后的情景匹配的背景动画。当当前调用的背景动画播放完毕后,若发现输入信息的当前情景没有发生变化,则继续循环播放当前的背景动画。When calling the background animation, it is necessary to analyze the current scene of the input information. Based on the results of the analysis, dynamically select the animation that matches the current scene from the preset multiple animations as the background animation. For example, if the analysis knows that the current input information is excited, then the excited background animation is called. When a change in the scene of the input information is detected, even if the currently called background animation has not been played, it will switch to call the background animation matching the changed scene. After the currently called background animation is played, if it is found that the current scene of the input information has not changed, the current background animation will continue to be played in a loop.
S104,基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。S104: Generate a final animation related to the input information based on the first animation and the background animation.
通过将预测产生的第一动画与背景动画进行叠加,便生成了与所述输入信息相关的最终动画。最终动画可以以视频文件的方式进行存储或传播。By superimposing the predicted first animation and the background animation, the final animation related to the input information is generated. The final animation can be stored or disseminated in the form of video files.
在生成最终动画的过程中,可能会存在背景动画进行切换的情况,对此,可以判断所述背景动画是否有多个不同类型的动画拼接组成,若是,则对不同类型的动画进行平滑处理。使作为背景动画的不同类型的动画之间的过渡衔接更加自然。In the process of generating the final animation, there may be situations where the background animation is switched. For this, it can be determined whether the background animation is composed of multiple different types of animation splicing, and if so, smooth processing is performed on the different types of animation. Make the transition between different types of animation as background animation more natural.
重构模型通过神经网络训练得到,为此,在获取与目标对象的特定区域、第一元素及第二元素相关的重构模型之前,还需要采集包含目标对象的多个图像,基于所述多个图像,训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型。The reconstruction model is obtained through neural network training. For this reason, before obtaining the reconstruction model related to the specific area, the first element and the second element of the target object, it is also necessary to collect multiple images containing the target object. Images, training a reconstruction model related to the specific area, the first element and the second element of the target object.
参见图2,可以通过根据本公开实施例的一种具体实现方式,训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,可以包括:Referring to FIG. 2, training a reconstruction model related to a specific region, first element, and second element of the target object can be performed through a specific implementation manner according to an embodiment of the present disclosure, which can include:
S201,对所述多个图像上的特定区域进行检测,得到目标区域。S201: Detect specific areas on the multiple images to obtain a target area.
可以采用多个图像进行目标匹配,从而得到目标区域,作为一个例子,目标区域可以是人脸区域,此时,可以通过人脸检测的方式检测多个图像上存在的人脸区域。Multiple images can be used for target matching to obtain a target area. As an example, the target area may be a face area. In this case, face areas existing on multiple images can be detected by means of face detection.
S202,对所述目标区域进行3D重建,得到3D区域对象。S202: Perform 3D reconstruction on the target area to obtain a 3D area object.
获取了与目标对象相关的多个图像之后,便可以选择目标对象上的构成对象来对目标对象进行建模。对目标区域进行3D建模可以采用多种方法,例如,可以在所述多个图像平面二维像素值的基础之上,增加深度值,深度值可以基于所述多个图像的亮度值获得。After acquiring multiple images related to the target object, the constituent objects on the target object can be selected to model the target object. Various methods can be used for 3D modeling of the target area. For example, a depth value can be added on the basis of the two-dimensional pixel values of the multiple image planes, and the depth value can be obtained based on the brightness values of the multiple images.
S203,获取所述3D区域对象的三维网格,所述三维网格包含预设坐标值。S203: Acquire a three-dimensional grid of the 3D area object, where the three-dimensional grid includes preset coordinate values.
3D区域对象通过三维网格来描述其具体的位置,为此对三维网格设置具体的坐标值,例如可以通过设置平面二维坐标及空间高度坐标的方式来描述三维网格。The 3D area object uses a three-dimensional grid to describe its specific location. For this purpose, specific coordinate values are set for the three-dimensional grid. For example, the three-dimensional grid can be described by setting two-dimensional plane coordinates and spatial height coordinates.
S204,基于不同三维网格坐标上的像素值,确定所述特定区域的纹理图。S204: Determine a texture map of the specific region based on pixel values on different three-dimensional grid coordinates.
可以将不同三维网格坐标上的像素值连接在一起,形成网格平面,该网格平面形成特定区域的纹理图。The pixel values on different three-dimensional grid coordinates can be connected together to form a grid plane, which forms a texture map of a specific area.
除了对目标区域进行训练之外,还需要对第一元素进行训练,参见图3,根据本公开实施例的一种具体实现方式,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:In addition to training the target area, it is also necessary to train the first element. Referring to FIG. 3, according to a specific implementation of an embodiment of the present disclosure, the training is related to the specific area of the target object, the first element and Refactored models related to the second element, including:
S301,在所述多个图像上对所述第一元素进行特征点检测。S301: Perform feature point detection on the first element on the multiple images.
对第一元素进行特征设置,可以得到与第一元素相关的特征点。以眼睛为例,可以将眼睛的轮廓、瞳孔作为特征对象,通过特征检测的方式获得与眼睛轮廓、瞳孔相关的特征点。Perform feature setting on the first element to obtain feature points related to the first element. Taking the eye as an example, the contour and pupil of the eye can be used as feature objects, and feature points related to the contour and pupil of the eye can be obtained through feature detection.
S302,将检测得到的特征点划分为第一类型特征点和第二类型特征点,所述第一类型特征点用于形成第一闭合区域,所述第二类型特征点用于形成第二闭合区域。S302. Divide the detected feature points into a first type feature point and a second type feature point, the first type feature point is used to form a first closed area, and the second type feature point is used to form a second closed area. area.
第一元素为至少包含两个组成部分(例如,瞳孔和巩膜),基于特征点所属的部分不同,将检测得到的特征点划分为第一类型特征点和第二类型特征点,例如,第一类型特征点是与瞳孔相关的特征点,第二类型的特征点是与巩膜相关的特征点。通过将第一类型特征点和第二类型的特征点形成闭合曲线,可以分别形成第一闭合区域和第二闭合区域。The first element is composed of at least two components (for example, pupil and sclera). Based on the different parts to which the feature points belong, the detected feature points are divided into first type feature points and second type feature points, for example, the first Type feature points are feature points related to the pupil, and the second type feature points are feature points related to the sclera. By forming the first type feature points and the second type feature points into a closed curve, the first closed area and the second closed area can be formed respectively.
S303,在所述第一闭合区域内填充第一颜色,在所述第二闭合区域内填充第二颜色,所述第一颜色与所述第二颜色不同。S303: Fill a first color in the first closed area, and a second color in the second closed area, where the first color is different from the second color.
例如,可以在第一闭合区域内填充蓝色,在第二闭合区域内填充白色,通过在不同的闭合区域内填充不同的第一颜色和第二颜色,使得预测后的目标对象的形象更加的逼真。For example, you can fill blue in the first closed area and white in the second closed area. By filling different first and second colors in different closed areas, the image of the predicted target object can be more improved. lifelike.
对于仅有一个组成部分的第二元素,可以采用简单的处理方式,从而提高数据处理的效率。参见图4,根据本公开实施例的一种具体实现方式,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:For the second element with only one component, a simple processing method can be used to improve the efficiency of data processing. Referring to FIG. 4, according to a specific implementation of the embodiment of the present disclosure, the training of a reconstruction model related to a specific area, a first element, and a second element of the target object includes:
S401,在所述多个图像上对所述第二元素进行特征点检测。S401: Perform feature point detection on the second element on the multiple images.
对第二元素进行特征设置,可以检测到与第二元素相关的特征点,以嘴巴为例,可以将嘴巴整体作为特征对象,通过特征检测的方式获得嘴巴相关的特征点。The feature setting of the second element can detect the feature points related to the second element. Taking the mouth as an example, the entire mouth can be used as a feature object, and the feature points related to the mouth can be obtained through feature detection.
S402,基于检测到的所有特征点,形成第三闭合区域。S402: Form a third closed area based on all the detected feature points.
在获得第二元素的所有特征点之后,可以通过将所有特征点连接在一起的方式,形成闭合区域,进而得到第三闭合区域。After all the feature points of the second element are obtained, a closed area can be formed by connecting all the feature points together, and then a third closed area can be obtained.
S403,在所述第三闭合区域内填充第三颜色。S403, filling a third color in the third closed area.
在得到第三闭合区域之后,可以在第三闭合区域内填充与第二元素匹配的颜色,以嘴巴为例,可以在第三闭合区域内填充与嘴巴匹配的红色。After the third closed area is obtained, the third closed area can be filled with a color matching the second element. Taking the mouth as an example, the third closed area can be filled with red that matches the mouth.
在实际对目标对象进行预测的过程中,将所述目标对象的特定区域的轮廓进行预测,并在预测的轮廓内填充所述重构模型确定的纹理图,从而得到了目标对象特定区域的轮廓图(例如,人脸轮廓图)。In the process of actually predicting the target object, the contour of the specific area of the target object is predicted, and the texture map determined by the reconstruction model is filled in the predicted contour, thereby obtaining the contour of the specific area of the target object Figure (e.g., face contour map).
将所述输入信息解析后的得到运动参数匹配到所述第一元素及第二元素,便形成了第一元素和第二元素的动作。具体的,输入信息可以是多种方式,例如,输入信息可以是文本或音频的形式。输入信息通过数据解析后转化为第一解析结果,该第一解析结果包含与纹理图及第一元素、第二元素的形状约束图相匹配的参数,利用训练后得到的所述重建模型,通过调用所述纹理图及所述形状约束图来最终完成目标对象预测动作的生成。The motion parameters obtained after parsing the input information are matched to the first element and the second element, and the actions of the first element and the second element are formed. Specifically, the input information may be in multiple ways, for example, the input information may be in the form of text or audio. The input information is transformed into a first analysis result after data analysis. The first analysis result includes parameters matching the texture map and the shape constraint map of the first element and the second element. The reconstruction model obtained after training is used to pass The texture map and the shape constraint map are called to finally complete the generation of the target object prediction action.
第一解析结果中包含有针对目标对象上第一元素和第二元素的运动幅度参数,以嘴巴为例,嘴巴全部张开的时候运动幅度可以量化为1,嘴巴全部闭合的时候,运动运动幅度可以量化为0,通过量化0和1之间的数值,可以描述嘴巴在完全张开和完全闭合之间的中间状态。The first analysis result contains the motion amplitude parameters for the first element and the second element on the target object. Take the mouth as an example. When the mouth is fully opened, the motion amplitude can be quantified as 1. When the mouth is fully closed, the motion amplitude It can be quantified as 0. By quantifying the value between 0 and 1, the intermediate state of the mouth between fully opened and fully closed can be described.
与上面的方法实施例相对应,参见图5,本公开实施例还公开了一种包含动态背景的动画生成装置50,包括:Corresponding to the above method embodiment, referring to FIG. 5, an embodiment of the present disclosure also discloses an animation generating device 50 containing a dynamic background, including:
获取模块501,用于获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内。The obtaining module 501 is configured to obtain a reconstruction model related to a specific area, a first element, and a second element of a target object, the specific area belongs to a part of the target object, and the first element and the second element are located in the target object. Said within a specific area.
目标对象的动作和表情是本公开的方案所要模拟和预测的内容,作为一个例子,目标对象可以是真实的能够进行网络播报的人,也可以是其他的具 有信息传播功能的对象,例如电视节目主持人、新闻节目播音员、进行授课的教师等。The actions and expressions of the target object are the content to be simulated and predicted by the solution of the present disclosure. As an example, the target object can be a real person who can broadcast on the Internet, or other objects with information dissemination functions, such as TV programs. Hosts, news program announcers, teachers who give lectures, etc.
目标对象通常为具有传播功能的人,由于该类型的人通常具有一定的知名度,当有海量的内容需要目标对象进行包含语音和/或视频动作的播报时,通常需要花费较多的成本。同时,对于直播类的节目,目标对象通常无法在同一个时间内出现在多个直播间(或多个直播频道)。此时如果想呈现出例如“主播分身”这样的效果,通过真人进行现场播报通常难以达到这种效果。The target object is usually a person with a communication function. Since this type of person usually has a certain reputation, when there is a large amount of content that the target object needs to broadcast including voice and/or video actions, it usually costs more. At the same time, for live broadcast programs, the target object usually cannot appear in multiple live broadcast rooms (or multiple live channels) at the same time. At this time, if you want to present an effect such as "Anchor Clone", it is usually difficult to achieve this effect through live broadcast by real people.
为此,需要预先训练重构模型,重构模型能够基于输入信息预测目标对象的形象和动作,从而产生与输入信息匹配的动作和表情。例如,针对一段需要播报的新闻,目标对象能够以新闻播报员的形式播报该新闻。For this reason, it is necessary to pre-train the reconstruction model, which can predict the image and actions of the target object based on the input information, thereby generating actions and expressions that match the input information. For example, for a piece of news that needs to be broadcast, the target audience can broadcast the news in the form of a news broadcaster.
为了能够预测出目标对象的动作及表情,通常需要模仿目标对象上的较多组成要素(例如,眼睛,耳朵,眉毛,鼻子,嘴巴等)。但是,预测较多的组成要素会影响预测的效率,消耗较多的资源。为此,本公开的方案仅选取与目标对象的辨识最相关的组成要素:特定区域、第一元素及第二元素。作为一个例子,特定区域为目标对象的脸部区域,第一元素为目标对象的眼睛,第二元素为目标对象的嘴巴。In order to be able to predict the actions and expressions of the target object, it is usually necessary to imitate more components on the target object (for example, eyes, ears, eyebrows, nose, mouth, etc.). However, the more predictive components will affect the efficiency of the forecast and consume more resources. For this reason, the solution of the present disclosure only selects the component elements that are most relevant to the identification of the target object: the specific area, the first element and the second element. As an example, the specific area is the face area of the target object, the first element is the eyes of the target object, and the second element is the mouth of the target object.
确定模块502,用于基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入信息相关的第一动画。The determining module 502 is configured to determine, based on the reconstruction model, the texture feature of the specific area, the action of the first element and the action of the second element related to the input information, the texture feature of the specific area, the The action of the first element and the action of the second element form a first animation related to the input information.
基于重建模型,可以通过视频动画的方式来预测目标对象在视频中的各种动作和表情。具体的,可以通过产生保真图像的方式来生成包含目标对象动作和表情的视频文件,保真图像可以作为视频文件的全部帧或关键帧,保真图像包含了与所述输入信息相匹配的一个或多个预测动作的多个图像的集合。Based on the reconstruction model, the various actions and expressions of the target object in the video can be predicted through video animation. Specifically, the fidelity image can be used to generate a video file containing the actions and expressions of the target object. The fidelity image can be used as all frames or key frames of the video file, and the fidelity image contains information that matches the input information. A collection of multiple images for one or more predicted actions.
输入信息可以是多种方式,例如,输入信息可以是文本或音频的形式。输入信息通过数据解析后转化为与所述纹理图及所述形状约束图相匹配的参数,利用训练后得到的所述重建模型,通过调用所述纹理图及所述形状约束图来最终完成保证图像的生成。The input information can be in a variety of ways. For example, the input information can be in the form of text or audio. After data analysis, the input information is transformed into parameters matching the texture map and the shape constraint map, and the reconstruction model obtained after training is used to finally complete the guarantee by calling the texture map and the shape constraint map Image generation.
在进行预测的阶段,可以给出目标对象的特定区域纹理图和第一元素、 第二元素的形状约束,利用所训练的重建模型,预测二维主播形象的图像,通过连续的第一元素、第二元素的形状约束和固定的特定区域的纹理作为输入,预测出连续的目标对象动画形象。In the stage of prediction, the texture map of the specific area of the target object and the shape constraints of the first element and the second element can be given, and the trained reconstruction model can be used to predict the image of the two-dimensional anchor image through the continuous first element, The shape constraint of the second element and the fixed texture of the specific area are used as input to predict the continuous animation image of the target object.
选择模块503,用于在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画。The selection module 503 is configured to dynamically select an animation matching the first animation from a plurality of preset animations as the background animation of the first animation while generating the first animation.
为了提高生成画面的真实度,预先录制与目标对象相关的背景动画,背景动画可以是多种不同类型的背景动画,例如,可以是严肃、活泼、高兴、悲伤等多种不同风格的动画。背景动画具有一定的长度,以便于被调用。In order to improve the realism of the generated screen, the background animation related to the target object is recorded in advance. The background animation can be a variety of different types of background animation, for example, it can be serious, lively, happy, sad, and other different styles of animation. The background animation has a certain length so that it can be called easily.
在调用背景动画时,需要对输入信息的当前情景进行分析,基于分析的结果,从预先设置的多个动画中动态选择选择与当前情景匹配的动画作为背景动画。例如,分析得知当前输入信息的情景是兴奋,则调用兴奋风格的背景动画。当侦测到输入信息的情景发生变化时,即便是当前调用的背景动画没有播放完毕,也会切换调用与变化后的情景匹配的背景动画。当当前调用的背景动画播放完毕后,若发现输入信息的当前情景没有发生变化,则继续循环播放当前的背景动画。When calling the background animation, it is necessary to analyze the current scene of the input information, and based on the analysis result, dynamically select an animation that matches the current scene from a plurality of preset animations as the background animation. For example, if the analysis knows that the current input information is excited, then the excited background animation is called. When a change in the scene of the input information is detected, even if the currently called background animation has not been played, it will switch to call the background animation matching the changed scene. After the currently called background animation is played, if it is found that the current scene of the input information has not changed, the current background animation will continue to be played in a loop.
生成模块504,用于基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。The generating module 504 is configured to generate a final animation related to the input information based on the first animation and the background animation.
通过将预测产生的第一动画与背景动画进行叠加,便生成了与所述输入信息相关的最终动画。最终动画可以以视频文件的方式进行存储或传播。By superimposing the predicted first animation and the background animation, the final animation related to the input information is generated. The final animation can be stored or disseminated in the form of video files.
在生成最终动画的过程中,可能会存在背景动画进行切换的情况,对此,可以判断所述背景动画是否有多个不同类型的动画拼接组成,若是,则对不同类型的动画进行平滑处理。使得不同类型的动画之间的过渡衔接更加自然。In the process of generating the final animation, there may be situations where the background animation is switched. For this, it can be determined whether the background animation is composed of multiple different types of animation splicing, and if so, smooth processing is performed on the different types of animation. Make the transition between different types of animation more natural.
图5所示装置可以对应的执行上述方法实施例中的内容,本实施例未详细描述的部分,参照上述方法实施例中记载的内容,在此不再赘述。The device shown in FIG. 5 can correspondingly execute the content in the foregoing method embodiment. For parts that are not described in detail in this embodiment, refer to the content recorded in the foregoing method embodiment, which will not be repeated here.
参见图6,本公开实施例还提供了一种电子设备60,该电子设备包括:Referring to FIG. 6, an embodiment of the present disclosure also provides an electronic device 60, which includes:
至少一个处理器;以及,At least one processor; and,
与该至少一个处理器通信连接的存储器;其中,A memory communicatively connected with the at least one processor; wherein,
该存储器存储有可被该至少一个处理器执行的指令,该指令被该至少一个处理器执行,以使该至少一个处理器能够执行前述方法实施例中包含动态背景的动画生成方法。The memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute the animation generation method including the dynamic background in the foregoing method embodiment.
本公开实施例还提供了一种非暂态计算机可读存储介质,该非暂态计算机可读存储介质存储计算机指令,该计算机指令用于使该计算机执行前述方法实施例中。The embodiments of the present disclosure also provide a non-transitory computer-readable storage medium that stores computer instructions, and the computer instructions are used to make the computer execute the foregoing method embodiments.
本公开实施例还提供了一种计算机程序产品,该计算机程序产品包括存储在非暂态计算机可读存储介质上的计算程序,该计算机程序包括程序指令,当该程序指令被计算机执行时,使该计算机执行前述方法实施例中的包含动态背景的动画生成方法。The embodiments of the present disclosure also provide a computer program product, the computer program product includes a calculation program stored on a non-transitory computer-readable storage medium, the computer program includes program instructions, when the program instructions are executed by a computer, The computer executes the animation generation method including the dynamic background in the foregoing method embodiment.
下面参考图6,其示出了适于用来实现本公开实施例的电子设备60的结构示意图。本公开实施例中的电子设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图6示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。Reference is now made to FIG. 6, which shows a schematic structural diagram of an electronic device 60 suitable for implementing embodiments of the present disclosure. Electronic devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablets), PMPs (portable multimedia players), vehicle-mounted terminals (for example, Mobile terminals such as car navigation terminals) and fixed terminals such as digital TVs, desktop computers, etc. The electronic device shown in FIG. 6 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present disclosure.
如图6所示,电子设备60可以包括处理装置(例如中央处理器、图形处理器等)601,其可以根据存储在只读存储器(ROM)602中的程序或者从存储装置608加载到随机访问存储器(RAM)603中的程序而执行各种适当的动作和处理。在RAM 603中,还存储有电子设备60操作所需的各种程序和数据。处理装置601、ROM 602以及RAM 603通过总线604彼此相连。输入/输出(I/O)接口605也连接至总线604。As shown in FIG. 6, the electronic device 60 may include a processing device (such as a central processing unit, a graphics processor, etc.) 601, which may be loaded into a random access device according to a program stored in a read-only memory (ROM) 602 or from a storage device 608. The program in the memory (RAM) 603 executes various appropriate actions and processing. The RAM 603 also stores various programs and data required for the operation of the electronic device 60. The processing device 601, the ROM 602, and the RAM 603 are connected to each other through a bus 604. An input/output (I/O) interface 605 is also connected to the bus 604.
通常,以下装置可以连接至I/O接口605:包括例如触摸屏、触摸板、键盘、鼠标、图像传感器、麦克风、加速度计、陀螺仪等的输入装置606;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置607;包括例如磁带、硬盘等的存储装置608;以及通信装置609。通信装置609可以允许电子设备60与其他设备进行无线或有线通信以交换数据。虽然图中示出了具有各种装置的电子设备60,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。Generally, the following devices can be connected to the I/O interface 605: including input devices 606 such as touch screen, touch panel, keyboard, mouse, image sensor, microphone, accelerometer, gyroscope, etc.; including, for example, liquid crystal display (LCD), speakers, An output device 607 such as a vibrator; a storage device 608 such as a magnetic tape, a hard disk, etc.; and a communication device 609. The communication device 609 may allow the electronic device 60 to perform wireless or wired communication with other devices to exchange data. Although the figure shows the electronic device 60 having various devices, it should be understood that it is not required to implement or have all the illustrated devices. It may alternatively be implemented or provided with more or fewer devices.
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信 装置609从网络上被下载和安装,或者从存储装置608被安装,或者从ROM602被安装。在该计算机程序被处理装置601执行时,执行本公开实施例的方法中限定的上述功能。In particular, according to an embodiment of the present disclosure, the process described above with reference to the flowchart can be implemented as a computer software program. For example, the embodiments of the present disclosure include a computer program product, which includes a computer program carried on a computer-readable medium, and the computer program contains program code for executing the method shown in the flowchart. In such an embodiment, the computer program may be downloaded and installed from the network through the communication device 609, or installed from the storage device 608, or installed from the ROM602. When the computer program is executed by the processing device 601, the above-mentioned functions defined in the method of the embodiment of the present disclosure are executed.
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。It should be noted that the above-mentioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the two. The computer-readable storage medium may be, for example, but not limited to, an electric, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the above. More specific examples of computer-readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above. In the present disclosure, a computer-readable storage medium may be any tangible medium that contains or stores a program, and the program may be used by or in combination with an instruction execution system, apparatus, or device. In the present disclosure, a computer-readable signal medium may include a data signal propagated in a baseband or as a part of a carrier wave, and a computer-readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing. The computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium. The computer-readable signal medium may send, propagate, or transmit the program for use by or in combination with the instruction execution system, apparatus, or device . The program code contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to: wire, optical cable, RF (Radio Frequency), etc., or any suitable combination of the above.
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。The above-mentioned computer-readable medium may be included in the above-mentioned electronic device; or it may exist alone without being assembled into the electronic device.
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:获取至少两个网际协议地址;向节点评价设备发送包括所述至少两个网际协议地址的节点评价请求,其中,所述节点评价设备从所述至少两个网际协议地址中,选取网际协议地址并返回;接收所述节点评价设备返回的网际协议地址;其中,所获取的网际协议地址指示内容分发网络中的边缘节点。The above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device: obtains at least two Internet protocol addresses; and sends to the node evaluation device including the at least two A node evaluation request for an Internet Protocol address, wherein the node evaluation device selects an Internet Protocol address from the at least two Internet Protocol addresses and returns it; receives the Internet Protocol address returned by the node evaluation device; wherein, the obtained The Internet Protocol address indicates the edge node in the content distribution network.
或者,上述计算机可读介质承载有一个或者多个程序,当上述一个或者 多个程序被该电子设备执行时,使得该电子设备:接收包括至少两个网际协议地址的节点评价请求;从所述至少两个网际协议地址中,选取网际协议地址;返回选取出的网际协议地址;其中,接收到的网际协议地址指示内容分发网络中的边缘节点。Alternatively, the aforementioned computer-readable medium carries one or more programs, and when the aforementioned one or more programs are executed by the electronic device, the electronic device: receives a node evaluation request including at least two Internet Protocol addresses; Among the at least two Internet Protocol addresses, select an Internet Protocol address; return the selected Internet Protocol address; wherein, the received Internet Protocol address indicates an edge node in the content distribution network.
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。The computer program code used to perform the operations of the present disclosure may be written in one or more programming languages or a combination thereof. The above-mentioned programming languages include object-oriented programming languages—such as Java, Smalltalk, C++, and also conventional Procedural programming language-such as "C" language or similar programming language. The program code can be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server. In the case of a remote computer, the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, using an Internet service provider to pass Internet connection).
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。The flowcharts and block diagrams in the accompanying drawings illustrate the possible implementation architecture, functions, and operations of the system, method, and computer program product according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagram can represent a module, program segment, or part of code, and the module, program segment, or part of code contains one or more for realizing the specified logical function Executable instructions. It should also be noted that, in some alternative implementations, the functions marked in the block may also occur in a different order from the order marked in the drawings. For example, two blocks shown in succession can actually be executed substantially in parallel, or they can sometimes be executed in the reverse order, depending on the functions involved. It should also be noted that each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart, can be implemented by a dedicated hardware-based system that performs the specified functions or operations Or it can be realized by a combination of dedicated hardware and computer instructions.
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。The units involved in the embodiments described in the present disclosure may be implemented in a software manner, or may be implemented in a hardware manner. Wherein, the name of the unit does not constitute a limitation on the unit itself under certain circumstances. For example, the first obtaining unit can also be described as "a unit for obtaining at least two Internet Protocol addresses."
应当理解,本公开的各部分可以用硬件、软件、固件或它们的组合来实现。It should be understood that each part of the present disclosure can be implemented by hardware, software, firmware or a combination thereof.
以上所述,仅为本公开的具体实施方式,但本公开的保护范围并不局限 于此,任何熟悉本技术领域的技术人员在本公开揭露的技术范围内,可轻易想到的变化或替换,都应涵盖在本公开的保护范围之内。因此,本公开的保护范围应以权利要求的保护范围为准。The above are only specific implementations of the present disclosure, but the protection scope of the present disclosure is not limited thereto. Any person skilled in the art can easily think of changes or substitutions within the technical scope disclosed in the present disclosure. All should be covered within the protection scope of this disclosure. Therefore, the protection scope of the present disclosure should be subject to the protection scope of the claims.

Claims (12)

  1. 一种包含动态背景的动画生成方法,其特征在于,包括:A method for generating animation containing dynamic background, characterized in that it comprises:
    获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内;Acquiring a reconstruction model related to a specific area, a first element, and a second element of a target object, the specific area belongs to a part of the target object, and the first element and the second element are located in the specific area;
    基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入信息相关的第一动画;Based on the reconstruction model, determine the texture feature of the specific area related to the input information, the action of the first element and the action of the second element, the texture feature of the specific area, the action of the first element And the action of the second element forms a first animation related to the input information;
    在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画;While generating the first animation, dynamically selecting an animation matching the first animation from a plurality of preset animations as the background animation of the first animation;
    基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。Based on the first animation and the background animation, a final animation related to the input information is generated.
  2. 根据权利要求1所述的方法,其特征在于,获取与目标对象的特定区域、第一元素及第二元素相关的重构模型之前,所述方法还包括:The method according to claim 1, wherein before obtaining the reconstruction model related to the specific area, the first element and the second element of the target object, the method further comprises:
    采集包含目标对象的多个图像,基于所述多个图像,训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型。A plurality of images including the target object are acquired, and based on the plurality of images, a reconstruction model related to the specific area, the first element and the second element of the target object is trained.
  3. 根据权利要求2所述的方法,其特征在于,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:The method according to claim 2, wherein the training a reconstruction model related to the specific area, the first element and the second element of the target object comprises:
    对所述多个图像上的特定区域进行检测,得到目标区域;Detecting specific areas on the multiple images to obtain a target area;
    对所述目标区域进行3D重建,得到3D区域对象;Performing 3D reconstruction on the target area to obtain a 3D area object;
    获取所述3D区域对象的三维网格,所述三维网格包含预设坐标值;Acquiring a three-dimensional grid of the 3D area object, where the three-dimensional grid includes preset coordinate values;
    基于不同三维网格坐标上的像素值,确定所述特定区域的纹理图。Based on the pixel values on different three-dimensional grid coordinates, the texture map of the specific area is determined.
  4. 根据权利要求2所述的方法,其特征在于,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:The method according to claim 2, wherein the training a reconstruction model related to the specific area, the first element and the second element of the target object comprises:
    在所述多个图像上对所述第一元素进行特征点检测;Performing feature point detection on the first element on the multiple images;
    将检测得到的特征点划分为第一类型特征点和第二类型特征点,所述第一类型特征点用于形成第一闭合区域,所述第二类型特征点用于形成第二闭合区域;Divide the detected feature points into a first type feature point and a second type feature point, the first type feature point is used to form a first closed area, and the second type feature point is used to form a second closed area;
    在所述第一闭合区域内填充第一颜色,在所述第二闭合区域内填充第二 颜色,所述第一颜色与所述第二颜色不同。A first color is filled in the first closed area, and a second color is filled in the second closed area, and the first color is different from the second color.
  5. 根据权利要求2所述的方法,其特征在于,所述训练与所述目标对象的特定区域、第一元素及第二元素相关的重构模型,包括:The method according to claim 2, wherein the training a reconstruction model related to the specific area, the first element and the second element of the target object comprises:
    在所述多个图像上对所述第二元素进行特征点检测;Performing feature point detection on the second element on the multiple images;
    基于检测到的所有特征点,形成第三闭合区域;Based on all the detected feature points, a third closed area is formed;
    在所述第三闭合区域内填充第三颜色。A third color is filled in the third closed area.
  6. 根据权利要求3所述的方法,其特征在于,所述基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,包括:The method according to claim 3, wherein the determining the texture feature of the specific area related to the input information, the action of the first element and the action of the second element based on the reconstruction model, include:
    将所述目标对象的特定区域的轮廓进行预测,在预测的轮廓内填充所述重构模型确定的纹理图;Predicting the contour of the specific area of the target object, and filling the predicted contour with the texture map determined by the reconstruction model;
    将所述输入信息解析后的得到运动参数匹配到所述第一元素及第二元素,形成所述第一元素及所述第二元素的动作。The motion parameters obtained after parsing the input information are matched to the first element and the second element to form an action of the first element and the second element.
  7. 根据权利要求1所述的方法,其特征在于,所述从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画,包括:The method according to claim 1, wherein the dynamically selecting an animation matching the first animation from a plurality of preset animations as the background animation of the first animation comprises:
    对输入信息的当前情景进行分析,从预先设置的多个动画中动态选择选择与当前情景匹配的动画作为背景动画。Analyze the current scene of the input information, dynamically select an animation that matches the current scene from a plurality of preset animations as the background animation.
  8. 根据权利要求1所述的方法,其特征在于,所述基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画,包括:The method according to claim 1, wherein the generating a final animation related to the input information based on the first animation and the background animation comprises:
    判断所述背景动画是否有多个不同类型的动画拼接组成;Judging whether the background animation is composed of multiple animations of different types;
    若是,则对不同类型的动画进行平滑处理。If it is, smooth the different types of animations.
  9. 根据权利要求1所述的方法,其特征在于:The method according to claim 1, wherein:
    所述特定区域为脸部区域,所述第一元素为眼睛,所述第二元素为嘴巴。The specific area is a face area, the first element is an eye, and the second element is a mouth.
  10. 一种包含动态背景的动画生成装置,其特征在于,包括:An animation generating device containing a dynamic background, characterized in that it comprises:
    获取模块,用于获取与目标对象的特定区域、第一元素及第二元素相关的重构模型,所述特定区域属于所述目标对象的一部分,所述第一元素及第二元素位于所述特定区域内;The acquiring module is used to acquire a reconstruction model related to a specific area, a first element and a second element of a target object, the specific area belongs to a part of the target object, and the first element and the second element are located in the Within a specific area
    确定模块,用于基于所述重构模型,确定与输入信息相关的特定区域的纹理特征、所述第一元素的动作以及所述第二元素的动作,所述特定区域的 纹理特征、所述第一元素的动作以及所述第二元素的动作形成了与所述输入信息相关的第一动画;The determining module is configured to determine, based on the reconstruction model, the texture feature of the specific area related to the input information, the action of the first element and the action of the second element, the texture feature of the specific area, the The actions of the first element and the actions of the second element form a first animation related to the input information;
    选择模块,用于在生成所述第一动画的同时,从预先设置的多个动画中动态选择与所述第一动画匹配的动画作为所述第一动画的背景动画;The selection module is configured to dynamically select an animation matching the first animation from a plurality of preset animations as the background animation of the first animation while generating the first animation;
    生成模块,用于基于所述第一动画及所述背景动画,生成与所述输入信息相关的最终动画。A generating module is used to generate a final animation related to the input information based on the first animation and the background animation.
  11. 一种电子设备,其特征在于,所述电子设备包括:An electronic device, characterized in that, the electronic device includes:
    至少一个处理器;以及,At least one processor; and,
    与所述至少一个处理器通信连接的存储器;其中,A memory communicatively connected with the at least one processor; wherein,
    所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行前述任一权利要求1-9所述的包含动态背景的动画生成方法。The memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor, so that the at least one processor can execute any of the preceding claims 1-9. Including dynamic background animation generation method.
  12. 一种非暂态计算机可读存储介质,该非暂态计算机可读存储介质存储A non-transitory computer-readable storage medium that stores
    计算机指令,该计算机指令用于使该计算机执行前述任一权利要求1-9所述的包含动态背景的动画生成方法。Computer instructions, which are used to make the computer execute the animation generating method including dynamic background according to any one of claims 1-9.
PCT/CN2020/074369 2019-03-20 2020-02-05 Method, apparatus, and electronic device for generating animation containing dynamic background WO2020186934A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910214896.6 2019-03-20
CN201910214896.6A CN110047119B (en) 2019-03-20 2019-03-20 Animation generation method and device comprising dynamic background and electronic equipment

Publications (1)

Publication Number Publication Date
WO2020186934A1 true WO2020186934A1 (en) 2020-09-24

Family

ID=67273996

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/074369 WO2020186934A1 (en) 2019-03-20 2020-02-05 Method, apparatus, and electronic device for generating animation containing dynamic background

Country Status (2)

Country Link
CN (1) CN110047119B (en)
WO (1) WO2020186934A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110047119B (en) * 2019-03-20 2021-04-13 北京字节跳动网络技术有限公司 Animation generation method and device comprising dynamic background and electronic equipment
CN113554734A (en) * 2021-07-19 2021-10-26 深圳东辉盛扬科技有限公司 Animation model generation method and device based on neural network
CN114549706A (en) * 2022-02-21 2022-05-27 成都工业学院 Animation generation method and animation generation device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100759364B1 (en) * 2006-05-02 2007-09-19 한국과학기술원 Composition method of user action pattern realtime graphics and high quality animation
CN103198508A (en) * 2013-04-07 2013-07-10 河北工业大学 Human face expression animation generation method
CN106648071A (en) * 2016-11-21 2017-05-10 捷开通讯科技(上海)有限公司 Social implementation system for virtual reality
CN109285208A (en) * 2018-09-29 2019-01-29 吉林动画学院 Virtual role expression cartooning algorithm based on expression dynamic template library
CN110047119A (en) * 2019-03-20 2019-07-23 北京字节跳动网络技术有限公司 Animation producing method, device and electronic equipment comprising dynamic background

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8937620B1 (en) * 2011-04-07 2015-01-20 Google Inc. System and methods for generation and control of story animation
CN102867333A (en) * 2012-07-18 2013-01-09 西北工业大学 DI-GUY-based virtual character behavior visualization method
CN103854306A (en) * 2012-12-07 2014-06-11 山东财经大学 High-reality dynamic expression modeling method
US11803993B2 (en) * 2017-02-27 2023-10-31 Disney Enterprises, Inc. Multiplane animation system
CN107392984B (en) * 2017-07-26 2020-09-15 厦门美图之家科技有限公司 Method for generating animation based on face image and computing equipment
CN108629821A (en) * 2018-04-20 2018-10-09 北京比特智学科技有限公司 Animation producing method and device
CN109118579A (en) * 2018-08-03 2019-01-01 北京微播视界科技有限公司 The method, apparatus of dynamic generation human face three-dimensional model, electronic equipment
CN109272543B (en) * 2018-09-21 2020-10-02 北京字节跳动网络技术有限公司 Method and apparatus for generating a model
CN109462776B (en) * 2018-11-29 2021-08-20 北京字节跳动网络技术有限公司 Video special effect adding method and device, terminal equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100759364B1 (en) * 2006-05-02 2007-09-19 한국과학기술원 Composition method of user action pattern realtime graphics and high quality animation
CN103198508A (en) * 2013-04-07 2013-07-10 河北工业大学 Human face expression animation generation method
CN106648071A (en) * 2016-11-21 2017-05-10 捷开通讯科技(上海)有限公司 Social implementation system for virtual reality
CN109285208A (en) * 2018-09-29 2019-01-29 吉林动画学院 Virtual role expression cartooning algorithm based on expression dynamic template library
CN110047119A (en) * 2019-03-20 2019-07-23 北京字节跳动网络技术有限公司 Animation producing method, device and electronic equipment comprising dynamic background

Also Published As

Publication number Publication date
CN110047119B (en) 2021-04-13
CN110047119A (en) 2019-07-23

Similar Documents

Publication Publication Date Title
US20210029305A1 (en) Method and apparatus for adding a video special effect, terminal device and storage medium
WO2020186934A1 (en) Method, apparatus, and electronic device for generating animation containing dynamic background
WO2021004247A1 (en) Method and apparatus for generating video cover and electronic device
CN110047121B (en) End-to-end animation generation method and device and electronic equipment
WO2020228383A1 (en) Mouth shape generation method and apparatus, and electronic device
US20200234478A1 (en) Method and Apparatus for Processing Information
JP6936298B2 (en) Methods and devices for controlling changes in the mouth shape of 3D virtual portraits
JP2020064616A (en) Virtual robot interaction method, device, storage medium, and electronic device
CN112492380B (en) Sound effect adjusting method, device, equipment and storage medium
CN110035271B (en) Fidelity image generation method and device and electronic equipment
CN112669417B (en) Virtual image generation method and device, storage medium and electronic equipment
WO2020151491A1 (en) Image deformation control method and device and hardware device
JP7473676B2 (en) AUDIO PROCESSING METHOD, APPARATUS, READABLE MEDIUM AND ELECTRONIC DEVICE
US11587280B2 (en) Augmented reality-based display method and device, and storage medium
JP2022505118A (en) Image processing method, equipment, hardware equipment
WO2022007565A1 (en) Image processing method and apparatus for augmented reality, electronic device and storage medium
US20230162452A1 (en) Implementation and display of augmented reality
US20230131975A1 (en) Music playing method and apparatus based on user interaction, and device and storage medium
CN110288532B (en) Method, apparatus, device and computer readable storage medium for generating whole body image
CN110060324B (en) Image rendering method and device and electronic equipment
WO2024067159A1 (en) Video generation method and apparatus, electronic device, and storage medium
US20230401789A1 (en) Methods and systems for unified rendering of light and sound content for a simulated 3d environment
CN116958344A (en) Animation generation method and device for virtual image, computer equipment and storage medium
CN109816791A (en) Method and apparatus for generating information
CN113628324A (en) Wisdom highway VR interactive teaching system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20774803

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 02/02/2022)

122 Ep: pct application non-entry in european phase

Ref document number: 20774803

Country of ref document: EP

Kind code of ref document: A1