CN113347465A - Video generation method and device, electronic equipment and storage medium - Google Patents

Video generation method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN113347465A
CN113347465A CN202110597266.9A CN202110597266A CN113347465A CN 113347465 A CN113347465 A CN 113347465A CN 202110597266 A CN202110597266 A CN 202110597266A CN 113347465 A CN113347465 A CN 113347465A
Authority
CN
China
Prior art keywords
video
elements
template
production
multimedia
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110597266.9A
Other languages
Chinese (zh)
Other versions
CN113347465B (en
Inventor
唐小辉
张宏伟
段一苗
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN202110597266.9A priority Critical patent/CN113347465B/en
Publication of CN113347465A publication Critical patent/CN113347465A/en
Application granted granted Critical
Publication of CN113347465B publication Critical patent/CN113347465B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/23424Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for inserting or substituting an advertisement
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/812Monomedia components thereof involving advertisement data

Abstract

The application discloses a video generation method, a video generation device, electronic equipment and a storage medium, in the method, a server responds to a video production request of a first terminal and determines video production information at least comprising production elements, wherein the production elements comprise a video template and a first multimedia element, the video template is obtained by combining at least two second multimedia elements in advance, the video template is rendered to obtain an intermediate video corresponding to the video template, and then the intermediate video and the first multimedia element corresponding to the video template are utilized to generate a target video. Therefore, the video template is converted into the intermediate video, the intermediate video belongs to the multimedia elements and can be combined with other multimedia elements, so that a user can freely combine the video template and the first multimedia element to generate the target video, and the richness of the target video is favorably improved.

Description

Video generation method and device, electronic equipment and storage medium
Technical Field
The present application relates to the field of video processing technologies, and in particular, to a video generation method and apparatus, an electronic device, and a storage medium.
Background
The advertising video production has a high technical threshold, and a great deal of effort is required to be invested in processing such as video editing and special effect synthesis. Since the threshold of the video editing technology is high, a method for producing an advertisement video by applying a video template is created.
In the related technology, a user selects a video template meeting the requirement, and modifies video materials in the template according to prompts to generate an advertisement video. However, since the style of the video of the same video template is fixed, the styles of the videos generated by using the same video template are also relatively similar, which easily results in low richness of the videos generated by the user.
Disclosure of Invention
The embodiment of the application provides a video generation method and device, electronic equipment and a storage medium, which are used for solving the problem of low video richness generated by applying a video template in the related art.
According to a first aspect of embodiments of the present application, there is provided a video generation method, including:
responding to a video production request of a first terminal, determining video production information, wherein the video production information at least comprises production elements, the production elements comprise a video template and first multimedia elements, and the video template is obtained by combining at least two second multimedia elements in advance;
rendering the video template to obtain an intermediate video corresponding to the video template;
and generating a target video by using the intermediate video corresponding to the video template and the first multimedia element.
In some possible embodiments, the video production information further includes occurrence segment description information of the production element, and the method further includes:
determining fragment synthesis information according to the occurrence fragment description information of the production elements, wherein the fragment synthesis information at least comprises a video template and/or a first multimedia element which occur in a single fragment;
generating a target video by using the intermediate video corresponding to the video template and the first multimedia element, wherein the generating comprises:
respectively generating video clips according to the intermediate video and/or the first multimedia elements corresponding to the video template appearing in the single clip;
and generating a target video according to the video clips.
In some possible embodiments, the clip composition information further includes video templates and/or first multimedia elements appearing in all clips, and generating the target video from each video clip includes:
synthesizing each video clip to obtain a synthesized video;
and modifying the synthesized video by utilizing the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in all the segments to obtain the target video.
In some possible embodiments, after generating the target video, the method further includes:
and responding to an element editing instruction, and updating the target video, wherein the element editing instruction is an element adding instruction, an element deleting instruction or an element replacing instruction.
In some possible embodiments, after generating the target video or updating the target video, the method further includes:
sending sharable elements corresponding to the target video to the first terminal, and displaying the sharable elements by the first terminal, wherein the sharable elements comprise any combination of the target video, production elements of the target video and video clips in the target video;
responding to an element sharing request of the first terminal, and determining sharable elements selected by the first terminal;
and saving the sharable elements selected by the first terminal.
In some possible embodiments, the method further comprises:
and responding to a manufacturing element acquisition request of a second terminal, selecting at least one sharable element from the stored sharable elements, and sending the at least one sharable element to the second terminal, so that the second terminal generates a new video by using the at least one sharable element.
According to a second aspect of embodiments of the present application, there is provided a video generation apparatus, including:
a response unit configured to execute a video production request in response to a first terminal, and determine video production information, where the video production information at least includes production elements, the production elements include a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance;
the rendering unit is configured to perform rendering on the video template to obtain an intermediate video corresponding to the video template;
and the generating unit is configured to generate a target video by using the intermediate video corresponding to the video template and the first multimedia element.
In some possible embodiments, the video production information further includes occurrence segment description information of the production element, and the apparatus further includes:
a determining unit configured to perform determining clip composition information including at least a video template and/or a first multimedia element appearing in a single clip according to the appearance clip description information of the production element;
the generating unit is specifically configured to execute intermediate video and/or first multimedia elements corresponding to video templates appearing in a single segment, and respectively generate video segments; and generating a target video according to the video clips.
In some possible embodiments, the clip composition information further comprises a video template and/or a first multimedia element occurring at all clips, and
the generating unit is specifically configured to perform synthesis on each video segment to obtain a synthesized video; and modifying the synthesized video by utilizing the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in all the segments to obtain the target video.
In some possible embodiments, the method further comprises:
the updating unit is configured to update the target video in response to an element editing instruction after the target video is generated, wherein the element editing instruction is an element adding instruction, an element deleting instruction or an element replacing instruction.
In some possible embodiments, the method further comprises:
the sending unit is configured to send sharable elements corresponding to a target video to the first terminal after the target video is generated or updated, and the first terminal displays the sharable elements, wherein the sharable elements comprise any combination of the target video, production elements of the target video and video clips in the target video;
the sharing unit is configured to execute element sharing request responding to the first terminal and determine sharable elements selected by the first terminal;
a saving unit configured to perform saving of the sharable element selected by the first terminal.
In some possible embodiments, the method further comprises:
the selecting unit is further configured to execute, in response to a production element acquisition request of a second terminal, selecting at least one sharable element from the stored sharable elements and sending the at least one sharable element to the second terminal, so that the second terminal generates a new video by using the at least one sharable element.
According to a third aspect of embodiments of the present application, there is provided an electronic apparatus, including:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement a video generation method;
according to a fourth aspect of embodiments herein, there is provided a storage medium having instructions that, when executed by a processor of an electronic device, enable the electronic device to perform a video generation method;
according to a fifth aspect of embodiments herein, there is provided a computer program product comprising at least one processor; and a memory communicatively coupled to the at least one processor; the memory stores instructions executable by the at least one processor, and the instructions are executed by the at least one processor to enable the at least one processor to execute the video generation method provided by the embodiment of the application.
The technical scheme provided by the embodiment of the application at least has the following beneficial effects:
in the embodiment of the application, a server responds to a video production request of a first terminal, and determines video production information at least comprising production elements, wherein the production elements comprise a video template and a first multimedia element, the video template is obtained by combining at least two second multimedia elements in advance, the video template is rendered to obtain an intermediate video corresponding to the video template, and then the intermediate video and the first multimedia element corresponding to the video template are utilized to generate a target video. Therefore, the video template is converted into the intermediate video, the intermediate video belongs to the multimedia elements and can be combined with other multimedia elements, so that a user can freely combine the video template and the first multimedia element to generate the target video, and the richness of the target video is favorably improved.
Additional features and advantages of the application will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by the practice of the application. The objectives and other advantages of the application may be realized and attained by the structure particularly pointed out in the written description and claims hereof as well as the appended drawings.
Drawings
The accompanying drawings, which are included to provide a further understanding of the application and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the application and together with the description serve to explain the application and not to limit the application. In the drawings:
fig. 1 is a diagram illustrating an application scenario of a video generation method according to an exemplary embodiment.
Fig. 2 is a flow diagram illustrating a video generation method according to an example embodiment.
Fig. 3 is a flow chart illustrating yet another video generation method according to an exemplary embodiment.
Fig. 4 is a flow chart illustrating yet another video generation method according to an exemplary embodiment.
Fig. 5 is a flow chart illustrating yet another video generation method according to an exemplary embodiment.
Fig. 6 is a flow chart illustrating yet another video generation method according to an exemplary embodiment.
Fig. 7 is a block diagram illustrating a video generation apparatus according to an example embodiment.
Fig. 8 is a schematic structural diagram of an electronic device for implementing a video generation method according to an exemplary embodiment.
Detailed Description
In order to solve the problem of low video richness generated by applying a video template in the related art, embodiments of the present application provide a video generation method, an apparatus, an electronic device, and a storage medium. In order to better understand the technical solution provided by the embodiments of the present application, the following brief description is made on the basic principle of the solution:
it should be noted that the terms "first," "second," and the like in the description and claims of this application and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application described herein are capable of operation in sequences other than those illustrated or described herein. The embodiments described in the following exemplary embodiments do not represent all embodiments consistent with the present application. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present application, as detailed in the appended claims.
The preferred embodiments of the present application will be described below with reference to the accompanying drawings of the specification, it should be understood that the preferred embodiments described herein are merely for illustrating and explaining the present application, and are not intended to limit the present application, and that the embodiments and features of the embodiments in the present application may be combined with each other without conflict.
The data to which the present disclosure relates (e.g., video templates, material, etc. used to produce the video) may be data that is authorized by the user or sufficiently authorized by various parties.
AE (after effects) is a video creation tool, and a user can create an AE template by using AE, and further create a video by using the AE template. However, when the user generates a video by using the AE template, only one AE template can be selected at a time, and only the video material in the AE template can be replaced, but other materials in the AE template cannot be replaced, and the AE template cannot be combined with other production elements. Thus, the video styles generated by different users by using the same AE template are similar, and the video is lack of richness.
The above description is only an AE template, and other video templates such as various Artificial Intelligence (AI) face changing templates have the same problem.
In order to solve the above problems, the present application discloses a scheme capable of freely combining a video template and multimedia elements (such as videos, pictures, music, characters, etc.) to generate a target video, which is not only simple in video production mode and easy for a user to get on hand, but also beneficial to improving the richness of the finally generated target video.
The technical solution of the present application will be described below with reference to the drawings of the specification.
Fig. 1 is a diagram illustrating an application scenario of a video generation method according to an exemplary embodiment, and includes a first terminal, a second terminal, and a server connected to the first terminal and the second terminal through a communication network, where:
the first terminal is used for providing a video production page, wherein various production elements can be displayed in the video production page, and a production element uploading inlet and a video production button can be provided. The user of the first terminal can upload some production elements through the upload entrance, and can freely select production elements which are wanted to be used when the video is produced on the video production page, such as: AE template, AI face-changing template, etc. After the user of the first terminal clicks the video production button, the first terminal may be triggered to send a video production request to the server, where the video production request may include an element identifier of a production element.
The server is used for responding to a video production request sent by the first terminal, determining video production information according to an element identifier included in the production request, wherein the video production information at least comprises a first multimedia element and a video template selected by a user of the first terminal, the first multimedia element is at least one of music, characters, pictures, videos and the like, the video template is obtained by combining at least two second multimedia elements in advance, the second multimedia elements are at least one of music, characters, pictures and the like, and the second multimedia elements may or may not be associated with the first multimedia elements. And then, rendering the video template to obtain an intermediate video corresponding to the video template, and generating the target video by using the intermediate video corresponding to the video template and the first multimedia element.
Therefore, the user of the first terminal can freely combine the video template and the multimedia elements on the video production page provided by the first terminal to generate the video, the video production mode is simpler, and the probability of different users selecting the same video template and the same multimedia elements is lower, so that the richness of the target video can be improved, and the condition that the style of the target video using the same video template is single is avoided.
And the server is also used for sending the target video to the first terminal and updating the target video in response to an element editing instruction of the first terminal, wherein the element editing instruction is an element adding instruction, an element deleting instruction or an element replacing instruction.
That is to say, after the user of the first terminal generates the target video, the user can also edit the target video, so as to improve the playing effect of the finally generated target video.
In addition, the server is further configured to send the sharable elements corresponding to the target video to the first terminal after the target video is generated or updated, so that the first terminal displays the sharable elements, determine the sharable elements selected by the first terminal in response to an element sharing request of the first terminal, and store the sharable elements selected by the first terminal, where the sharable elements include any combination of the target video, production elements of the target video, and video clips in the target video.
That is, after the target video is generated, the user of the first terminal may select to share the target video, the production element of the target video, or the video clip in the target video, so that the production element or the video clip may exist independently of the target video and be independently authored and shared, thereby implementing fragmented creative and sharing.
And the second terminal is used for providing a video production page and displaying a button for acquiring production elements on the video production page, and a user of the second terminal can trigger the second terminal to send a production element acquisition request to the server by clicking the button.
And the server is also used for responding to the element extraction request sent by the second terminal, selecting at least one sharable element from the stored sharable elements and sending the sharable element to the second terminal.
And the second terminal is further used for generating a new video by utilizing the at least one sharable element after receiving the at least one sharable element sent by the server.
In this way, creatives of users of the first terminal can be used by users of the second terminal, thereby enabling fragmented creative flows.
Fig. 2 is a flowchart illustrating a video generation method for use in a server, as shown in fig. 1, according to an exemplary embodiment, the flowchart including the following steps.
S201: the method comprises the steps of responding to a video production request of a first terminal, and determining video production information, wherein the video production information at least comprises production elements, the production elements comprise a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance.
Among them, video templates such as various AE templates, various AI face-changing templates; the multimedia device comprises a first multimedia element such as a picture, a video, a text, an audio and the like, a second multimedia element such as a picture, a video, a text, an audio and the like, and the second multimedia element may or may not have an association with the first multimedia element.
S202: and rendering the video template to obtain an intermediate video corresponding to the video template.
Generally, the video template can realize some special effects such as background blurring, character superposition and the like, so that the intermediate video obtained by rendering the video template is also the video with the special effects.
S203: and generating a target video by using the intermediate video and the first multimedia element corresponding to the video template.
If the first multimedia element is music, the music can be synthesized in the intermediate video corresponding to the video template, so as to obtain the target video.
If the first multimedia element is an icon, the icon can be superimposed in the intermediate video corresponding to the video template, so that the target video is obtained.
In the embodiment of the application, the video template is converted into the intermediate video which can be combined with the multimedia elements, so that a user can freely combine the video template and the first multimedia elements to generate the target video, and the richness of the target video is favorably improved.
In specific implementation, the occurrence segment description information of the production element can be set, for example, whether the production element occurs in a certain segment or all segments, and the occurrence position, occurrence opportunity, end opportunity and other information of the production element in the segments, so as to produce a target video meeting the requirements and further improve the richness of the target video.
Fig. 3 is a flowchart illustrating yet another video generation method, as shown in fig. 1, for use in a server according to an exemplary embodiment, the flowchart including the following steps.
S301: the method comprises the steps of responding to a video production request of a first terminal, and determining video production information, wherein the video production information at least comprises production elements, the production elements comprise a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance.
S302: and rendering the video template to obtain an intermediate video corresponding to the video template.
S303: and determining the fragment composition information according to the occurrence fragment description information of the production element, wherein the fragment composition information at least comprises the video template and/or the first multimedia element which occur in a single fragment.
That is, either the video template or the first multimedia element may appear in one segment alone.
S304: and respectively generating the video clips according to the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in the single clip.
Assuming that a certain segment appears independently as an AE template, the intermediate video corresponding to the AE template can be directly used as a video segment.
Assuming that a certain segment appears separately as video and music, the video and the music can be synthesized, so as to make a video segment.
Assuming that a certain section appears independently as an AE template and a video, a video section can be produced by using the video and the intermediate video corresponding to the AE template according to the production rule in the video production request. For example, the video and the intermediate video corresponding to the AE template are spliced into one video segment, the intermediate video corresponding to the AE template is displayed on the video in a picture-in-picture manner, and the video is displayed on the intermediate video corresponding to the AE template in a picture-in-picture manner.
S305: and generating a target video according to the video clips.
For example, the video segments are spliced to obtain the target video.
In the embodiment of the application, the video template is converted into the intermediate video capable of being combined with the multimedia elements, so that a user can freely combine the video template and the first multimedia elements to generate the target video, and the occurrence information of the video template and/or the first multimedia elements appearing independently in one segment can be set, thereby facilitating the production of rich and diverse videos and further improving the richness of the target video.
Fig. 4 is a flowchart illustrating yet another video generation method, as shown in fig. 1, for use in a server according to an exemplary embodiment, the flowchart including the following steps.
S401: the method comprises the steps of responding to a video production request of a first terminal, and determining video production information, wherein the video production information at least comprises production elements, the production elements comprise a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance.
S402: and rendering the video template to obtain an intermediate video corresponding to the video template.
S403: and determining the fragment composition information according to the occurrence fragment description information of the production elements, wherein the fragment composition information comprises the video template and/or the first multimedia elements which occur in a single fragment and the video template and/or the first multimedia elements which occur in all fragments.
That is, the video template or the first multimedia element may appear in a certain section alone or in all sections. When the video template or the first multimedia element appears in a certain fragment independently, the video template or the first multimedia element is a segmentation element; when the video template or the first multimedia element appears in all the segments, the video template or the first multimedia element is a global element.
S404: and respectively generating the video clips according to the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in the single clip.
S405: and synthesizing the video clips to obtain a synthesized video.
For example, the video segments are first spliced, and then some transition effects are added between the video segments, so as to obtain the composite video.
S406: and modifying the synthesized video by utilizing the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in all the segments to obtain the target video.
If all the segments appear as the video template, modifying the composite video by using the intermediate video corresponding to the video template, and if the intermediate video is clipped into the composite video in a picture-in-picture mode, obtaining the target video.
If the first multimedia element appears in all the segments and is music, the target video can be obtained by using the music to dub the synthesized video.
Assuming that all the segments appear as a video template and a first multimedia element, and the first multimedia element is music, the intermediate video corresponding to the video template can be edited into a composite video in a picture-in-picture manner, and then the music is used to dub the edited video to obtain the target video.
In the embodiment of the application, the video template is converted into the intermediate video capable of being combined with the multimedia elements, so that a user can freely combine the video template and the first multimedia elements to generate the target video, and the occurrence information of the video template and/or the first multimedia elements appearing in a single segment and the occurrence information of the video template and/or the first multimedia elements appearing in all the segments can be set, so as to produce complicated and diversified videos, and further improve the richness of the target video.
In any of the above embodiments, after obtaining the target video, the server may further update the target video in response to an element editing instruction, where the element editing instruction is an element addition instruction, an element deletion instruction, or an element replacement instruction. Therefore, a video satisfying the user can be obtained conveniently, and the user experience is further improved.
In addition, after the target video is generated or updated, the target video, the production elements of the target video or the video clips in the target video can be collected or shared to other users as templates, and the production efficiency of video creativity is improved.
Fig. 5 is a flowchart illustrating yet another video generation method, as shown in fig. 1, for use in a server according to an exemplary embodiment, the flowchart including the following steps.
S501: the method comprises the steps of responding to a video production request of a first terminal, and determining video production information, wherein the video production information at least comprises production elements, the production elements comprise a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance.
S502: and rendering the video template to obtain an intermediate video corresponding to the video template.
S503: and determining the fragment composition information according to the occurrence fragment description information of the production elements, wherein the fragment composition information comprises the video template and/or the first multimedia elements which occur in a single fragment and the video template and/or the first multimedia elements which occur in all fragments.
S504: and respectively generating the video clips according to the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in the single clip.
S505: and synthesizing the video clips to obtain a synthesized video.
S506: and modifying the synthesized video by utilizing the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in all the segments to obtain the target video.
S507: and responding to an element editing instruction, and updating the target video, wherein the element editing instruction is an element adding instruction, an element deleting instruction or an element replacing instruction.
S508: and sending the sharable elements corresponding to the target video to the first terminal, so that the first terminal displays the sharable elements, wherein the sharable elements comprise any combination of the target video, the production elements of the target video and the video clips in the target video.
S509: and responding to the element sharing request of the first terminal, and determining the sharable elements selected by the first terminal.
S510: and saving the sharable elements selected by the first terminal.
In specific implementation, after determining the sharable elements selected by the first terminal, the sharable elements selected by the first terminal may be compared with the stored elements, and after determining that there is no element with similarity higher than a preset value with the sharable elements in the stored elements, the sharable elements selected by the first terminal are stored. In this way, duplicate or similar elements may be avoided from being saved, saving storage space.
S511: and responding to a production element acquisition request of the second terminal, selecting at least one sharable element from the stored sharable elements, and sending the at least one sharable element to the second terminal, so that the second terminal generates a new video by using the at least one sharable element.
In the embodiment of the application, the video template is converted into the intermediate video which can be combined with the multimedia elements, so that a user can freely combine the video template and the first multimedia elements to generate the target video, and the richness of the target video is favorably improved. Moreover, the target video, the production elements of the target video or the video clips in the target video can be independently shared as templates, so that the production elements or the video clips can exist separately from the target video and can be independently shared and used, and fragmented creative production and sharing are realized.
The following describes the generation process of the target video in the present application example by taking the video template as the AE template. In addition, the target video or the segments in the target video can be shared and used as the template, so the process of generating the target video can also be regarded as the process of generating the template.
In the embodiment of the application, the organizational relationship among the manufacturing elements such as audio, AE (automatic evolution) templates, pictures, videos and characters can be presented in the form of template protocols, the process of replacing the template protocols is the process of replacing the templates, the template replacing mode is convenient and fast, new elements are easy to add in the template protocols, and the expansibility of the elements is good.
In practical applications, a video may include more than one segment, for example, a video includes four segments of spring, summer, autumn and winter, and different segments are not affected by each other. In addition, each fragment can have its own attributes such as title, trailer, and general fragment, and transition can be defined between fragments.
In order to produce various rich target videos and improve the richness of the target videos, the structure of the videos in the embodiment of the present application may include global elements and template fragments, where the global elements may run through the entire video cycle, and the template fragments may include segment attributes, segment elements and segment transitions, the segment attributes may include a slice header, a slice trailer, and a general segment, and are used to describe the type of a video fragment, and the segment elements may be elements that may appear in a single fragment, such as a text template, a picture template, a video template, an audio template, and an AE template, and are used to describe transition modes and transition effects between different segments.
In practical application, all production elements have some general attributes, such as element size, element position, element layer, in-out animation, display animation, frame style and the like, and the general display effect of the production elements can be defined through the general attributes. Each production element can also have own specific attribute, such as the font and the font size of the characters, which are the specific attributes of the characters, and the specific display effect of the production elements can be defined through the specific attributes.
In general, a video may include a plurality of segments, such as a slice header, a slice trailer, and other segments, which are independent of each other, without temporal overlapping, and may further include elements that appear in all segments. In order to enable a user to conveniently generate rich and diverse templates, elements in a template protocol can be divided into global elements and segmented elements, when the elements are defined as the global elements, the elements can run through the whole time, when the elements are defined as the segmented elements, the elements only appear in corresponding segments, and a plurality of elements can be defined in a single segment so as to enrich the playing effect of the single segment.
In specific implementation, after determining the AE template and the multimedia elements used for generating the target video, the AE template may be converted into an intermediate video, so that the AE template becomes a video element, and then the target video is generated by using the intermediate video and the multimedia elements converted by the AE template based on the template protocol. In addition, after the target video is generated, operations such as replacing, updating, deleting elements in the target video and the like can be performed on the target video, so that the playing effect of the target video is improved.
Therefore, the AE template is used as a new manufacturing element, and the AE template and the multimedia elements are freely combined, so that the AE template is favorable for a user to generate abundant and diverse target videos by using the AE template, and the user is supported to freely edit the generated target videos, thereby achieving the effects of generating the abundant and diverse target videos and flexibly editing the videos.
In any of the above embodiments, the steps may be implemented in a single thread of the server or in multiple threads. Fig. 6 is a flowchart illustrating a video generation method according to an exemplary embodiment, in which a background service, an asynchronous production service, a video composition service, and other basic services can be regarded as one thread and can be disposed in the same server, and the flowchart includes the following steps:
s601: the web front end sends a video production request to the backend service.
S602: and the background service responds to the video production request, determines and stores the video production parameters.
S603: the background service sends a video asynchronous production instruction to the asynchronous production service.
S604: and asynchronously making a service analysis template protocol.
S605: the asynchronous production service interacts with the video composition service to convert the video template into an intermediate video and compose the text into an image.
S606: the asynchronous production service interacts with other basic services to convert text to audio.
S607: the asynchronous production service sends the intermediate video and the image to the video synthesis service, and the video synthesis service synthesizes the target video.
S608: the video synthesis service sends the target video to the Web front end through the asynchronous production service and the background service.
The video production scheme in the embodiment of the application has the following advantages:
1. having video clip features
Templates can be matched, such as: the AE template is edited, is similar to video clips, overcomes the defect that the template cannot be used with multimedia elements simultaneously when a video is generated, and has good flexibility and expansibility.
2. Threshold for reducing template manufacturing
The video making process is simple, a user does not need to have strong professional knowledge, and the using page of the template is the making page of the template, so that the user can make the template easily, and creative production and sharing are facilitated.
3. Can produce complex special effect video
Templates are supported as special production elements, and multiple templates can be used in producing video. Therefore, the method realizes the production of the complex video clip by virtue of the advantages of the template, is simple and flexible, and has better video playing effect.
In the above process, after the target video is generated, the user may actively share the target video, the elements or segments in the target video to the shared library, and other users may directly select favorite elements or segments from the shared library to generate the video. In addition, when other users browse the target video, the specified elements or fragments are selected from the target video for collection, and when the video is generated, the corresponding elements or fragments are selected from the personal collection library for use. Both of these ways can accelerate the streaming of creatives.
In the embodiment of the application, the fragments or elements can be separated from the target videos to which the fragments or elements belong, and can be shared and used independently, so that when a user generates videos, the user can select the essence fragments or elements from a plurality of target videos to be used by the user, the video making quality is improved, the user can focus on the fragment or element making which is good at the user, the limitation of the whole video making is avoided, the concept of a complete template is broken, and the fragmented creative making and sharing are realized.
Based on the same inventive concept, the application also provides a video generation device. Fig. 7 is a schematic structural diagram of a video generating apparatus according to the present application. The device comprises a response unit 701, a rendering unit 702 and a generating unit 703, wherein:
a response unit 701 configured to perform determining video production information in response to a video production request of a first terminal, where the video production information at least includes production elements, the production elements include a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance;
a rendering unit 702, configured to perform rendering on the video template, to obtain an intermediate video corresponding to the video template;
a generating unit 703 configured to execute generating a target video by using the intermediate video corresponding to the video template and the first multimedia element.
In some possible embodiments, the video production information further includes occurrence segment description information of the production element, and the apparatus further includes:
a determining unit 704 configured to perform determining clip composition information including at least a video template and/or a first multimedia element appearing in a single clip according to the occurrence clip description information of the production element;
the generating unit 703 is specifically configured to perform respective generation of video segments according to intermediate video and/or first multimedia elements corresponding to video templates appearing in a single segment; and generating a target video according to the video clips.
In some possible embodiments, the clip composition information further comprises a video template and/or a first multimedia element occurring at all clips, and
the generating unit 703 is specifically configured to perform synthesis on each video segment to obtain a synthesized video; and modifying the synthesized video by utilizing the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in all the segments to obtain the target video.
In some possible embodiments, the method further comprises:
an updating unit 705 configured to perform updating of the target video in response to an element editing instruction, which is an element addition instruction, an element deletion instruction, or an element replacement instruction, after generating the target video.
In some possible embodiments, the method further comprises:
a sending unit 706, configured to send, after a target video is generated or updated, a sharable element corresponding to the target video to the first terminal, where the sharable element is displayed by the first terminal, and the sharable element includes any combination of the target video, a production element of the target video, and a video clip in the target video;
a sharing unit 707 configured to execute determining a sharable element selected by the first terminal in response to an element sharing request of the first terminal;
a saving unit 708 configured to perform saving the sharable element selected by the first terminal.
In some possible embodiments, the method further comprises:
the selecting unit 709 is further configured to execute, in response to a production element obtaining request of the second terminal, selecting at least one sharable element from the stored sharable elements and sending the at least one sharable element to the second terminal, so that the second terminal generates a new video by using the at least one sharable element.
With regard to the apparatus in the above-described embodiment, the specific manner in which each module performs the operation has been described in detail in the embodiment related to the method, and will not be elaborated here.
The division of the modules in the embodiments of the present application is schematic, and only one logical function division is provided, and in actual implementation, there may be another division manner, and in addition, each functional module in each embodiment of the present application may be integrated in one processor, may also exist alone physically, or may also be integrated in one module by two or more modules. The coupling of the various modules to each other may be through interfaces that are typically electrical communication interfaces, but mechanical or other forms of interfaces are not excluded. Thus, modules described as separate components may or may not be physically separate, may be located in one place, or may be distributed in different locations on the same or different devices. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode.
Fig. 8 is a schematic structural diagram of an electronic device according to an exemplary embodiment, where the electronic device includes a transceiver 801 and a physical device such as a processor 802, where the processor 802 may be a Central Processing Unit (CPU), a microprocessor, an application specific integrated circuit, a programmable logic circuit, a large scale integrated circuit, or a digital Processing Unit. The transceiver 801 is used for data transmission and reception between electronic devices and other devices.
The electronic device may further comprise a memory 803 for storing software instructions executed by the processor 802, but may also store some other data required by the electronic device, such as identification information of the electronic device, encryption information of the electronic device, user data, etc. The Memory 803 may be a Volatile Memory (Volatile Memory), such as a Random-Access Memory (RAM); the Memory 803 may also be a Non-Volatile Memory (Non-Volatile Memory) such as, but not limited to, a Read-Only Memory (ROM), a Flash Memory (Flash Memory), a Hard Disk Drive (HDD) or a Solid-State Drive (SSD), or the Memory 803 may be any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer. The memory 803 may be a combination of the above.
The specific connection medium between the processor 802, the memory 803 and the transceiver 801 is not limited in the embodiments of the present application. In fig. 8, the embodiment of the present application is described by taking only the case where the memory 803, the processor 802, and the transceiver 801 are connected by the bus 804 as an example, the bus is shown by a thick line in fig. 8, and the connection manner between other components is merely illustrative and not limiting. The bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one thick line is shown in FIG. 8, but this is not intended to represent only one bus or type of bus.
The processor 802 may be dedicated hardware or a processor running software, and when the processor 802 may run software, the processor 802 reads software instructions stored in the memory 803 and executes the video generation method involved in the foregoing embodiments under the drive of the software instructions.
The embodiment of the present application also provides a storage medium, and when instructions in the storage medium are executed by a processor of an electronic device, the electronic device is capable of executing the video generation method in the foregoing embodiment.
In some possible implementations, the aspects of the video generation method provided in this application may also be implemented in the form of a program product including program code for causing an electronic device to perform the video generation method referred to in the foregoing embodiments when the program product is run on the electronic device.
The program product may employ any combination of one or more readable media. The readable medium may be a readable signal medium or a readable storage medium. A readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples (a non-exhaustive list) of the readable storage medium include: an electrical connection having one or more wires, a portable Disk, a hard Disk, a Random Access Memory (RAM), a Read Only Memory (ROM), an Erasable Programmable Read-Only Memory (EPROM), a flash Memory, an optical fiber, a Compact disc Read-Only Memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
The program product for video production in the embodiments of the present application may be a CD-ROM and include program code, and may be run on a computing device. However, the program product of the present application is not limited thereto, and in this document, a readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A readable signal medium may include a propagated data signal with readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A readable signal medium may also be any readable medium that is not a readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, Radio Frequency (RF), etc., or any suitable combination of the foregoing.
Program code for carrying out operations of the present application may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, C + + or the like and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computing device, partly on the user's device, as a stand-alone software package, partly on the user's computing device and partly on a remote computing device, or entirely on the remote computing device or server. In situations involving remote computing devices, the remote computing devices may be connected to the user computing device through any kind of Network, including a Local Area Network (LAN) or Wide Area Network (WAN), or may be connected to an external computing device (e.g., through the internet using an internet service provider).
It should be noted that although several units or sub-units of the apparatus are mentioned in the above detailed description, such division is merely exemplary and not mandatory. Indeed, the features and functions of two or more units described above may be embodied in one unit, according to embodiments of the application. Conversely, the features and functions of one unit described above may be further divided into embodiments by a plurality of units.
Further, while the operations of the methods of the present application are depicted in the drawings in a particular order, this does not require or imply that these operations must be performed in this particular order, or that all of the illustrated operations must be performed, to achieve desirable results. Additionally or alternatively, certain steps may be omitted, multiple steps combined into one step execution, and/or one step broken down into multiple step executions.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
While the preferred embodiments of the present application have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all alterations and modifications as fall within the scope of the application.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present application without departing from the spirit and scope of the application. Thus, if such modifications and variations of the present application fall within the scope of the claims of the present application and their equivalents, the present application is intended to include such modifications and variations as well.

Claims (10)

1. A method of video generation, comprising:
responding to a video production request of a first terminal, determining video production information, wherein the video production information at least comprises production elements, the production elements comprise a video template and first multimedia elements, and the video template is obtained by combining at least two second multimedia elements in advance;
rendering the video template to obtain an intermediate video corresponding to the video template;
and generating a target video by using the intermediate video corresponding to the video template and the first multimedia element.
2. The method of claim 1, wherein the video production information further includes occurrence segment description information of the production elements, the method further comprising:
determining fragment synthesis information according to the occurrence fragment description information of the production elements, wherein the fragment synthesis information at least comprises a video template and/or a first multimedia element which occur in a single fragment;
generating a target video by using the intermediate video corresponding to the video template and the first multimedia element, wherein the generating comprises:
respectively generating video clips according to the intermediate video and/or the first multimedia elements corresponding to the video template appearing in the single clip;
and generating a target video according to the video clips.
3. The method of claim 2, wherein the clip composition information further comprises video templates and/or first multimedia elements occurring in all clips, and wherein generating the target video from each video clip comprises:
synthesizing each video clip to obtain a synthesized video;
and modifying the synthesized video by utilizing the intermediate video and/or the first multimedia elements corresponding to the video templates appearing in all the segments to obtain the target video.
4. The method of claim 1, after generating the target video, further comprising:
and responding to an element editing instruction, and updating the target video, wherein the element editing instruction is an element adding instruction, an element deleting instruction or an element replacing instruction.
5. The method according to any one of claims 1-4, further comprising, after generating or updating the target video:
sending sharable elements corresponding to the target video to the first terminal, and displaying the sharable elements by the first terminal, wherein the sharable elements comprise any combination of the target video, production elements of the target video and video clips in the target video;
responding to an element sharing request of the first terminal, and determining sharable elements selected by the first terminal;
and saving the sharable elements selected by the first terminal.
6. The method of claim 5, further comprising:
and responding to a manufacturing element acquisition request of a second terminal, selecting at least one sharable element from the stored sharable elements, and sending the at least one sharable element to the second terminal, so that the second terminal generates a new video by using the at least one sharable element.
7. A video generation apparatus, comprising:
a response unit configured to execute a video production request in response to a first terminal, and determine video production information, where the video production information at least includes production elements, the production elements include a video template and a first multimedia element, and the video template is obtained by combining at least two second multimedia elements in advance;
the rendering unit is configured to perform rendering on the video template to obtain an intermediate video corresponding to the video template;
and the generating unit is configured to generate a target video by using the intermediate video corresponding to the video template and the first multimedia element.
8. The apparatus of claim 7, wherein the video production information further includes occurrence segment description information of the production elements, the apparatus further comprising:
a determining unit configured to perform determining clip composition information including at least a video template and/or a first multimedia element appearing in a single clip according to the appearance clip description information of the production element;
the generating unit is specifically configured to execute intermediate video and/or first multimedia elements corresponding to video templates appearing in a single segment, and respectively generate video segments; and generating a target video according to the video clips.
9. An electronic device, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the video generation method of any of claims 1 to 6.
10. A storage medium, wherein instructions in the storage medium, when executed by a processor of an electronic device, enable the electronic device to perform the video generation method of any of claims 1 to 6.
CN202110597266.9A 2021-05-31 2021-05-31 Video generation method and device, electronic equipment and storage medium Active CN113347465B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110597266.9A CN113347465B (en) 2021-05-31 2021-05-31 Video generation method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110597266.9A CN113347465B (en) 2021-05-31 2021-05-31 Video generation method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN113347465A true CN113347465A (en) 2021-09-03
CN113347465B CN113347465B (en) 2023-04-28

Family

ID=77472077

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110597266.9A Active CN113347465B (en) 2021-05-31 2021-05-31 Video generation method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113347465B (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101448089A (en) * 2007-11-26 2009-06-03 新奥特(北京)视频技术有限公司 Non-linear editing system
US20130272679A1 (en) * 2012-04-12 2013-10-17 Mario Luis Gomes Cavalcanti Video Generator System
US20150371679A1 (en) * 2012-05-01 2015-12-24 Wochit, Inc. Semi-automatic generation of multimedia content
CN109168028A (en) * 2018-11-06 2019-01-08 北京达佳互联信息技术有限公司 Video generation method, device, server and storage medium
CN109600628A (en) * 2018-12-21 2019-04-09 广州酷狗计算机科技有限公司 Video creating method, device, computer equipment and storage medium
CN110532426A (en) * 2019-08-27 2019-12-03 新华智云科技有限公司 It is a kind of to extract the method and system that Multi-media Material generates video based on template
CN110708596A (en) * 2019-09-29 2020-01-17 北京达佳互联信息技术有限公司 Method and device for generating video, electronic equipment and readable storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101448089A (en) * 2007-11-26 2009-06-03 新奥特(北京)视频技术有限公司 Non-linear editing system
US20130272679A1 (en) * 2012-04-12 2013-10-17 Mario Luis Gomes Cavalcanti Video Generator System
US20150371679A1 (en) * 2012-05-01 2015-12-24 Wochit, Inc. Semi-automatic generation of multimedia content
CN109168028A (en) * 2018-11-06 2019-01-08 北京达佳互联信息技术有限公司 Video generation method, device, server and storage medium
CN109600628A (en) * 2018-12-21 2019-04-09 广州酷狗计算机科技有限公司 Video creating method, device, computer equipment and storage medium
CN110532426A (en) * 2019-08-27 2019-12-03 新华智云科技有限公司 It is a kind of to extract the method and system that Multi-media Material generates video based on template
CN110708596A (en) * 2019-09-29 2020-01-17 北京达佳互联信息技术有限公司 Method and device for generating video, electronic equipment and readable storage medium

Also Published As

Publication number Publication date
CN113347465B (en) 2023-04-28

Similar Documents

Publication Publication Date Title
CN108989885B (en) Video file transcoding system, segmentation method, transcoding method and device
US9277198B2 (en) Systems and methods for media personalization using templates
CN102473189B (en) Providing in real time to the link of portion of media object during social networks upgrades
US20140172856A1 (en) Method and system for storytelling on a computing device
CN104065979A (en) Method for dynamically displaying information related with video content and system thereof
US11974013B2 (en) Modifying digital video content
US20220277775A1 (en) Video generating method, apparatus, electronic device, and computer-readable medium
CN112333536A (en) Audio and video editing method, equipment and computer readable storage medium
CN111327968A (en) Short video generation method, short video generation platform, electronic equipment and storage medium
WO2021238084A1 (en) Voice packet recommendation method, apparatus and device, and storage medium
US20200186869A1 (en) Method and apparatus for referencing, filtering, and combining content
CN112015927B (en) Method and device for editing multimedia file, electronic equipment and storage medium
CN114466222B (en) Video synthesis method and device, electronic equipment and storage medium
CN111797061A (en) Multimedia file processing method and device, electronic equipment and storage medium
US10153003B2 (en) Method, system, and apparatus for generating video content
US10783319B2 (en) Methods and systems of creation and review of media annotations
US9721321B1 (en) Automated interactive dynamic audio/visual performance with integrated data assembly system and methods
CN113190316A (en) Interactive content generation method and device, storage medium and electronic equipment
EP4276828A1 (en) Integrated media processing pipeline
CN113347465B (en) Video generation method and device, electronic equipment and storage medium
CN115314732B (en) Multi-user collaborative film examination method and system
CN113711575A (en) System and method for instantly assembling video clips based on presentation
CN104111768B (en) Interactive window and method and system for customizing, quoting and synchronizing interactive window
CN108616768A (en) Synchronous broadcast method, device, storage location and the electronic device of multimedia resource
CN105487769A (en) Media file display method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant