CN117425033A - Video generation method, device, electronic equipment and storage medium - Google Patents

Video generation method, device, electronic equipment and storage medium Download PDF

Info

Publication number
CN117425033A
CN117425033A CN202311416310.7A CN202311416310A CN117425033A CN 117425033 A CN117425033 A CN 117425033A CN 202311416310 A CN202311416310 A CN 202311416310A CN 117425033 A CN117425033 A CN 117425033A
Authority
CN
China
Prior art keywords
video
target
page
information
script
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311416310.7A
Other languages
Chinese (zh)
Inventor
鲁国珍
杨帆
赵巍
彭荣贵
田翔
黄严
周鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Zitiao Network Technology Co Ltd
Original Assignee
Beijing Zitiao Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Zitiao Network Technology Co Ltd filed Critical Beijing Zitiao Network Technology Co Ltd
Priority to CN202311416310.7A priority Critical patent/CN117425033A/en
Publication of CN117425033A publication Critical patent/CN117425033A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The disclosure relates to a video generation method, a device, an electronic device and a storage medium, wherein the method comprises the following steps: displaying a target page; the target page comprises video generation options; responding to the selection operation of the video generation options, and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video; displaying a video preview page according to video creation destination information input in the information collection page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information. By adopting the technical scheme provided by the application, automatic video editing can be realized, the video editing threshold can be reduced, the video editing time consumption is shortened, and the video editing efficiency is improved.

Description

Video generation method, device, electronic equipment and storage medium
Technical Field
The present disclosure relates to the field of computer vision, and in particular, to a video generating method, apparatus, electronic device, and storage medium.
Background
Video editing refers to fusing shots of different scenes together by editing multiple video materials to express a particular story or intent.
Currently, video editing mainly relies on manual operation, and has high requirements on professional quality and artistic maintenance of editors. This makes the threshold for video editing higher. To solve this problem, implementing automated video editing has become a current urgent problem to be solved.
Disclosure of Invention
In order to solve the technical problems described above or at least partially solve the technical problems described above, the present disclosure provides a video generation method, apparatus, electronic device, and storage medium.
In a first aspect, the present disclosure provides a video generating method, including:
displaying a target page; the target page comprises video generation options;
responding to the selection operation of the video generation options, and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video;
if the video creation destination information is detected in the information collection page, displaying a video preview page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information.
In a second aspect, the present disclosure also provides a video generating apparatus, including:
the first display module is used for displaying the target page; the target page comprises video generation options;
the second display module is used for responding to the selection operation of the video generation options and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video;
the third display module is used for displaying a video preview page if the video creation destination information is detected in the information collection page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information.
In a third aspect, the present disclosure also provides an electronic device, including:
one or more processors;
a storage means for storing one or more programs;
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the video generation method as described above.
In a fourth aspect, the present disclosure also provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements a video generation method as described above.
Compared with the prior art, the technical scheme provided by the embodiment of the disclosure has the following advantages:
according to the technical scheme provided by the embodiment of the disclosure, the information collection page is displayed by setting a selection operation in response to the video generation options; the information collection page is used for collecting video creation destination information of the generated video; displaying a video preview page according to video creation destination information input in the information collection page; the video preview page is used for displaying the target video; the target video is a video generated by editing video materials in the video material library based on video creation destination information, and the essence is that under the condition that only the video creation destination information is input, the electronic equipment automatically edits the video materials to obtain the target video capable of reflecting the video creation destination.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the disclosure and together with the description, serve to explain the principles of the disclosure.
In order to more clearly illustrate the embodiments of the present disclosure or the solutions in the prior art, the drawings that are required for the description of the embodiments or the prior art will be briefly described below, and it will be obvious to those skilled in the art that other drawings can be obtained from these drawings without inventive effort.
Fig. 1 is a flowchart of a video generating method according to an embodiment of the present disclosure;
FIGS. 2-5 are schematic diagrams of several electronic device display interfaces provided by embodiments of the present disclosure;
FIG. 6 is a flow chart of a method of generating a target video provided by an embodiment of the present disclosure;
FIG. 7 is a schematic diagram of generating a target video according to an embodiment of the present disclosure;
fig. 8 is a schematic structural diagram of a video generating apparatus in an embodiment of the present disclosure;
fig. 9 is a schematic structural diagram of an electronic device in an embodiment of the disclosure.
Detailed Description
In order that the above objects, features and advantages of the present disclosure may be more clearly understood, a further description of aspects of the present disclosure will be provided below. It should be noted that, without conflict, the embodiments of the present disclosure and features in the embodiments may be combined with each other.
In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present disclosure, but the present disclosure may be practiced otherwise than as described herein; it will be apparent that the embodiments in the specification are only some, but not all, embodiments of the disclosure.
Fig. 1 is a flowchart of a video generating method provided in an embodiment of the present disclosure, where the embodiment may be suitable for a case of video editing in a client, and the method may be performed by a video generating device, where the device may be implemented in software and/or hardware, and the device may be configured in an electronic device, for example, a terminal, and specifically includes, but is not limited to, a smart phone, a palm top computer, a tablet computer, a wearable device with a display screen, a desktop computer, a notebook computer, an all-in-one machine, a smart home device, and so on.
As shown in fig. 1, the method specifically may include:
s110, displaying a target page; the target page includes video generation options.
The target page may be, for example, a page that is opened after the user performs some operation. The video generation option is a starting point with a video generation function that can be edited. The video generation options may specifically be controls, menu items, or specific text, etc.
S120, responding to the selection operation of the video generation options, and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video.
The selection operation of the video generation option may be, for example, a click operation or a slide operation of the video generation option. The user performs a selection operation on the video generation option for communicating to the electronic device an intent of the user to edit the video.
The information collection page may be, for example, a page for collecting the creation purpose of the edited video. The video creation destination information may be, for example, creation destination information for reflecting an edited video. By way of example, a video creation may be, for example, advertising a good, brand, service, store, or merchant to attract the attention, purchase, use of a user viewing a video work. Alternatively, the video creation purpose may be, for example, to make a video work of a certain style (such as a fun style, romance style, martial arts style) for the purpose of recreation and entertainment. Alternatively, the video creation purpose may be, for example, to make a user action for conveying knowledge or ideas, or inspiring guidance to view the video work.
In practice, the video creation destination information includes different contents for different video creation destinations. Illustratively, the video creation purpose is to promote a certain commodity of a certain store, and the video creation purpose information includes at least one of: video names, merchant names, store descriptions, merchandise names, core selling points, and promotional levels that are desired to be generated. The video creation purpose is to explain a certain knowledge point, and the video creation purpose information may include at least one of the following, for example: the name of the video desired to be generated, the content of the knowledge point, and the audience object of the video.
S130, displaying a video preview page according to video creation destination information input in the information collection page; the video preview page is used for displaying the target video; the target video is video generated by editing video materials in the video material library based on the video creation destination information.
The video material library may be, for example, a collection of video material. Video material is an element in a library of video material, which may include video and/or pictures in particular.
The target video may be, for example, a video generated during the current execution of the video generating method, that is, a video editing result. The target video is generated according to the video creation purpose, accords with the video creation purpose, and can enable viewers of the target video to obtain corresponding experience and feel so as to achieve expected effects.
Optionally, the solution provided in the present application is executed once, and may be regarded as a video generating task, where the task is executed to obtain one or more target videos. A set of all target videos obtained through one video generation task is defined as a target video set, and the target video set comprises one or more target videos.
The video preview page may be, for example, a page that presents the target video in a preview prior to release of the target video. The specific display mode of the target video displayed on the video preview page can comprise directly playing the target video or firstly summarizing and displaying the summary information of all the target videos in the target video set, and playing the selected target video after receiving the selection operation of a user on a certain target video. The summary information of the target video may include, for example, at least one of a cover, a duration, and a video name of the target video. The direct playing of the target video may be, for example, playing each target video in the target video set in turn, or playing only one target video in the target video set.
Referring to fig. 2, a "create video task" control, which is a video generation option, is included on the target page. After the user clicks on the "Create video task" control, see FIG. 3, the information collection page is presented. The information collection page is provided with a plurality of configuration items, including: the "task name" configuration item, the "merchant name" configuration item, the "store description" configuration item, the "commodity name" configuration item, the "core selling point" configuration item, and the "promotion level" configuration item. The user can input proper information in the information input area corresponding to each configuration item according to the video creation purpose of the user, and click on the option of generating video. Information input in the information input area corresponding to all the configuration items jointly form video creation destination information. And editing the video materials in the video material library according to the video creation destination information input in the information collection page to generate a target video. Referring to fig. 4, a video preview page is presented in which a target video is presented.
According to the technical scheme, the information collection page is displayed by setting the selection operation responding to the video generation options; the information collection page is used for collecting video creation destination information of the generated video; displaying a video preview page according to video creation destination information input in the information collection page; the video preview page is used for displaying the target video; the target video is a video generated by editing video materials in the video material library based on video creation destination information, and the essence is that under the condition that only the video creation destination information is input, the electronic equipment automatically edits the video materials to obtain the target video capable of reflecting the video creation destination.
Optionally, the information collection page is also used to collect video material. Illustratively, referring to fig. 3, the information collection page also includes a material upload option, which the user can upload video material by triggering.
On the basis of the above technical solution, optionally, after S130, the method further includes: responding to the selection operation of the release rule configuration options, and displaying a release rule configuration page; the release rule configuration page is used for configuring release rules; the release rule configuration options are located in the target page; and distributing the target videos in the target video set according to the distribution rules input in the distribution rule configuration page.
The publishing rules may be, for example, rules that need to be followed when publishing the target video on some platforms. Alternatively, the publication rules may define through which accounts the target video is published, in what manner (e.g., directly or if certain conditions are met), with what frequency, etc. After the target video is published on the platform, other users of the platform can browse the target video.
If one video generation task is completed, a plurality of target videos are obtained, and a release rule configuration page is used for configuring release rules for a target video set; or the release rule configuration page is used for configuring release rules for part of the target videos in the target video set.
The publication rule configuration page is used for configuring the publication rule for the target video set, namely, the publication rule configured by the publication rule configuration page is specific to all target videos.
The distribution rule configuration page is used for configuring distribution rules for part of target videos in the target video set, and the distribution rules configured by the distribution rule configuration page are specific to one or a few target videos. In this case, different target videos may be published using different publication rules, or may be published using the same publication rules.
The release rule configuration options refer to options with a video generation function, and after triggering, release rule configuration pages of the target video are displayed, wherein the release rule configuration options can be controls, menu items or specific texts.
The selection operation of the distribution rule configuration option may be, for example, a click or slide operation of the distribution rule configuration option or the like.
In some embodiments, if the distribution rule configuration page is used to configure the distribution rule for the target video set, the target page may be set to further include summary information of the target video set, where a display position of the distribution rule configuration option of the target video is adjacent to a display position of the summary information of the target video set.
The target video set may have a plurality of attributes after being generated, for example, if the video is created for the purpose of commodity promotion, the attributes of the target video set include, for example, a generation time, a task name, a target video content, a name of a commodity promoted, a name of a merchant selling the commodity promoted, and the like. Summary information of a target video set may be, for example, a set of partial attributes of the target video set.
The display position of the release rule configuration options is adjacent to the display position of the summary information of the target video set, for example, the target page may include a plurality of sub-display areas arranged vertically or horizontally, and the release rule configuration options and the summary information corresponding to the same target object set are located in the same sub-display area.
For example, referring to fig. 2, in the target page, a plurality of sub-presentation areas are arranged vertically, and distribution rule configuration options corresponding to the same target object set and summary information of the target video are arranged horizontally to occupy the same line.
In practice, when the video generating method provided in the present application is performed multiple times, it can be regarded that a plurality of video generating tasks are performed, and a plurality of target video sets will be obtained. Because each target object set is provided with the corresponding release rule configuration options, a plurality of release rule configuration options are required to be displayed in the target page, and the user can understand which target video set each release rule configuration option corresponds to by means of the abstract information of the target video set by setting the display position of the release rule configuration option of the same target video set adjacent to the display position of the abstract information.
Illustratively, suppose that the video generation method provided herein was previously performed 4 times in total, see fig. 2, the target page includes summary information of four target video sets and four distribution rule configuration options. In fig. 2, each row represents a target video set. The user can know which target video set each row represents through the summary information in each row. If the user selects the publishing rule configuration option in the first row, see fig. 5, a publishing rule configuration page of the target video set represented by the first row is presented. The distribution rule configuration page includes a plurality of configuration items such as an account number for distributing a target video, POI (address of a store selling a commodity advertised by the target video) setting, distribution mode, and distribution period. When the user configures each configuration item in the release rule configuration page and selects a release option, the release rule defined by the configuration content of each configuration item together releases the target videos in the target video set.
In practice, all video materials in the video material library may be required in the process of generating the target video, or only part of the video materials in the video material library may be used.
Those skilled in the art will appreciate that in practice, the same video material, if applied to make multiple target videos, can be tiring for a user viewing the target videos. In this regard, optionally, each time a video material is used in generating the target video, the video material is marked. The purpose of marking the video material is to indicate that the video material has been applied to generate a target video. On the basis, optionally, if the material consumption proportion of the video material library is greater than or equal to a set threshold value, displaying prompt information for prompting the supplementary video material; the material consumption ratio is the ratio of the number of video materials that have been applied to generate the target video to the total number of video materials in the video material library.
If the material consumption ratio of the video material library is greater than or equal to the set threshold, the number of video materials which are not applied to generating the target video in the video material library is small, and the video materials need to be timely supplemented so as to ensure that the target video can be generated based on the video materials in the video material library.
In the above technical solution, the specific implementation method of S130 is various, which is not limited in this application. Illustratively, the specific implementation method of S130 includes: creating target video according to video creation destination information input in the information collection page; and displaying a video preview page, wherein the video preview page comprises target videos in a target video set.
Further, there are various specific implementation methods for generating the target video, which are not limited in this application. Illustratively, referring to fig. 6, generating the target video includes:
s210, determining a video script based on the video creation destination information.
The video script may be, for example, a planning scheme for the target video, which defines that several shots are required for generating the target video, the content or expressed emotion to be exhibited by each shot, and the like.
Optionally, the video script includes a plurality of shot scripts, and the shot scripts are planning schemes for one shot. Illustratively, one shot script includes at least one of a shot name, a document, screen information, and highlight information. The document is information describing the contents of the shots. The screen information is information describing features of the shot content. The highlight information is information that is desired to be presented in a highlighted form in the shot. Further, it may also be set that the screen information includes at least one of a scene, a subject, and a style. The scene refers to a background environment in a picture, the main object refers to a person, an animal or an object which takes the dominant role in the picture, and the main object is a core object for showing the lens. Style refers to the overall visual effect of a picture. The style may convey the emotional mood of the picture, such as happy, sad, serious, or relaxed, etc.
A target video is formed by arranging a plurality of shots in a certain order. Accordingly, in one video script, a plurality of shot scripts are arranged in a certain order.
There are various ways to implement this step, and this application is not limited thereto. In one embodiment, the implementation method of the step includes: extracting keywords from the video creation destination information to obtain target keywords; and determining the video script corresponding to the target keyword in the video script library.
Illustratively, assume that the video creation purpose is to advertise a certain merchandise. And extracting keywords from the video creation destination information, wherein the obtained target keywords are commodity propagandas, and searching video scripts for carrying out commodity propaganda in a video script library.
In practice, the video creation purposes are various, and even if the video creation purposes are all used for advertising goods, angles and modes of different goods are slightly different. This makes it possible that the method for generating the video script of the target video determined by the method for querying the video script library may occur that the determined video script is not well suited for promoting the merchandise that the user wishes to promote, eventually resulting in a failure to achieve the video creation purpose well.
In another embodiment, the implementation method of this step includes: and inputting the video creation destination information into a video script generation model to obtain a video script corresponding to the video creation destination. The video script generation model is a neural network model taking video creation destination description information as input information and taking video script as output information. The "video script generation model" mentioned in this paragraph is a trained neural network model. By the arrangement, the determined video script can be well adapted to the video creation purpose.
In practice, a good quality sample video including multiple shots may be selected to train the video script generation model. The specific training process can comprise the following steps: obtaining a sample video; converting the sample video into a sample video script and creating target information of the sample video; the video script generation model is trained based on the sample video script and sample video creation destination information.
Further, the method further comprises: acquiring play data of the released target video; and adjusting the video script generation model according to the play data.
The play data may be, for example, an indicator reflecting the quality and popularity of the target video. Illustratively, the play data includes at least one of: the completion rate, the praise number, the comment number and the play amount. The playback completion rate may be, for example, a ratio of the number of times the target video is viewed to the number of times the target video is played.
"adjusting the video script generation model according to the play data" means optimizing the video script generation model so that it can output a higher quality video script. The better the video script, the better the quality of the generated target video, and the higher its popularity.
The evaluation criteria for the high-quality video script are various. Such as high video playback rates or high popularity.
Illustratively, the preset shot script includes: the method comprises the steps of dividing a scene into a plurality of parts, and dividing the scene into a plurality of parts, wherein the part comprises a division name, a text, picture information and highlight information, and the picture information comprises a scene, a main body object and a style. Optionally, an auxiliary rule may be further set, such as that the screen information does not include a shop name, the highlight information needs to select text that can attract a video viewer when determining, the word count limit condition of the text, and the style in the text and the screen information is determined according to the promotion degree.
In one example, if the video creation destination information includes: the store is described as a newly opened XX coffee shop, is positioned beside a YY subway station, has a core selling point of low price and good environment, and has the trade name of osmanthus latte which only needs zz yuan and has low propaganda degree. Wherein XX represents a coffee shop name, YY represents a subway station name, and zz represents the price of the osmanthus latte. After the video creation destination information is input into a video script generation model, the obtained video script comprises 3 shot scripts. Specifically, the video script is as follows.
Video script title: enjoying calm afternoon hours in a warm cafe
Minute mirror 1 name: experience Process 01
Document case: the XX is newly arranged beside the YY subway station, and a plurality of people working nearby walk around
Picture information:
scene: street beside subway station
The main body: signboard for coffee shop
Style: city
Highlighting information: YY subway station
Split mirror 2 name: experience Process 02
Document case: sit on this reading and reading cup and drink a cup of coffee, enjoy calm for a while
Picture information:
scene: in the room of the coffee shop,
the main body: one person can read and drink coffee
Style: elegant and delicate
Highlighting information: without any means for
Split mirror 3 name: expanding introduction environments
Document case: his home decoration is simple and warm, and sunlight is transmitted through a large-surface glass window
Picture information:
scene: coffee shop indoor
The main body: sunlight-permeable large glass window
Style: bright, relax
Highlighting information: without any means for
In another example, if the video creation destination information includes: the store is described as a newly opened XX coffee shop, is positioned beside a YY subway station, has a core selling point of low price and good environment, and has the trade name of osmanthus latte, only needing zz yuan and high propaganda degree. Wherein XX represents a coffee shop name, YY represents a subway station name, and zz represents the price of the osmanthus latte. After the video creation destination information is input into a video script generation model, the obtained video script comprises 3 shot scripts. Specifically, the video script is as follows.
Video script title: is very economical, only 8.8 of sweet osmanthus latte is needed, and sisters can wash
Minute mirror 1 name: manufacturing suspense
Document case: the new XX is too cheap, and only 8.8 picture information is needed for the osmanthus latte:
scene: coffee shop indoor
The main body: coffee cup
Style: elegant and delicate
Highlighting information: is too cheap
Split mirror 2 name: development introduction product
Document case: carefully selecting coffee beans, finely grinding and brewing, and diffusing aroma
Picture information:
scene: in the room of the coffee shop,
the main body: coffee bean grinding by a coffer
Style: specialized work
Highlighting information: without any means for
Split mirror 3 name: introduction of the group product preference
Document case: only 8.8 of osmanthus latte is needed, and the user can get to store to experience bar quickly
Picture information:
scene: coffee shop indoor
The main body: one cup of coffee and some desserts
Style: food for delicacies
Highlighting information: only 8.8
S220, selecting target video materials from the video material library based on the video script.
There are various ways to implement this step, and this application is not limited thereto. Illustratively, the video script comprises a plurality of shot scripts; the shot script comprises shot content information; the implementation method of the steps comprises the following steps: determining first characteristic information of each video material in a video material library; matching the shot content information in the shot script with first characteristic information of the video material; and taking the video material in the matching as a target video material.
If the shot script comprises shot names, texts, picture information and highlight information. The shot content information may specifically include one or more of a shot name, a document, screen information, and highlight information.
The first characteristic information may be, for example, descriptive information of content reflected by the video material. The first characteristic information is used for judging whether a video material is used as a target video material. Illustratively, the first characteristic information may include at least one of: the picture comprises a picture inner part, a picture created atmosphere, a picture main body object, a picture reflected emotion, a picture style and a picture scene.
For example, a specific implementation method of "determining first characteristic information of each video material in a video material library" may include: and extracting images of a preset frame number aiming at each video material, and then carrying out feature extraction on the extracted images to obtain first feature information of the video material. In some scenarios, feature extraction may be performed on the extracted image using a neural network model that is capable of understanding semantic links between the image and text, thereby obtaining first feature information of the video material.
The specific implementation method for matching the shot content information in the shot script with the first characteristic information of the video material can comprise the following steps: feature information of each shot can be obtained through shot content information in the shot script; and calculating the similarity between the characteristic information of the sub-lens and the first characteristic information of each video material, and taking the video material with the maximum similarity as the target video material corresponding to the sub-lens.
And S230, synthesizing the target video materials according to the video script to obtain the target video.
Optionally, the implementation method of the step includes that the target video materials corresponding to the sub-shots are spliced together according to the arrangement sequence of the sub-shots in the video script, so as to obtain the target video.
Illustratively, referring to fig. 7, the video material library includes 10 video materials. Matching the shot content information in the shot script with first characteristic information of the video material to obtain video materials 6, 7 and 8 which are matched with the shot script 1, wherein the video materials 6, 7 and 8 are target video materials corresponding to the shot script 1; the video materials 3, 2 and 9 are matched with the shot script 2, and the video materials 3, 2 and 9 are target video materials corresponding to the shot script 2; the video materials 4, 1 and 10 are matched with the shot script 3, and the video materials 4, 1 and 10 are target video materials corresponding to the shot script 3; one of the video materials 6, 7 and 8, one of the video materials 3, 2 and 9, and one of the video materials 4, 1 and 10 are spliced to obtain three target videos.
According to the technical scheme, the video script is determined by setting the video-based creation destination information; selecting a target video material from a video material library based on the video script; according to the video script, synthesizing the target video material to obtain the target video, and providing a method capable of automatically producing the target video.
Optionally, after S230, the method further includes: determining target music according to the video creation destination information; synthesizing target music and target video; the video preview page is used for displaying the target video synthesized with the target music.
The target music may be, for example, music for background music as a target video. Determining the target music according to the video creation destination information means that the target music is not randomly selected, but is determined according to the video creation destination information, and can assist the target video to achieve the expected effect.
Illustratively, the music library is set to include a plurality of pieces of music, different pieces of music correspond to different video creation purposes, for example, the video creation purpose corresponding to the music a is a commodity propaganda, the video creation purpose corresponding to the music b is a leisure entertainment, and the video creation purpose corresponding to the music c is a delivery of knowledge. If the video creation destination information indicates that the video creation destination is commodity propaganda, taking the music a as target music, and then combining the music a with the target video. And then, displaying the target video synthesized with the music a in the video preview page, and releasing the target video synthesized with the music a.
Further, after S120, the video generating method further includes: determining a target editing element in the editing element library according to the second characteristic information of the target video material and the matching relation of all editing elements in the editing element library; synthesizing the target editing element with the target video; the video preview page is used for displaying the target video synthesized with the target editing element.
Editing elements may include, for example, recommended animations, transition manners of transition, and special effects. The recommended animation may be, for example, a specific animation effect for enhancing visual impact or narrative ability, which is recommended for a single shot. Such as zoom, rotate, move, etc. The transition mode may be, for example, a mode in which two shots or a plurality of shots are connected smoothly. The special effects may be, for example, special effects in which a sticker or a label is added to an image frame of the target video; or beautifying or stylizing the object in the image frame, blurring, clearing, adding noise or halation, changing the tone or perspective effect of the image frame, and making the image frame show 3d effect. Among other effects, the effects of the decal or label include, but are not limited to, literal effects, outline effects, and graffiti effects.
The second characteristic information of the target video material may be, for example, descriptive information of content reflected by the target video material. The second characteristic information is used for judging whether to synthesize an editing element with the target video. Illustratively, the second characteristic information may include at least one of: the picture comprises a picture inner part, a picture created atmosphere, a picture main body object, a picture reflected emotion, a picture style and a picture scene.
In some scenarios, the first feature information and the second feature information of the target video material may include identical content. In other scenarios, the first feature information and the second feature information of the target video material may not include exactly the same content. Illustratively, the first characteristic information includes an inside of the picture, an atmosphere created by the picture, a scene of the picture, and a picture subject. The second characteristic information includes a picture subject, a picture-reflected emotion, a picture style, and a scene of the picture.
Alternatively, the target video material may be input into an image classification model to obtain second feature information of the target video material. And determining the correlation degree of the target video material and each editing element in the editing element library, and determining the editing element with the correlation degree higher than the set degree threshold as the target editing element corresponding to the video material. And synthesizing the target editing element with the target video. And the video preview page is used for displaying the target video synthesized with the target editing element. And the target video synthesized with the target editing element is released.
Optionally, a display style of the highlighted text may also be preset. After S120, the video generating method further includes: synthesizing the highlight text and the target video; the display style of the highlighted text is consistent with the preset display style. The video preview page is used to present a target video that includes highlighted text. Published is a target video that includes highlighted text.
It will be appreciated that prior to using the technical solutions disclosed in the embodiments of the present disclosure, the user should be informed and authorized of the type, usage range, usage scenario, etc. of the personal information related to the present disclosure in an appropriate manner according to the relevant legal regulations.
For example, in response to receiving an active request from a user, a prompt is sent to the user to explicitly prompt the user that the operation it is requesting to perform will require personal information to be obtained and used with the user. Thus, the user can autonomously select whether to provide personal information to software or hardware such as an electronic device, an application program, a server or a storage medium for executing the operation of the technical scheme of the present disclosure according to the prompt information.
As an alternative but non-limiting implementation, in response to receiving an active request from a user, the manner in which the prompt information is sent to the user may be, for example, a popup, in which the prompt information may be presented in a text manner. In addition, a selection control for the user to select to provide personal information to the electronic device in a 'consent' or 'disagreement' manner can be carried in the popup window.
It will be appreciated that the above-described notification and user authorization process is merely illustrative and not limiting of the implementations of the present disclosure, and that other ways of satisfying relevant legal regulations may be applied to the implementations of the present disclosure.
It should be noted that, for simplicity of description, the foregoing method embodiments are all described as a series of acts, but it should be understood by those skilled in the art that the present invention is not limited by the order of acts described, as some steps may be performed in other orders or concurrently in accordance with the present invention. Further, those skilled in the art will also appreciate that the embodiments described in the specification are all preferred embodiments, and that the acts and modules referred to are not necessarily required for the present invention.
Fig. 8 is a schematic structural diagram of a video generating apparatus in an embodiment of the present disclosure. The video generating apparatus provided by the embodiment of the present disclosure may be configured in a client or may be configured in a server. Referring to fig. 8, the video generating apparatus specifically includes:
a first display module 310, configured to display a target page; the target page comprises video generation options;
A second display module 320, configured to display an information collection page in response to a selection operation of the video generation option; the information collection page is used for collecting video creation destination information of the generated video;
a third display module 330, configured to display a video preview page according to the video creation destination information input in the information collection page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information.
Optionally, the device further comprises a publishing module, which is used for responding to the selection operation of the publishing rule configuration option after the video preview page is displayed, and displaying the publishing rule configuration page; the release rule configuration page is used for configuring release rules; the release rule configuration options are located in the target page;
and distributing the target video according to the distribution rule input in the distribution rule configuration page of the target video.
Further, the device also comprises a fourth display module for:
if the material consumption proportion of the video material library is greater than or equal to a set threshold value, displaying prompt information for prompting the supplement of video materials; the material consumption ratio is a ratio of the number of video materials that have been applied to generate the target video to the total number of video materials in the video material library.
Further, the apparatus further comprises a video generation module for:
determining a video script based on the video creation destination information;
selecting a target video material from the video material library based on the video script;
and synthesizing the target video material according to the video script to obtain a target video.
Further, the video generation module is used for:
and inputting the video creation destination information into a video script generation model to obtain a video script corresponding to the video creation destination.
Further, the video generation module is further configured to:
acquiring the play data of the target video after being released;
and adjusting the video script generation model according to the play data.
Further, the video script comprises a plurality of shot scripts; the shot script comprises shot content information; the video generation module is further used for:
determining first characteristic information of each video material in the video material library;
matching the shot content information in the shot script with first characteristic information of the video material;
and taking the video material in the matching as a target video material.
Further, the video generation module is further configured to:
synthesizing the target video materials according to the video script, and determining target music according to the video creation destination information after obtaining target video;
synthesizing the target music and the target video; the video preview page is used for displaying the target video synthesized with the target music.
Further, the video generation module is further configured to:
synthesizing the target video material according to the video script, and determining a target editing element in the editing element library according to the second characteristic information of the target video material and the matching relation of each editing element in the editing element library after obtaining the target video;
and synthesizing the target editing element with the target video, wherein the video preview page is used for displaying the target video synthesized with the target editing element.
The video generating device provided in the embodiment of the present disclosure may perform the steps performed by the client in the video generating method provided in the embodiment of the present disclosure, and have the performing steps and beneficial effects, which are not described herein again.
Fig. 9 is a schematic structural diagram of an electronic device in an embodiment of the disclosure. Referring now in particular to fig. 9, a schematic diagram of an electronic device 1000 suitable for use in implementing embodiments of the present disclosure is shown. The electronic device 1000 in the embodiments of the present disclosure may include, but is not limited to, mobile terminals such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g., in-vehicle navigation terminals), wearable electronic devices, and the like, and fixed terminals such as digital TVs, desktop computers, smart home devices, and the like. The electronic device shown in fig. 9 is merely an example, and should not impose any limitations on the functionality and scope of use of embodiments of the present disclosure.
As shown in fig. 9, the electronic apparatus 1000 may include a processing device (e.g., a central processing unit, a graphic processor, etc.) 1001 that may perform various appropriate actions and processes according to a program stored in a Read Only Memory (ROM) 1002 or a program loaded from a storage device 1008 into a Random Access Memory (RAM) 1003 to implement a video generation method of an embodiment as described in the present disclosure. In the RAM 1003, various programs and information necessary for the operation of the electronic apparatus 1000 are also stored. The processing device 1001, the ROM 1002, and the RAM 1003 are connected to each other by a bus 1004. An input/output (I/O) interface 1005 is also connected to bus 1004.
In general, the following devices may be connected to the I/O interface 1005: input devices 1006 including, for example, a touch screen, touchpad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, and the like; an output device 1007 including, for example, a Liquid Crystal Display (LCD), speaker, vibrator, etc.; storage 1008 including, for example, magnetic tape, hard disk, etc.; and communication means 1009. The communication means 1009 may allow the electronic device 1000 to communicate wirelessly or by wire with other devices to exchange information. While fig. 9 shows an electronic device 1000 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may be implemented or provided instead.
In particular, according to embodiments of the present disclosure, the processes described above with reference to flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a non-transitory computer readable medium, the computer program comprising program code for performing the method shown in the flowchart, thereby implementing the video generation method as described above. In such an embodiment, the computer program may be downloaded and installed from a network via the communication device 1009, or installed from the storage device 1008, or installed from the ROM 1002. The above-described functions defined in the method of the embodiment of the present disclosure are performed when the computer program is executed by the processing device 1001.
It should be noted that the computer readable medium described in the present disclosure may be a computer readable signal medium or a computer readable storage medium, or any combination of the two. The computer readable storage medium can be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the foregoing. More specific examples of the computer-readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this disclosure, a computer-readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In the present disclosure, however, the computer-readable signal medium may include an information signal propagated in baseband or as part of a carrier wave, with the computer-readable program code embodied therein. Such a propagated signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination of the foregoing. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, fiber optic cables, RF (radio frequency), and the like, or any suitable combination of the foregoing.
In some implementations, the clients, servers may communicate using any known or future developed network protocol, such as HTTP (HyperText Transfer Protocol ), and may be interconnected with digital information communication (e.g., a communication network) in any form or medium. Examples of communication networks include a local area network ("LAN"), a wide area network ("WAN"), the internet (e.g., the internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks), as well as any known or future developed networks.
The computer readable medium may be contained in the electronic device; or may exist alone without being incorporated into the electronic device.
The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to:
displaying a target page; the target page comprises video generation options;
responding to the selection operation of the video generation options, and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video;
displaying a video preview page according to video creation destination information input in the information collection page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information.
Alternatively, the electronic device may perform other steps described in the above embodiments when the above one or more programs are executed by the electronic device.
Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including, but not limited to, an object oriented programming language such as Java, smalltalk, C ++ and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or may be connected to an external computer (for example, through the Internet using an Internet service provider).
The flowcharts and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The units involved in the embodiments of the present disclosure may be implemented by means of software, or may be implemented by means of hardware. Wherein the names of the units do not constitute a limitation of the units themselves in some cases.
The functions described above herein may be performed, at least in part, by one or more hardware logic components. For example, without limitation, exemplary types of hardware logic components that may be used include: a Field Programmable Gate Array (FPGA), an Application Specific Integrated Circuit (ASIC), an Application Specific Standard Product (ASSP), a system on a chip (SOC), a Complex Programmable Logic Device (CPLD), and the like.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. The machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
According to one or more embodiments of the present disclosure, the present disclosure provides an electronic device comprising:
one or more processors;
a memory for storing one or more programs;
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement any of the video generation methods as provided by the present disclosure.
According to one or more embodiments of the present disclosure, the present disclosure provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, implements a video generation method as any one of the present disclosure provides.
The disclosed embodiments also provide a computer program product comprising a computer program or instructions which, when executed by a processor, implements a video generation method as described above.
It should be noted that in this document, relational terms such as "first" and "second" and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Moreover, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element.
The foregoing is merely a specific embodiment of the disclosure to enable one skilled in the art to understand or practice the disclosure. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the disclosure. Thus, the present disclosure is not intended to be limited to the embodiments shown and described herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (12)

1. A video generation method, comprising:
displaying a target page; the target page comprises video generation options;
responding to the selection operation of the video generation options, and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video;
displaying a video preview page according to video creation destination information input in the information collection page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information.
2. The method of claim 1, wherein after the presenting the video preview page, the method further comprises:
responding to the selection operation of the release rule configuration options, and displaying a release rule configuration page; the release rule configuration page is used for configuring release rules; the release rule configuration options are located in the target page;
and distributing the target video according to the distribution rule input in the distribution rule configuration page.
3. The method as recited in claim 1, further comprising:
if the material consumption proportion of the video material library is greater than or equal to a set threshold value, displaying prompt information for prompting the supplement of video materials; the material consumption ratio is a ratio of the number of video materials that have been applied to generate the target video to the total number of video materials in the video material library.
4. The method of claim 1, wherein the method of generating the target video comprises:
determining a video script based on the video creation destination information;
selecting a target video material from the video material library based on the video script;
and synthesizing the target video material according to the video script to obtain a target video.
5. The method of claim 4, wherein determining a video script based on the video creation destination information comprises:
and inputting the video creation destination information into a video script generation model to obtain a video script corresponding to the video creation destination.
6. The method as recited in claim 5, further comprising:
acquiring the play data of the target video after being released;
and adjusting the video script generation model according to the play data.
7. The method of claim 4, wherein the video script comprises a plurality of shot scripts; the shot script comprises shot content information; the selecting, based on the video script, a target video material from the video material library, including:
determining first characteristic information of each video material in the video material library;
matching the shot content information in the shot script with first characteristic information of the video material;
and taking the video material in the matching as a target video material.
8. The method of claim 4, wherein the synthesizing the target video material according to the video script, after obtaining a target video, further comprises:
Determining target music according to the video creation destination information;
synthesizing the target music and the target video; the video preview page is used for displaying the target video synthesized with the target music.
9. The method of claim 4, wherein the synthesizing the target video material according to the video script, after obtaining a target video, further comprises:
determining a target editing element in the editing element library according to the second characteristic information of the target video material and the matching relation of all editing elements in the editing element library;
and synthesizing the target editing element with the target video, wherein the video preview page is used for displaying the target video synthesized with the target editing element.
10. A video generating apparatus, comprising:
the first display module is used for displaying the target page; the target page comprises video generation options;
the second display module is used for responding to the selection operation of the video generation options and displaying an information collection page; the information collection page is used for collecting video creation destination information of the generated video;
the third display module is used for displaying a video preview page according to the video creation destination information input in the information collection page; the video preview page is used for displaying a target video; the target video is generated by editing video materials in a video material library based on the video creation destination information.
11. An electronic device, the electronic device comprising:
one or more processors;
a storage means for storing one or more programs;
the one or more programs, when executed by the one or more processors, cause the one or more processors to implement the method of any of claims 1-9.
12. A computer readable storage medium, on which a computer program is stored, characterized in that the program, when being executed by a processor, implements the method according to any one of claims 1-9.
CN202311416310.7A 2023-10-27 2023-10-27 Video generation method, device, electronic equipment and storage medium Pending CN117425033A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311416310.7A CN117425033A (en) 2023-10-27 2023-10-27 Video generation method, device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311416310.7A CN117425033A (en) 2023-10-27 2023-10-27 Video generation method, device, electronic equipment and storage medium

Publications (1)

Publication Number Publication Date
CN117425033A true CN117425033A (en) 2024-01-19

Family

ID=89529915

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311416310.7A Pending CN117425033A (en) 2023-10-27 2023-10-27 Video generation method, device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN117425033A (en)

Similar Documents

Publication Publication Date Title
CN107251006B (en) Gallery of messages with shared interests
US8229800B2 (en) System and method for an enhanced shopping experience
WO2019171128A1 (en) In-media and with controls advertisement, ephemeral, actionable and multi page photo filters on photo, automated integration of external contents, automated feed scrolling, template based advertisement post and actions and reaction controls on recognized objects in photo or video
US20120311623A1 (en) Methods and systems for obtaining still images corresponding to video
US20170220591A1 (en) Modular search object framework
US20100312596A1 (en) Ecosystem for smart content tagging and interaction
US20150317354A1 (en) Intent based search results associated with a modular search object framework
CN111435999A (en) Method, device, equipment and storage medium for displaying information on video
CN113792181A (en) Video recommendation method, device, equipment and medium
US9589296B1 (en) Managing information for items referenced in media content
US10440435B1 (en) Performing searches while viewing video content
CN114071179B (en) Live broadcast preview method, device, equipment and medium
US20160307599A1 (en) Methods and Systems for Creating, Combining, and Sharing Time-Constrained Videos
EP2940607A1 (en) Enhanced search results associated with a modular search object framework
US20200250369A1 (en) System and method for transposing web content
CN115190366B (en) Information display method, device, electronic equipment and computer readable medium
Mishra et al. A beginner's guide to mobile marketing
CN113849682A (en) Video searching method, device, equipment and medium
KR20160027486A (en) Apparatus and method of providing advertisement, and apparatus and method of displaying advertisement
US20200037034A1 (en) System and Method for Navigating in a Digital Environment
US20220005129A1 (en) Methods and systems for interaction with videos and other media files
CN114846812A (en) Abstract video generation method and device and server
KR102553332B1 (en) Method and apparatus for editing content on a live broadcasting platform
CN115860869A (en) Shop information recommendation method, equipment and storage medium
CN117425033A (en) Video generation method, device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination