CN112541353A - Video generation method, device, equipment and medium - Google Patents

Video generation method, device, equipment and medium Download PDF

Info

Publication number
CN112541353A
CN112541353A CN202011550514.6A CN202011550514A CN112541353A CN 112541353 A CN112541353 A CN 112541353A CN 202011550514 A CN202011550514 A CN 202011550514A CN 112541353 A CN112541353 A CN 112541353A
Authority
CN
China
Prior art keywords
segment
video
target material
word segmentation
segments
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011550514.6A
Other languages
Chinese (zh)
Inventor
郭瑞彪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202011550514.6A priority Critical patent/CN112541353A/en
Publication of CN112541353A publication Critical patent/CN112541353A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The present disclosure provides a video generation method, apparatus, device and medium, which relate to the technical field of multimedia, and in particular to the technical field of video processing and natural language processing. The implementation scheme is as follows: establishing a material library, wherein the material library comprises one or more material segments; receiving an input generation instruction; performing word segmentation on the generation instruction to obtain a first word segmentation set, wherein the first word segmentation set comprises at least one word segmentation; aiming at the first word segmentation set, acquiring target material segments respectively related to at least one word segmentation from a material library; editing each target material segment according to a preset rule; and generating a video based on the edited target material segment.

Description

Video generation method, device, equipment and medium
Technical Field
The present disclosure relates to the field of multimedia technologies, and in particular, to the field of video processing and natural language processing technologies, and in particular, to a method, an apparatus, a device, and a medium for generating a video.
Background
With the rapid development of the video industry and the popularization of intelligent equipment, more and more video resources can conveniently enter the visual field of the masses, meanwhile, a plurality of people begin to perform secondary creation on the video resources, and creative videos called as ghost animals are deeply loved and touted by young people. But the creative video has higher creation cost, extremely large workload and longer video production time.
The approaches described in this section are not necessarily approaches that have been previously conceived or pursued. Unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section. Similarly, unless otherwise indicated, the problems mentioned in this section should not be considered as having been acknowledged in any prior art.
Disclosure of Invention
The present disclosure provides a video generation method, apparatus, device, and medium.
According to an aspect of the present disclosure, there is provided a video generation method including: establishing a material library, wherein the material library comprises one or more material segments; receiving an input generation instruction; performing word segmentation on the generation instruction to obtain a first word segmentation set of the generation instruction, wherein the first word segmentation set comprises at least one word segmentation; aiming at the first word segmentation set, acquiring target material segments respectively related to the at least one word segmentation from the material library; editing each target material segment according to a preset rule; and generating a video aiming at the generation instruction based on the edited target material segment.
According to another aspect of the present disclosure, there is provided a video generating apparatus including: an establishing unit configured to establish a material library, the material library including one or more material segments; a receiving unit configured to receive an input generation instruction; the word segmentation unit is configured to segment the generated instruction to obtain a first word segmentation set of the generated instruction, wherein the first word segmentation set comprises at least one word segmentation; an obtaining unit, configured to obtain, for the first segmentation set, target material segments respectively related to the at least one segmentation from the material library; the editing unit is configured for editing each target material segment according to a preset rule; and a generation unit configured to generate a video for the generation instruction based on the edited target material segment.
According to another aspect of the present disclosure, there is also provided an electronic device including: a processor; and a memory storing a program comprising instructions which, when executed by the processor, cause the processor to perform a video management method according to the above.
According to another aspect of the present disclosure, there is also provided a computer-readable storage medium storing a program, the program comprising instructions that, when executed by a processor of an electronic device, cause the electronic device to perform the video management method according to the above.
According to another aspect of the present disclosure, there is also provided a computer program product comprising a computer program, wherein the computer program realizes the above-mentioned method when executed by a processor.
According to one or more embodiments of the present disclosure, a video producer can be freed from tedious and tedious work of extracting target segments from a large amount of video materials, thereby lowering production thresholds and improving content quality of creative videos.
It should be understood that the statements in this section do not necessarily identify key or critical features of the embodiments of the present disclosure, nor do they limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate exemplary embodiments of the embodiments and, together with the description, serve to explain the exemplary implementations of the embodiments. The illustrated embodiments are for purposes of illustration only and do not limit the scope of the claims. Throughout the drawings, identical reference numbers designate similar, but not necessarily identical, elements.
Fig. 1 shows a flow diagram of a video generation method according to an embodiment of the present disclosure;
FIG. 2 shows a flow diagram for building a material library according to an embodiment of the present disclosure;
FIG. 3 illustrates a flow diagram of obtaining target material segments from a material library that are respectively associated with at least one participle in accordance with an embodiment of the present disclosure;
FIG. 4 shows a flow diagram for generating a video according to an embodiment of the present disclosure;
fig. 5 shows a block diagram of an apparatus for video generation according to an embodiment of the present disclosure;
FIG. 6 illustrates a block diagram of an exemplary electronic device that can be used to implement embodiments of the present disclosure.
Detailed Description
Exemplary embodiments of the present disclosure are described below with reference to the accompanying drawings, in which various details of the embodiments of the disclosure are included to assist understanding, and which are to be considered as merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope of the present disclosure. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
In the present disclosure, unless otherwise specified, the use of the terms "first", "second", etc. to describe various elements is not intended to limit the positional relationship, the timing relationship, or the importance relationship of the elements, and such terms are used only to distinguish one element from another. In some examples, a first element and a second element may refer to the same instance of the element, and in some cases, based on the context, they may also refer to different instances.
The terminology used in the description of the various described examples in this disclosure is for the purpose of describing particular examples only and is not intended to be limiting. Unless the context clearly indicates otherwise, if the number of elements is not specifically limited, the elements may be one or more. Furthermore, the term "and/or" as used in this disclosure is intended to encompass any and all possible combinations of the listed items.
Embodiments of the present disclosure will be described in detail below with reference to the accompanying drawings.
Fig. 1 is a flowchart illustrating a video generation method according to an exemplary embodiment of the present disclosure. As shown in fig. 1, the video generation method may include: step S101, establishing a material library, wherein the material library comprises one or more material segments; step S102, receiving an input generation instruction; step S103, performing word segmentation on the generation instruction to obtain a first word segmentation set of the generation instruction, wherein the first word segmentation set comprises at least one word segmentation; step S104, aiming at the first participle set, obtaining target material segments respectively related to the at least one participle from the material library; s105, editing each target material segment according to a preset rule; and S106, generating a video aiming at the generation instruction based on the edited target material segment.
In the video generation method according to the embodiment of the disclosure, a word segmentation set of a generation instruction is obtained based on establishing a material library including a plurality of material segments, and performing word segmentation on the generation instruction, and a related target material segment is automatically obtained in the material library, so that a video is generated according to a preset rule. The method and the device can liberate a video producer from tedious and boring work of extracting target segments from a large amount of video materials, reduce production cost and threshold, and improve content quality of creative videos.
The server may pre-establish a repository. For example, when the generation instruction is: when the young people do not speak Wude, the generation instruction is segmented to obtain a first segmentation set of the generation instruction. The first set of terms may include "young", "not", "talk", "armed". Target material segments respectively related to the young people, the non-attended people, the talkback and the Wude can be obtained from a material library; sequencing the target material segments according to a preset rule, for example, a text sequence corresponding to the generation instruction or a user-specified sequence; and generating a final video.
According to some embodiments, as shown in fig. 2, the step S101 of establishing a material library may include: step S1011, acquiring one or more original videos; step S1012, obtaining a second word set of each of the one or more original videos, where the second word set includes at least one word; step S1013, based on the corresponding second word set, dividing the one or more original videos so that each original video is divided into one or more material segments, where each word included in the second word set corresponds to one of the material segments of the corresponding original video; step S1014, determining segment identifications corresponding to the plurality of material segments obtained by segmentation, wherein the segment identifications comprise at least one word segmentation corresponding to the material segments; step S1015, building the material library based on the plurality of material segments of the one or more original videos and the segment identifiers thereof. Therefore, the original video is automatically cut into a plurality of material segments according to the segmentation in the second segmentation set, and the material segments meeting the requirements of the user can be quickly and accurately obtained. The video producer no longer needs to manually segment the required segments from the original video, saving a lot of time.
For example, the acquired original video may be a video excerpt corresponding to a line excerpt of a certain star in a movie. Taking the example of the phrase section as "young people always do not understand one of the reasons", the obtained second set of words may include the words "young people", "always", "not", "understand", "one", "the reason". And cutting the original video into six corresponding material segments according to the six word segments, wherein each segment corresponds to one word segment, and the corresponding word segment is used as one of segment marks of the material segment. The material segments with segment identifications (i.e., corresponding participles) are stored in a material library.
According to some embodiments, the step S1011 of acquiring one or more original videos may include: receiving an input keyword instruction; collecting the plurality of original videos related to the keyword instruction based on a web crawler tool; one or more original videos are obtained from the plurality of original videos based on the set number. Therefore, a large number of original videos related to the keywords can be obtained, and video resources of the material library are enriched.
According to some embodiments, the step S1011 of acquiring one or more original videos may further include: and acquiring the one or more original videos by a direct import mode. By the method, the specific video which the user is interested in can be directly imported, so that the material library is more targeted and unique.
According to some embodiments, the segment identifications for each segment of material in the material library may also include one or more filter information, as shown in FIG. 3. In this case, the step S104, for the first segmentation set, obtaining target material segments respectively related to the at least one segmentation from the material library may include: step S1041, determining one or more related participles matched with the participles in the first participle set in the plurality of segment identifiers of the material library; step S1042, based on the selected filtering information, selecting at least one material segment from a plurality of material segments corresponding to the one or more related participles, and determining the at least one material segment as a related material segment; step S1043, determining a target story segment of each word segmentation in the first word segmentation set from the corresponding at least one related story segment. Therefore, the screening efficiency of the material segments is improved through automatic matching based on the first segmentation and the segment identification. And the most satisfactory material segments of the user can be quickly selected according to the filtering information, so that the video generation quality is improved.
According to some embodiments, the filtering information may include, but is not limited to, at least one of the following: video content, video people, video popularity, and semantic information.
For example, for a first set of terms comprising "youth", "not", "speak", "martial" terms, respective segments of material in the material library are determined that are identified as "youth", "not", "speak", "martial", i.e., at least one relevant segment of material from the material library is determined that is identified as "youth", at least one relevant segment of material is determined that is identified as "not", at least one relevant segment of material is determined that is identified as "speak", and at least one relevant segment of material is determined that is identified as "martial". Further, a target material segment determined as "young" may be selected from among at least one related material segment whose segment is identified as "young" based on filtering information (e.g., video heat), and similarly, target material segments of "not", "say", "wude", and "wude", respectively, may also be determined based on corresponding filtering information.
According to some embodiments, the editing each target material segment according to the preset rule at step S105 may include: and setting the playing sequence number of each target material segment according to the sequence of the participles in the generation instruction. And the one or more target material segments can be connected in sequence according to the corresponding playing sequence numbers to generate the video aiming at the generation instruction. Therefore, the time for a video producer to sort and list the target material segments can be saved, and the final video can be directly and simply generated.
For example, the playback order of each target material segment may be set in the text order of the generation instruction "young person does not speak wude", and the target material segments corresponding to "young person", "not", "speaking", "wude" may be connected in order.
According to some embodiments, the editing each target material segment according to the preset rule in step S105 may further include: and setting the playing sequence number of each target material segment based on the received sequence number setting instruction, wherein the one or more target material segments are sequentially connected according to the corresponding playing sequence number to generate a video aiming at the generation instruction. Therefore, the video producer can conveniently sort and list the target material fragments according to the favor and the creative idea, so that the final video is richer, and the user requirements are met.
For example, if a ghost video of "wude-lectured young person" is generated in accordance with a generation instruction "young person is not wude-lectured", and the video creator wants to embody a creative in the arrangement order, the target material segments whose segments are identified as "young person" may be arranged at the end, and the target material segments corresponding to "not", "speaking", "wude", and "young person" may be connected in order.
According to some embodiments, as shown in fig. 4, the step S106 of generating a video for the generation instruction based on the edited target material segment may include: step S1061, generating a preview video based on the edited target material segment; step S1062, in response to the received editing request, re-editing at least one target material segment to generate a new preview video; and step S1063, outputting a final video based on the generated new preview video. Therefore, through the editing request, manual intervention on the automatically generated preview video can be realized, the trial and error cost is reduced, and the video production efficiency is improved.
According to some embodiments, step S1062, in response to receiving an editing request, re-editing at least one target material segment to generate a new preview video, where the editing request includes at least one of the following requests: cropping the length of at least one of the target material segments, deleting at least one of the target material segments, replacing at least one of the target material segments, inserting a new one or more of the target material segments, repeating at least one of the target material segments, adjusting the playback parameters of at least one of the target material segments, and adjusting the playback order of the one or more target material segments. Therefore, the user can better highlight the required ghost livestock effect and video creative by selecting different editing modes according to the requirements, and the quality of creative videos is improved.
For example, based on target material segments corresponding to "young person", "not", "speaking", "wude", which have been connected in sequence, a preview video corresponding to "young person does not speak wude" is generated; the video producer can edit each target material segment as required, for example, repeat the target material segment corresponding to the "young" 3 times, regenerate a new preview video, and output the final video.
For example, based on target material segments corresponding to "young person", "not", "speaking", "wude", which have been connected in sequence, a preview video corresponding to "young person does not speak wude" is generated; the video producer may insert new target material segment(s) between target material segments corresponding to "young people" and target material segments corresponding to "not", e.g., target material segments corresponding to "people", as desired.
According to another aspect of the present disclosure, a video generating apparatus is also provided. As shown in fig. 5, the video generating apparatus 500 may include: an establishing unit 501 configured to establish a material library, the material library including one or more material segments; a receiving unit 502 configured to receive a generation instruction input by a user; a word segmentation unit 503 configured to perform word segmentation on the generation instruction to obtain a first word segmentation set of the generation instruction, where the first word segmentation set includes at least one word segmentation; an obtaining unit 504, configured to obtain, for the first segmentation set, target material segments respectively related to the at least one segmentation from the material library; an editing unit 505 configured to edit each target material segment according to a preset rule; and a generating unit 506 configured to generate a video for the generation instruction based on the edited target material segment.
Here, the operations of the above units 501 and 506 of the video generating apparatus 500 are similar to the operations of the steps S101 to S106 described above, and are not described again here.
According to some embodiments, the establishing unit may comprise: an acquisition subunit configured to acquire one or more original videos; a word segmentation subunit configured to obtain a second word set of each of the one or more original videos, the second word set including at least one word segmentation; a segmentation subunit configured to segment the one or more original videos based on the corresponding second set of terms such that each original video is segmented into one or more material segments, wherein each term included in the second set of terms corresponds to one of the material segments of the corresponding original video; the first determining subunit is configured to determine segment identifiers corresponding to the plurality of material segments obtained through segmentation, wherein the segment identifiers include at least one word segmentation corresponding to the material segments; a creating subunit configured to create the material library based on a plurality of material segments of the one or more original videos and segment identifications thereof.
Here, the operations of the above sub-units of the establishing unit are similar to the operations of steps S1011 to S1015 described above, and are not described again here.
According to some embodiments, wherein the segment identity may further comprise one or more filtering information. In this case, the acquiring unit may include: a matching subunit configured to determine one or more related participles of the plurality of segment identifications of the corpus that match the participles in the first participle set; a selecting subunit, configured to select at least one material segment from a plurality of material segments corresponding to the one or more related participles based on the selected filtering information, and determine the at least one material segment as a related material segment; a second determining subunit configured to determine a target material segment for each participle in the first participle set from the respective at least one related material segment.
Here, the operations of the foregoing sub-units of the acquiring unit are respectively similar to the operations of steps S1041 to S1043 described above, and are not described again here.
According to some embodiments, wherein the generating unit is configured to include: a generation subunit configured to generate a preview video based on the edited target material segment; an editing subunit configured to, in response to receiving the editing request, re-edit the at least one target material segment to generate a new preview video; an output subunit configured to output a final video based on the preview video.
Here, the operations of the above sub-units of the generating unit are similar to the operations of steps S1061-S1063, respectively, and are not described again here. According to an embodiment of the present disclosure, there is also provided an electronic device, a readable storage medium, and a computer program product.
Referring to fig. 6, a block diagram of a structure of an electronic device 600, which may be a server or a client of the present disclosure, which is an example of a hardware device that may be applied to aspects of the present disclosure, will now be described. Electronic device is intended to represent various forms of digital electronic computer devices, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other suitable computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 6, the apparatus 600 includes a computing unit 601, which can perform various appropriate actions and processes according to a computer program stored in a Read Only Memory (ROM)602 or a computer program loaded from a storage unit 608 into a Random Access Memory (RAM) 603. In the RAM 603, various programs and data required for the operation of the device 600 can also be stored. The calculation unit 601, the ROM 602, and the RAM 603 are connected to each other via a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
A number of components in the device 600 are connected to the I/O interface 605, including: an input unit 606, an output unit 607, a storage unit 608, and a communication unit 609. The input unit 606 may be any type of device capable of inputting information to the device 600, and the input unit 606 may receive input numeric or character information and generate key signal inputs related to user settings and/or function controls of the electronic device, and may include, but is not limited to, a mouse, a keyboard, a touch screen, a track pad, a track ball, a joystick, a microphone, and/or a remote control. Output unit 607 may be any type of device capable of presenting information and may include, but is not limited to, a display, speakers, a video/audio output terminal, a vibrator, and/or a printer. The storage unit 608 may include, but is not limited to, a magnetic disk, an optical disk. The communication unit 609 allows the device 600 to exchange information/data with other devices via a computer network, such as the internet, and/or various telecommunications networks, and may include, but is not limited to, a modem, a network card, an infrared communication device, a wireless communication transceiver, and/or a chipset, such as a bluetooth (TM) device, an 1302.11 device, a WiFi device, a WiMax device, a cellular communication device, and/or the like.
The computing unit 601 may be a variety of general and/or special purpose processing components having processing and computing capabilities. Some examples of the computing unit 601 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various dedicated Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable processor, controller, microcontroller, and so forth. The calculation unit 601 performs the respective methods and processes described above, such as method video generation. For example, in some embodiments, the method video generation may be implemented as a computer software program tangibly embodied in a machine-readable medium, such as storage unit 608. In some embodiments, part or all of the computer program may be loaded and/or installed onto the device 600 via the ROM 602 and/or the communication unit 609. When the computer program is loaded into the RAM 603 and executed by the computing unit 601, one or more steps of the method video generation described above may be performed. Alternatively, in other embodiments, the computing unit 601 may be configured to perform the method video generation by any other suitable means (e.g., by means of firmware).
Various implementations of the systems and techniques described here above may be implemented in digital electronic circuitry, integrated circuitry, Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (ASSPs), system on a chip (SOCs), load programmable logic devices (CPLDs), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
Program code for implementing the methods of the present disclosure may be written in any combination of one or more programming languages. These program codes may be provided to a processor or controller of a general purpose computer, special purpose computer, or other programmable data processing apparatus, such that the program codes, when executed by the processor or controller, cause the functions/operations specified in the flowchart and/or block diagram to be performed. The program code may execute entirely on the machine, partly on the machine, as a stand-alone software package partly on the machine and partly on a remote machine or entirely on the remote machine or server.
In the context of this disclosure, a machine-readable medium may be a tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. The machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium. A machine-readable medium may include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples of a machine-readable storage medium would include an electrical connection based on one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present disclosure may be performed in parallel, sequentially or in different orders, and are not limited herein as long as the desired results of the technical solutions disclosed in the present disclosure can be achieved.
Although embodiments or examples of the present disclosure have been described with reference to the accompanying drawings, it is to be understood that the above-described methods, systems and apparatus are merely exemplary embodiments or examples and that the scope of the present invention is not limited by these embodiments or examples, but only by the claims as issued and their equivalents. Various elements in the embodiments or examples may be omitted or may be replaced with equivalents thereof. Further, the steps may be performed in an order different from that described in the present disclosure. Further, various elements in the embodiments or examples may be combined in various ways. It is important that as technology evolves, many of the elements described herein may be replaced with equivalent elements that appear after the present disclosure.

Claims (17)

1. A video generation method, comprising:
establishing a material library, wherein the material library comprises one or more material segments;
receiving an input generation instruction;
performing word segmentation on the generation instruction to obtain a first word segmentation set of the generation instruction, wherein the first word segmentation set comprises at least one word segmentation;
aiming at the first word segmentation set, acquiring target material segments respectively related to the at least one word segmentation from the material library;
editing each target material segment according to a preset rule; and
and generating a video aiming at the generation instruction based on the edited target material segment.
2. The method of claim 1, wherein said creating a stories library comprises:
acquiring one or more original videos;
obtaining a second word set of each original video in the one or more original videos, wherein the second word set comprises at least one word;
based on the corresponding second word segmentation set, segmenting the one or more original videos to enable each original video to be segmented into one or more material segments, wherein each word segmentation included in the second word segmentation set corresponds to one material segment of the corresponding original video;
determining segment identifications corresponding to a plurality of material segments obtained by segmentation respectively, wherein the segment identifications comprise at least one word segmentation corresponding to the material segments;
and establishing the material library based on a plurality of material segments of the one or more original videos and segment identifications thereof.
3. The method of claim 2, wherein the segment identification further comprises one or more filter information,
wherein obtaining target material segments respectively associated with the at least one participle from the material library comprises:
determining one or more related participles in a plurality of segment identifications of the material library, wherein the related participles are matched with the participles in the first participle set;
selecting at least one material segment from a plurality of material segments corresponding to the one or more related participles based on the selected filtering information, and determining the at least one material segment as a related material segment;
a target story segment for each word in the first set of words is determined from the respective at least one related story segment.
4. The method of claim 2, wherein said obtaining one or more raw videos comprises:
receiving an input keyword instruction;
collecting a plurality of original videos related to the keyword instruction based on a web crawler tool;
one or more original videos are obtained from the plurality of original videos based on the set number.
5. The method of claim 2, wherein the one or more original videos are obtained by direct import.
6. The method of claim 3, wherein the filtering information comprises at least one of:
video content, video people, video popularity, and semantic information.
7. The method of claim 1, wherein the generating a video comprises:
generating a preview video based on the edited target material segment;
in response to receiving the editing request, re-editing the at least one target material segment to generate a new preview video;
outputting a final video based on the generated new preview video.
8. The method of claim 7, wherein the edit request comprises at least one of:
cropping the length of at least one of the target material segments, deleting at least one of the target material segments, replacing at least one of the target material segments, inserting a new one or more of the target material segments, repeating at least one of the target material segments, adjusting the playback parameters of at least one of the target material segments, and adjusting the playback order of the one or more target material segments.
9. The method of claim 1, wherein the preset rule comprises:
setting the playing sequence number of each target material segment according to the sequence of the participles in the generating instruction,
and sequentially connecting the one or more target material segments according to the corresponding playing sequence numbers to generate the video aiming at the generation instruction.
10. The method of claim 1, wherein the preset rule comprises:
setting a play sequence number of each of the target material segments based on the received sequence number setting instruction,
and sequentially connecting the one or more target material segments according to the corresponding playing sequence numbers to generate the video aiming at the generation instruction.
11. A video generation apparatus comprising:
an establishing unit configured to establish a material library, the material library including one or more material segments;
a receiving unit configured to receive an input generation instruction;
the word segmentation unit is configured to segment the generated instruction to obtain a first word segmentation set of the generated instruction, wherein the first word segmentation set comprises at least one word segmentation;
an obtaining unit, configured to obtain, for the first segmentation set, target material segments respectively related to the at least one segmentation from the material library;
the editing unit is configured for editing each target material segment according to a preset rule; and
a generating unit configured to generate a video for the generation instruction based on the edited target material segment.
12. The apparatus of claim 11, wherein the establishing means comprises:
an acquisition subunit configured to acquire one or more original videos;
a word segmentation subunit configured to obtain a second word set of each of the one or more original videos, the second word set including at least one word segmentation;
a segmentation subunit configured to segment the one or more original videos based on the corresponding second set of terms such that each original video is segmented into one or more material segments, wherein each term included in the second set of terms corresponds to one of the material segments of the corresponding original video;
the first determining subunit is configured to determine segment identifiers corresponding to the plurality of material segments obtained through segmentation, wherein the segment identifiers include at least one word segmentation corresponding to the material segments;
a creating subunit configured to create the material library based on a plurality of material segments of the one or more original videos and segment identifications thereof.
13. The apparatus of claim 12, wherein the segment identification further comprises one or more filtering information,
wherein the acquisition unit includes:
a matching subunit configured to determine one or more related participles of the plurality of segment identifications of the corpus that match the participles in the first participle set;
a selecting subunit, configured to select at least one material segment from a plurality of material segments corresponding to the one or more related participles based on the selected filtering information, and determine the at least one material segment as a related material segment;
a second determining subunit configured to determine a target material segment for each participle in the first participle set from the respective at least one related material segment.
14. The apparatus of claim 11, wherein the generating unit comprises:
a generation subunit configured to generate a preview video based on the edited target material segment;
an editing subunit configured to, in response to receiving the editing request, re-edit the at least one target material segment to generate a new preview video;
an output subunit configured to output a final video based on the preview video.
15. An electronic device, comprising:
a processor; and
a memory storing a program comprising instructions that, when executed by the processor, cause the processor to perform the method of any of claims 1-10.
16. A computer readable storage medium storing a program, the program comprising instructions that when executed by a processor of an electronic device cause the electronic device to perform the method of any of claims 1-10.
17. A computer program product comprising a computer program, wherein the computer program realizes the method of any one of claims 1-10 when executed by a processor.
CN202011550514.6A 2020-12-24 2020-12-24 Video generation method, device, equipment and medium Pending CN112541353A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011550514.6A CN112541353A (en) 2020-12-24 2020-12-24 Video generation method, device, equipment and medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011550514.6A CN112541353A (en) 2020-12-24 2020-12-24 Video generation method, device, equipment and medium

Publications (1)

Publication Number Publication Date
CN112541353A true CN112541353A (en) 2021-03-23

Family

ID=75017469

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011550514.6A Pending CN112541353A (en) 2020-12-24 2020-12-24 Video generation method, device, equipment and medium

Country Status (1)

Country Link
CN (1) CN112541353A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113364999A (en) * 2021-05-31 2021-09-07 北京达佳互联信息技术有限公司 Video generation method and device, electronic equipment and storage medium
CN113361284A (en) * 2021-06-30 2021-09-07 北京百度网讯科技有限公司 Target content generation method and device
CN113556486A (en) * 2021-07-27 2021-10-26 北京达佳互联信息技术有限公司 Video generation method and device, electronic equipment and storage medium
CN113630644A (en) * 2021-06-29 2021-11-09 北京搜狗科技发展有限公司 Editing method and device of video content editor and storage medium
CN115134659A (en) * 2022-06-15 2022-09-30 阿里巴巴云计算(北京)有限公司 Video editing and configuring method and device, browser, electronic equipment and storage medium
CN115460459A (en) * 2022-09-02 2022-12-09 百度时代网络技术(北京)有限公司 Video generation method and device based on AI (Artificial Intelligence) and electronic equipment
CN115942040A (en) * 2022-12-14 2023-04-07 上海幻电信息科技有限公司 Video processing method and device, storage medium and electronic equipment
CN116634233A (en) * 2023-04-12 2023-08-22 北京优贝卡科技有限公司 Media editing method, device, equipment and storage medium
CN116708945A (en) * 2023-04-12 2023-09-05 北京优贝卡科技有限公司 Media editing method, device, equipment and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110121107A (en) * 2018-02-06 2019-08-13 上海全土豆文化传播有限公司 Video material collection method and device
CN111935537A (en) * 2020-06-30 2020-11-13 百度在线网络技术(北京)有限公司 Music video generation method and device, electronic equipment and storage medium

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110121107A (en) * 2018-02-06 2019-08-13 上海全土豆文化传播有限公司 Video material collection method and device
CN111935537A (en) * 2020-06-30 2020-11-13 百度在线网络技术(北京)有限公司 Music video generation method and device, electronic equipment and storage medium

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113364999B (en) * 2021-05-31 2022-12-27 北京达佳互联信息技术有限公司 Video generation method and device, electronic equipment and storage medium
CN113364999A (en) * 2021-05-31 2021-09-07 北京达佳互联信息技术有限公司 Video generation method and device, electronic equipment and storage medium
CN113630644A (en) * 2021-06-29 2021-11-09 北京搜狗科技发展有限公司 Editing method and device of video content editor and storage medium
CN113630644B (en) * 2021-06-29 2024-01-30 北京搜狗科技发展有限公司 Editing method, device and storage medium of video content editor
CN113361284A (en) * 2021-06-30 2021-09-07 北京百度网讯科技有限公司 Target content generation method and device
CN113361284B (en) * 2021-06-30 2024-03-26 北京百度网讯科技有限公司 Method and device for generating target content
CN113556486A (en) * 2021-07-27 2021-10-26 北京达佳互联信息技术有限公司 Video generation method and device, electronic equipment and storage medium
CN113556486B (en) * 2021-07-27 2024-02-06 北京达佳互联信息技术有限公司 Video generation method, device, electronic equipment and storage medium
CN115134659A (en) * 2022-06-15 2022-09-30 阿里巴巴云计算(北京)有限公司 Video editing and configuring method and device, browser, electronic equipment and storage medium
CN115460459B (en) * 2022-09-02 2024-02-27 百度时代网络技术(北京)有限公司 Video generation method and device based on AI and electronic equipment
CN115460459A (en) * 2022-09-02 2022-12-09 百度时代网络技术(北京)有限公司 Video generation method and device based on AI (Artificial Intelligence) and electronic equipment
CN115942040A (en) * 2022-12-14 2023-04-07 上海幻电信息科技有限公司 Video processing method and device, storage medium and electronic equipment
CN116634233A (en) * 2023-04-12 2023-08-22 北京优贝卡科技有限公司 Media editing method, device, equipment and storage medium
CN116634233B (en) * 2023-04-12 2024-02-09 北京七彩行云数字技术有限公司 Media editing method, device, equipment and storage medium
CN116708945A (en) * 2023-04-12 2023-09-05 北京优贝卡科技有限公司 Media editing method, device, equipment and storage medium
CN116708945B (en) * 2023-04-12 2024-04-16 半月谈新媒体科技有限公司 Media editing method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN112541353A (en) Video generation method, device, equipment and medium
CN103456314B (en) A kind of emotion identification method and device
CN111145737B (en) Voice test method and device and electronic equipment
US8831953B2 (en) Systems and methods for filtering objectionable content
US20080177536A1 (en) A/v content editing
CN107301170B (en) Method and device for segmenting sentences based on artificial intelligence
WO2019001194A1 (en) Voice recognition method, device, apparatus, and storage medium
US10031714B2 (en) Method and device for processing audio files
CN114254158B (en) Video generation method and device, and neural network training method and device
CN113204667B (en) Method and device for training audio annotation model and audio annotation
US20230368448A1 (en) Comment video generation method and apparatus
JP2022120024A (en) Audio signal processing method, model training method, and their device, electronic apparatus, storage medium, and computer program
CN112203140A (en) Video editing method and device, electronic equipment and storage medium
CN114860995B (en) Video script generation method and device, electronic equipment and medium
CN104778221A (en) Music collaborate splicing method and device
CN108334491B (en) Text analysis method and device, computing equipment and storage medium
CN104978377A (en) Multimedia data processing method, multimedia data processing device and terminal
CN113923479A (en) Audio and video editing method and device
CN117171296A (en) Information acquisition method and device and electronic equipment
US20130198181A1 (en) Summarising a Set of Articles
KR20210050410A (en) Method and system for suppoting content editing based on real time generation of synthesized sound for video content
CN114238745A (en) Method and device for providing search result, electronic equipment and medium
CN112925889A (en) Natural language processing method, device, electronic equipment and storage medium
CN104978404A (en) Video album name generating method and apparatus
CN104978401B (en) A kind of the keyword setting method and device of video album

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination