WO2021121023A1 - 视频编辑方法、视频编辑装置、终端和可读存储介质 - Google Patents

视频编辑方法、视频编辑装置、终端和可读存储介质 Download PDF

Info

Publication number
WO2021121023A1
WO2021121023A1 PCT/CN2020/133047 CN2020133047W WO2021121023A1 WO 2021121023 A1 WO2021121023 A1 WO 2021121023A1 CN 2020133047 W CN2020133047 W CN 2020133047W WO 2021121023 A1 WO2021121023 A1 WO 2021121023A1
Authority
WO
WIPO (PCT)
Prior art keywords
duration
video
edited
editing
video segment
Prior art date
Application number
PCT/CN2020/133047
Other languages
English (en)
French (fr)
Inventor
陈佳雨
Original Assignee
Oppo广东移动通信有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oppo广东移动通信有限公司 filed Critical Oppo广东移动通信有限公司
Priority to EP20903904.9A priority Critical patent/EP4068793A4/en
Publication of WO2021121023A1 publication Critical patent/WO2021121023A1/zh
Priority to US17/841,641 priority patent/US20220312048A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/458Scheduling content for creating a personalised stream, e.g. by combining a locally stored advertisement with an incoming stream; Updating operations, e.g. for OS modules ; time-related management operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8455Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • H04N21/41407Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance embedded in a portable device, e.g. video client on a mobile phone, PDA, laptop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8106Monomedia components thereof involving special audio data, e.g. different tracks for different languages
    • H04N21/8113Monomedia components thereof involving special audio data, e.g. different tracks for different languages comprising music, e.g. song in MP3 format

Definitions

  • This application relates to the field of consumer electronics technology, and in particular to a video editing method, a video editing device, a terminal, and a non-volatile computer-readable storage medium.
  • the embodiments of the present application provide a video editing method, a video editing device, a terminal, and a non-volatile computer-readable storage medium.
  • the video editing method of the embodiment of the present application includes marking one or more video clips in the initial video; determining from the one or more video clips according to the interval between two adjacent rhythm points of the preset music clip The to-be-edited video segment, the duration of the to-be-edited video segment is greater than or equal to the interval duration; and the to-be-edited video segment is edited to obtain the edited video segment, with the start time and end time of the edited video segment Correspond to the two rhythm points respectively.
  • the video editing device of the embodiment of the present application includes an acquisition module, a determination module, and an editing module.
  • the acquiring module is used to mark one or more video clips in the initial video;
  • the determining module is used to determine from the one or more video clips according to the interval between two adjacent rhythm points of the preset music clip.
  • the video segment to be edited is determined in the segment, and the duration of the video segment to be edited is greater than or equal to the interval duration;
  • the editing module is used to edit the video segment to be edited to obtain the edited video segment, and use the edited video
  • the start time and end time of the segment respectively correspond to the two rhythm points.
  • the terminal of the embodiment of the present application includes a processor configured to mark one or more video clips in the initial video; according to the interval time between two adjacent rhythm points of the preset music clip, from the one or more A video segment to be edited is determined among a plurality of video segments, and the duration of the to-be-edited video segment is greater than or equal to the interval duration; and the to-be-edited video segment is edited to obtain the edited video segment, and the edited video segment is The start time and the end time correspond to the two rhythm points respectively.
  • a non-volatile computer-readable storage medium containing computer-executable instructions of the present application when the computer-executable instructions are executed by one or more processors, causes the processors to execute a video editing method.
  • the video editing method includes marking one or more video clips in the initial video; and determining the video to be edited from the one or more video clips according to the interval between two adjacent rhythm points of the preset music clip Segment, the duration of the to-be-edited video segment is greater than or equal to the interval duration; and the to-be-edited video segment is edited to obtain the edited video segment, and the start time and end time of the edited video segment respectively correspond to two The rhythm points.
  • FIG. 1 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 2 is a schematic diagram of a terminal structure of some embodiments of the present application.
  • FIG. 3 is a schematic diagram of modules of a video editing device according to some embodiments of the present application.
  • FIG. 4 is a schematic diagram of the principle of a video editing method in some embodiments of the present application.
  • FIG. 5 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 6 is a schematic diagram of modules of a video editing device according to some embodiments of the present application.
  • FIG. 7 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 8 is a schematic diagram of an editing module in some embodiments of the present application.
  • FIG. 9 is a schematic diagram of the principle of a video editing method in some embodiments of the present application.
  • FIG. 10 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 11 is a schematic diagram of an editing module in some embodiments of the present application.
  • FIG. 12 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 13 is a schematic diagram of an editing module in some embodiments of the present application.
  • FIG. 14 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 15 is a schematic diagram of an editing module in some embodiments of the present application.
  • FIG. 16 is a schematic diagram of the principle of a video editing method in some embodiments of the present application.
  • FIG. 17 is a schematic diagram of the principle of a video editing method in some embodiments of the present application.
  • FIG. 18 is a schematic flowchart of a video editing method according to some embodiments of the present application.
  • FIG. 19 is a schematic diagram of an editing module of some embodiments of the present application.
  • FIG. 20 is a schematic diagram of the connection between a processor and a computer-readable storage medium in some embodiments of the present application.
  • the video editing method of the embodiment of the present application includes the following steps:
  • the to-be-edited video segment is edited to obtain the edited video segment, and the start time and end time of the edited video segment respectively correspond to two rhythm points.
  • the video segment to be edited is determined from one or more video segments according to the interval between two adjacent rhythm points of the preset music segment, and the duration of the video segment to be edited
  • the interval length is greater than or equal to:
  • a video clip whose duration is greater than the interval duration and whose duration difference is the minimum of all duration differences is the video clip to be edited, and the duration difference is the difference between the duration of the video clip and the interval duration.
  • the to-be-edited video segment is edited to obtain the edited video segment.
  • the start and end times of the edited video segment respectively correspond to two rhythm points including:
  • the editing duration is equal to the duration of the part of the video clip to be edited that exceeds the interval duration; and when the second duration is greater than the editing duration, the video clip to be edited is edited from the end time forward to obtain the edited video clip
  • the second duration is the silence duration corresponding to the end moment of the to-be-edited video clip.
  • editing the to-be-edited video segment to obtain the edited video segment also include:
  • editing a video clip to be edited to obtain a clipped video clip also includes:
  • the video editing method further includes: identifying repetitive parts in multiple video clips selected from the same initial video; and recognizing any of the N video clips having the same repetitive part
  • the repeated parts of N-1 video clips are clipped, and N is a positive integer greater than 1.
  • the video editing method further includes: splicing all the edited video segments into one fusion video segment according to the start time and end time of each edited video segment; and integrating the fusion video segment And reserve music clips to output recommended videos.
  • the video editing device 10 of the embodiment of the present application includes an obtaining module 11, a determining module 12 and an editing module 13.
  • the acquiring module 11 is used to mark one or more video clips in the initial video;
  • the determining module 12 is used to determine the to-be-determined video clips from one or more video clips according to the interval between two adjacent rhythm points of the preset music clip. Edit video clips, the duration of the video clips to be edited is greater than or equal to the interval duration;
  • the editing module 13 is used to edit the video clips to be edited to obtain the clipped video clips, and the start and end moments of the clipped video clips respectively correspond to two rhythms point.
  • the determining module 12 includes a determining unit 121.
  • the determining unit 121 is used to determine that among one or more video clips, a video clip whose duration is greater than the interval duration and whose duration difference is the minimum of all duration differences is a video clip to be edited, and the duration difference is the difference between the duration of the video clip and the duration of the interval. Difference.
  • the editing module 13 includes an identifying unit 131, a first editing unit 132 and a second editing unit 133.
  • the identifying unit 131 is used to identify the start time and end time of silence in the video clip to be edited.
  • the first editing unit 132 is used to edit the to-be-edited video segment from the start time backward when the first time length is greater than the editing time length to obtain the edited video segment, and the first time length is the silent duration corresponding to the start time of the to-be-edited video segment ,
  • the editing duration is equal to the duration of the part of the video clip to be edited that exceeds the interval duration.
  • the second editing unit 133 is configured to edit the to-be-edited video segment from the end time forward to obtain the edited video segment when the second time length is greater than the editing time length, and the second time length is the silence duration corresponding to the end time of the to-be-edited video segment.
  • the editing module 13 further includes a judgment unit 134 and a third editing unit 135.
  • the determining unit 134 is configured to determine whether the sum of the first duration and the second duration is greater than the editing duration when the first duration and the second duration are both less than the editing duration.
  • the third editing unit 135 is configured to edit the to-be-edited video segment from the start time and the end time of the to-be-edited video segment, respectively, to obtain the edited video segment when the sum of the first time length and the second time length is greater than the editing time length.
  • the editing module 13 further includes an obtaining unit 136 and a fourth editing unit 137.
  • the acquiring unit 136 is used to acquire the silence duration corresponding to the moment of silence start when the sum of the first duration and the second duration is less than the editing duration;
  • the fourth editing unit 137 is used to acquire the silence duration corresponding to the moment the silence starts is greater than the editing duration, from The video clip to be edited is edited at the moment when the mute starts, and the edited video clip is obtained.
  • the terminal 100 of the embodiment of the present application includes a processor 20, and the processor 20 is configured to mark one or more video clips in the initial video; according to the interval between two adjacent rhythm points of the preset music clip Duration: Determine the video segment to be edited from one or more video segments, and the duration of the video segment to be edited is greater than or equal to the interval duration; and edit the video segment to be edited to obtain the edited video segment, with the start moment of the edited video segment And the end moment correspond to two rhythm points respectively.
  • the processor 20 is configured to determine that among one or more video clips, a video clip whose duration is greater than the interval duration and whose duration difference is the minimum of all duration differences is a video clip to be edited, and the duration difference is a video clip. The difference between the duration of the segment and the duration of the interval.
  • the processor 20 is also used to identify the mute start moment and the mute end moment in the video clip to be edited; when the first duration is greater than the editing duration, edit the video clip to be edited from the start moment backward to obtain
  • the first duration is the mute duration corresponding to the start time of the video clip to be edited
  • the editing duration is equal to the duration of the part of the video clip to be edited that exceeds the interval duration
  • the second duration is greater than the editing duration, it starts from the end
  • the video clip to be edited is clipped forward at all times to obtain the edited video clip, and the second duration is the silence duration corresponding to the end moment of the video clip to be edited.
  • the processor 20 is further configured to determine whether the sum of the first duration and the second duration is greater than the editing duration when the first duration and the second duration are both less than the editing duration; and when the first duration and the second duration are both less than the editing duration; When the sum of the durations is greater than the editing duration, the to-be-edited video segments are respectively edited from the start and end moments of the to-be-edited video segments to obtain the edited video segments.
  • the processor 20 is further configured to, when the sum of the first duration and the second duration is less than the editing duration, obtain the mute duration corresponding to the moment of the mute start, and when the mute duration corresponding to the mute start moment is greater than the editing duration. , Edit the video clip to be edited from the moment of silence, and get the edited video clip.
  • the processor 20 is also used to identify repeated parts in multiple video segments selected from the same initial video, and any N-1 videos among the N video segments that will have the same repeated part.
  • the repetitive part of the clip is clipped.
  • the processor 20 is further configured to splice all the clipped video clips into one fusion video clip according to the start moment and end moment of each clipped video clip, and to integrate the fusion video clip and the predetermined music clip. Output recommended videos.
  • the processor 20 is caused to execute the video editing method of any of the above-mentioned embodiments.
  • the video editing method of the embodiment of the present application includes the following steps:
  • 011 Mark one or more video clips in the initial video
  • 012 Determine the video segment to be edited from one or more video segments according to the interval between two adjacent rhythm points of the preset music segment, and the duration of the video segment to be edited is greater than or equal to the interval duration;
  • 013 Edit the to-be-edited video segment to obtain the edited video segment.
  • the start time and end time of the edited video segment correspond to two rhythm points respectively.
  • the video editing device 10 includes an obtaining module 11, a determining module 12 and an editing module 13.
  • the acquiring module 11 is used to mark one or more video clips in the initial video;
  • the determining module 12 is used to determine the to-be-determined video clips from one or more video clips according to the interval between two adjacent rhythm points of the preset music clip.
  • Edit video clips, the duration of the video clips to be edited is greater than or equal to the interval duration;
  • the editing module 13 is used to edit the video clips to be edited to obtain the clipped video clips, and the start and end moments of the clipped video clips respectively correspond to two rhythms point.
  • step 011, step 012, and step 013 can be implemented by the acquiring module 11, the determining module 12, and the editing module 13, respectively.
  • the terminal 100 further includes a processor 20, which is configured to mark one or more video clips in the initial video; according to the interval length between two adjacent rhythm points of the preset music clip, Determine the to-be-edited video segment from one or more video segments, and the duration of the to-be-edited video segment is greater than or equal to the interval duration; and edit the to-be-edited video segment to obtain the edited video segment, with the start time and end of the edited video segment
  • the moments correspond to two rhythm points respectively. That is to say, step 011 and step 012 and step 013 can be implemented by the processor 20.
  • the terminal 100 includes a housing 30 and a processor 20.
  • the processor 20 is installed in the housing 30.
  • the terminal 100 may be a mobile phone, a tablet computer, a display, a notebook computer, a teller machine, a gate, a smart watch, a head-mounted display device, a game console, and the like.
  • the embodiment of the present application is described by taking the terminal 100 as a mobile phone as an example. It can be understood that the specific form of the terminal 100 is not limited to a mobile phone.
  • the housing 30 can also be used to install functional modules such as an imaging device, a power supply device, and a communication device of the terminal 100, so that the housing 30 provides protections such as dustproof, anti-drop, and waterproof for the functional modules.
  • the processor 20 will automatically classify the initial video taken.
  • the initial video can be an instant video.
  • the video is temporarily sent to the other party in a chat software, and this kind of video does not need to be saved to the system album.
  • the initial video can also be a saved video, such as a video taken by the user using the system camera, or a video that the user manually downloads and saves to the system album from the network.
  • the processor 20 can divide the initial video into multiple categories such as selfie, landscape, pet, cute baby, party, etc. according to the content of the initial video.
  • the initial video of the self-portrait type is defined as the number of frames of the self-portrait in the video occupying the frame of the entire video.
  • the self-portrait image is an image in which the proportion of the human face occupying the current image is greater than the second predetermined ratio (for example, the second predetermined ratio is 1/2);
  • the type of initial video is defined as whether the ratio of landscape images in the video to the number of frames of the entire video is greater than a third predetermined ratio (for example, the third predetermined ratio is greater than 2/3).
  • the landscape image is an image that does not include images of people;
  • the initial type of pet Video is defined as whether the proportion of pet images in the video to the number of frames of the entire video is greater than the fourth predetermined proportion (for example, the fourth predetermined proportion is greater than 2/3), and the proportion of pet images that are pets to the current image is greater than the fifth proportion (for example, The fifth predetermined ratio is greater than 1/3);
  • the initial video of the party type is defined as whether the ratio of the party image in the video to the number of frames of the entire video is greater than the sixth predetermined ratio (for example, the sixth predetermined ratio is greater than 3/4), the party The image is an image containing multiple people. In this way, the type of initial video can be quickly determined through image recognition.
  • the numerical values of the first predetermined ratio to the sixth predetermined ratio are only examples, and should not be regarded as a limitation of the present application.
  • the terminal 100 can mark the highlight segments in the initial video to obtain one or more video segments.
  • the highlight segments are video segments with higher quality in the initial video, for example, in the self-portrait type of initial video.
  • the part of the video with the face image of the person can be regarded as the higher quality part of the initial video; for another example, in the initial video of the party type, the part of the video where multiple people exist at the same time can be regarded as the higher quality part of the initial video.
  • the highlight segment can also be marked by the user manually selecting a more favorite part.
  • the terminal 100 can store the video parameters corresponding to the video clip (including the video path, the video type, and the time information of the video clip in the initial video (playing start time and playback end time), etc.) in the video clip library.
  • the video clip library When obtaining a video clip, first obtain the video parameters of the video clip from the video clip library, and then edit the video clip from the corresponding initial video according to the video parameters, for example, find the corresponding initial video according to the video path of the video clip, Then the video clip is clipped from the initial video according to the time information of the video clip.
  • the video clip library does not need to store all the video files of the video clips that have been marked as highlight clips, but only the video parameters corresponding to the video clips need to be stored, and the required storage space is small.
  • the terminal 100 may obtain one or more video clips.
  • the terminal 100 may obtain one or more video clips from multiple initial videos of the same type.
  • the initial video is edited, it will be based on the video parameters of the selected video clips.
  • Edit one or more video clips from the corresponding videos, and then splice them to generate a new edited video for example, edit multiple initial videos of self-timer types to generate a self-timer type edited video; or, for multiple Editing the initial video of the landscape type to generate a clipped video of the landscape type; or editing multiple initial videos of the cute pet type to generate the edited video of the cute pet type, etc.
  • the edited video includes one or more video clips selected from one or more initial videos. Multiple video clips can be from the same initial video (for example, 3 video clips are selected from the same initial video), and multiple video clips can also be selected. Divided into multiple initial videos (for example, select 1, 2, and 3 video clips from 3 initial videos).
  • the edited video After the edited video is generated, it can only be saved as a temporary file. When the user is satisfied with the edited video, he can manually choose to save it, but when the user is not satisfied with the temporary edited video, or there is no temporary edited video for a certain period of time Perform the operation to delete the clip video. In this way, the user’s preference of editing videos can be analyzed.
  • the scenery-type edited video For example, if the user never saves the scenery-type edited video, it can be determined that the user does not like the scenery-type edited video, and then the scenery-type edited video can no longer be generated in the future.
  • the scenery-type edited video can no longer be generated in the future.
  • all the edited videos of the Selfie type are saved, it can be determined that the user prefers the edited videos of the self-timer type, and more Selfie type edited videos can be generated subsequently.
  • the processor 20 can equip the edited video with background music.
  • the role of the background music in the edited video cannot be ignored.
  • the background music can adjust the emotions of the viewer, and an edited video can be matched with it.
  • the expressiveness and impact of the background music will be greatly improved.
  • the background music is a piece of music that is played synchronously with the clipped video when the clipped video is played.
  • the predetermined music segment is the background music, and the predetermined music segment will be played together with the edited video to serve as the background music in the edited video.
  • a suitable piece of music can be selected as background music from the background music library preset in the terminal 100 or the cloud according to the types of multiple initial videos that need to be edited, or from the user Choose a piece of music as background music from downloaded music and music frequently played by users.
  • the user can manually create a background music library and store his favorite background music in the background music library.
  • the processor 20 may select a suitable piece of music from the background music library as the background music.
  • the original audio clips in the video clip such as the audio clip including the voice of the character, the environmental noise, background music, etc.
  • the background music in the segment and the background music is cut out from the audio segment, so that the video segment only retains the voice, environmental noise, etc. Therefore, when the clip video generated from the video clip and the predetermined music clip are played synchronously, only one predetermined music clip is played as background music, and the viewing effect of the clip video is better.
  • Rhythm points are key time points derived from music fluctuations or its own beat.
  • the rhythm point can be a drum point or a beat point in music. When people enjoy some music, they will beat according to the rhythm of the music, and the beat point can also be a rhythm point.
  • the rhythm point can obtain the rhythm point of the background music through frequency domain analysis according to the frequency spectrum of the background music. Generally, the change of the frequency spectrum of the position corresponding to the rhythm point is more obvious.
  • the rhythm point can be determined in advance, and the processor 20 can also obtain the rhythm point information when acquiring the background music, thereby improving the efficiency of video editing.
  • the processor 20 After the processor 20 obtains the background music and rhythm point information, it will determine the video segment matching the two adjacent rhythm points from the video segment library, thereby determining the video segment to be edited. It can be understood that the closer the duration of the video clip is to the duration between two rhythm points, the less part that needs to be edited. Therefore, the processor 20 may determine from the video clip library a video clip that matches the length of the interval between two adjacent rhythm points as the to-be-edited video clip corresponding to the two rhythm points, for example, the one with the smallest time difference The video clip is determined to be the video clip to be edited, and the duration difference is the difference between the duration of the video clip and the interval duration.
  • the duration of video clip A is 6 seconds (S), and the duration of video clip B is 8S, the duration of video clip C is 10S, background music M includes four rhythm points (respectively S1, S2, S3 and S4), the interval between rhythm point S1 and rhythm point S2 is 5.5S, rhythm point S2 and The interval between rhythm point S3 is 9.5S, and the interval between rhythm point S3 and rhythm point S4 is 7.5S. After a simple calculation, it can be determined that the interval between rhythm point S1 and rhythm point S2 has the smallest difference in duration.
  • the video segment is video segment A (hereinafter referred to as video segment A to be edited), and the video segment with the smallest time difference between the interval length of rhythm point S2 and rhythm point S3 is video segment C (hereinafter referred to as video segment C to be edited), and the rhythm
  • the video clip with the smallest difference in interval between point S3 and rhythm point S4 is video clip B (hereinafter referred to as video clip B to be edited).
  • each video clip only needs to edit the part that exceeds the corresponding interval time. Video content, and the part corresponding to the interval length is reserved.
  • the part of the video clip A to be edited is part a, and the reserved part is the part of the video clip A to be edited except for part a; the video clip B to be edited is edited
  • the dropped part is part b
  • the remaining part is the part of the to-be-edited video segment B except for part b
  • the part to be edited is the part c of the to-be-edited video segment C
  • the remaining part is the part of the to-be-edited video segment C except c
  • the edited parts a, b, and c are all 0.5S, and the video content that is edited is less, and the remaining part is exactly the same as the corresponding interval.
  • the obtained edited video segment A' (that is, the part of the to-be-edited video segment A except for part a) is exactly equal to the length of the interval between the rhythm point S1 and the rhythm point S2; the obtained edited video segment A' The duration of the video segment B'(that is, the part of the to-be-edited video segment B except part b) is exactly equal to the length of the interval between the rhythm point S2 and the rhythm point S3; the obtained edited video segment C'(ie, the part to be edited The duration of the video clip C except for part c) is exactly equal to the interval between the rhythm point S3 and the rhythm point S4, and the start time and end time of the edited video segment are two corresponding rhythm points respectively, so as to achieve The clipped video clip and the background music corresponding to the two rhythm points are played synchronously.
  • the video editing method, video editing device, and terminal 100 of the present application can select matching video clips according to the interval between adjacent rhythm points of the background music, so that each video clip can match the rhythm point, and the video content that is clipped Less, to ensure the integrity of each video clip. And there is no need to manually match the rhythm points of the video clips and the background music, which saves the user's time and energy, and the editing efficiency is also high.
  • step 012 includes the following steps:
  • 0121 Determine that among one or more video clips, the video clip whose duration is greater than the interval duration and the duration difference is the minimum of all duration differences is the video clip to be edited, and the duration difference is the difference between the duration of the video clip and the interval duration.
  • the determining module 12 includes a determining unit 121.
  • the determining unit 121 is used to determine that among one or more video clips, a video clip whose duration is greater than the interval duration and whose duration difference is the minimum of all duration differences is a video clip to be edited, and the duration difference is the difference between the duration of the video clip and the duration of the interval. Difference.
  • step 0121 can be implemented by the determining unit 121.
  • the processor 20 is configured to determine that among one or more video clips, a video clip whose duration is greater than the interval duration and whose duration difference is the minimum of all duration differences is a video clip to be edited, and the duration difference is a video clip. The difference between the duration of the segment and the duration of the interval.
  • step 014 can be implemented by the processor 20.
  • the segment is a video segment to be edited.
  • the time length difference between the interval length of the video segment A and the rhythm point S1 and the rhythm point S2 is 0.5S
  • the video segment The time difference between the interval between B and the rhythm point S1 and the rhythm point S2 is 2.5S
  • the time difference between the interval time between the video clip C and the rhythm point S1 and the rhythm point S2 is 4.5S.
  • the video clip A with the smallest time difference is used as the rhythm
  • the to-be-edited video segments corresponding to point S1 and rhythm point S2 need the least video content to be edited.
  • the duration of the video segment is equal to the interval duration, it can be determined that the video segment is the to-be-edited video segment corresponding to the interval duration.
  • step 013 includes the following steps:
  • the first duration is the silent duration corresponding to the start time of the to-be-edited video segment, and the editing time is equal to the standby time. Edit the duration of the part of the video clip that exceeds the interval duration;
  • the second duration is the silence duration corresponding to the end of the to-be-edited video segment.
  • the editing module 13 includes an identification unit 131, a first editing unit 132 and a second editing unit 133.
  • the identifying unit 131 is used to identify the start time and end time of silence in the video clip to be edited.
  • the first editing unit 132 is used to edit the to-be-edited video segment from the start time backward when the first time length is greater than the editing time length to obtain the edited video segment, and the first time length is the silent duration corresponding to the start time of the to-be-edited video segment ,
  • the editing duration is equal to the duration of the part of the video clip to be edited that exceeds the interval duration.
  • the second editing unit 133 is configured to edit the to-be-edited video segment from the end time forward to obtain the edited video segment when the second time length is greater than the editing time length, and the second time length is the silence duration corresponding to the end time of the to-be-edited video segment.
  • step 0131, step 0132, and step 0133 can be implemented by the identification unit 131, the first editing unit 132, and the second editing unit 13, respectively.
  • the processor 20 is also used to identify the mute start moment and the mute end moment in the video clip to be edited; when the first duration is greater than the editing duration, edit the video clip to be edited from the start moment backward to obtain
  • the first duration is the mute duration corresponding to the start time of the video clip to be edited
  • the editing duration is equal to the duration of the part of the video clip to be edited that exceeds the interval duration
  • the second duration is greater than the editing duration, it starts from the end
  • the video clip to be edited is clipped forward at all times to obtain the edited video clip, and the second duration is the silence duration corresponding to the end moment of the video clip to be edited.
  • step 0131, step 0132, and step 0133 can be implemented by the processor 20.
  • the position of the clip It may happen to be the position where the user hasn't finished a sentence. If you edit directly, it may cause a complete section of the to-be-edited video clip to be cut off, which greatly affects the viewing experience.
  • the processor 20 can first identify the mute start moment and the mute end moment of the video clip to be edited, where the mute start moment refers to the moment when the character in the video clip starts to speak, and the silence end moment refers to the moment when the character in the video clip begins to speak.
  • the processor 20 It can be determined that the moment is the end of silence when there is no user speaking within the preset time period (for example, the preset time length is 1S) after the time when the character finishes speaking for a period of time. In this way, the mute start time and the mute end time can be accurately determined.
  • the video segment E includes 4 mute start moments (respectively P0, P2, P4, and P6) and 3 mute end moments (respectively P1, P3, and P5).
  • the editing duration is 0.5S.
  • the mute duration corresponding to the starting time P0 ie, the first duration P0P1
  • the editing duration is edited backward from the starting time P0 (that is, to the end time P7).
  • the edited parts are all within the silent duration corresponding to the starting time P0, which will not affect the role.
  • the processor 20 can edit the portion of the video clip to be edited backwards at the starting time that exceeds the interval duration, that is, from the beginning You can edit the video content for the duration of the editing backwards at any time.
  • the silence duration corresponding to the end time P7 (ie, the second duration P6P7) is greater than or equal to the editing duration, it means that there is no character speaking between the end time P7 of the video clip and the silence start time P6 nearest to the end time P7.
  • the editing duration is edited forward at the end time P7 (ie, in the direction of the start time P0), and the edited parts are all within the silence duration corresponding to the end time P7, and the integrity of a passage spoken by the character will not be affected.
  • the processor 20 can edit the part of the to-be-edited video clip ahead of the end time that exceeds the interval duration, that is, from the end time forward Just cut out the video content for the length of the clip. In this way, editing is performed under the premise of ensuring the integrity of the character's speech in the video clip, so as to clip the part of the video clip that exceeds the interval time.
  • step 013 further includes the following steps:
  • 0134 When the first duration and the second duration are both less than the editing duration, judge whether the sum of the first duration and the second duration is greater than the editing duration;
  • 0135 When the sum of the first duration and the second duration is greater than the editing duration, edit the to-be-edited video segment from the start and end moments of the to-be-edited video segment, respectively, to obtain the edited video segment.
  • the editing module 13 further includes a judgment unit 134 and a third editing unit 135.
  • the determining unit 134 is configured to determine whether the sum of the first duration and the second duration is greater than the editing duration when the first duration and the second duration are both less than the editing duration.
  • the third editing unit 135 is used for editing the to-be-edited video segment from the start time and the end time of the to-be-edited video segment respectively when the sum of the first time length and the second time length is greater than the editing time length to obtain the edited video segment.
  • step 0134 and step 0135 can be performed by the judgment unit 134 and the third editing unit 135, respectively.
  • the processor 20 is further configured to determine whether the sum of the first duration and the second duration is greater than the editing duration when the first duration and the second duration are both less than the editing duration; and when the first duration and the second duration are both less than the editing duration; When the sum of the duration is greater than the editing duration, the to-be-edited video segment is edited from the start time and the end time of the to-be-edited video segment, respectively, to obtain the edited video segment.
  • step 0134 and step 0135 can be implemented by the processor 20.
  • the processor 20 can determine at this time whether the sum of the first duration P0P1 and the second duration P6P7 is greater than or equal to the editing duration, when the sum of the first duration P0P1 and the second duration P6P7 is greater than or equal to the editing duration (as shown in FIG.
  • the processor 20 can respectively edit the first editing duration that is less than or equal to the first duration P0P1 from the starting time P0 (for example, the first editing duration is 1S) , Edit the second editing duration less than or equal to the second duration P6P7 from the end time P7 forward (for example, the first editing duration is 1S), so that the sum of the first editing duration and the second editing duration is exactly equal to the editing duration. Edit video clips.
  • the processor 20 can start from the start time P0 and the end time P7 when the first duration P0P1 and the second duration P6P7 are both less than the editing duration, and the sum of the first duration P0P1 and the second duration P6P7 is greater than or equal to the editing duration.
  • the part of the edited video clip that exceeds the interval duration can ensure that the remaining part after editing (ie, the edited video clip) corresponds to the interval duration, and the integrity of the character speech in the edited video clip can be ensured.
  • step 013 further includes the following steps:
  • the editing module 13 further includes an obtaining unit 136 and a fourth editing unit 137.
  • the acquiring unit 136 is used to acquire the silence duration corresponding to the moment of silence start when the sum of the first duration and the second duration is less than the editing duration;
  • the fourth editing unit 137 is used to acquire the silence duration corresponding to the moment the silence starts is greater than the editing duration, from The video clip to be edited is edited at the moment when the mute starts, and the edited video clip is obtained.
  • step 0136 and step 0137 can be implemented by the acquisition unit 136 and the fourth editing unit 137, respectively.
  • the processor 20 is further configured to, when the sum of the first duration and the second duration is less than the editing duration, obtain the mute duration corresponding to the moment of the mute start, and when the mute duration corresponding to the mute start moment is greater than the editing duration. , Edit the video clip to be edited from the moment of silence, and get the edited video clip.
  • step 0136 and step 0137 can be implemented by the processor 20.
  • the editing duration at this time is 3S.
  • the first duration P0P1 and the second duration P6P7 are both less than the editing duration, and the sum of the first duration P0P1 and the second duration P6P7 is less than the editing duration, this Editing backwards from the starting time P0, editing forwards from the ending time P7, or editing backwards from the starting time P0 and editing forwards from the ending time P7 will all make the characters in the edited video clip speak The integrity is destroyed.
  • the edited part is within the silent duration, so the edited part will not affect the role in the edited video clip Integrity of speech.
  • the processor 20 splices the remaining two parts of the video clips into one clipped video clip, and the duration of the clipped video clip is equal to the interval duration between the two corresponding rhythm points.
  • the video editing method further includes:
  • the video editing device 10 further includes an identification module 14 and a editing module 15.
  • the recognition module 14 is used for recognizing repeated parts in multiple video clips selected from the same initial video.
  • the editing module 15 is used for editing the repeated part of any N-1 video fragments among the N video fragments with the same repeated part.
  • step 014 and step 015 can be implemented by the recognition module 14 and the editing module 15 respectively.
  • the processor 20 is also used to identify repeated parts in multiple video segments selected from the same initial video, and any N-1 videos among the N video segments that will have the same repeated part.
  • the repetitive part of the clip is clipped.
  • step 014 and step 015 can be implemented by the processor 20.
  • the multiple video clips acquired by the processor 20 may have multiple video clips from the same initial video, and multiple video clips selected from the same initial video may have duplicate parts. If there are multiple clips in the video, For video clips with the same repeated part, the edited video will play the repeated part multiple times during playback. Generally, the audience does not want to see the repeated part for the second time. Therefore, the repeated playback of the repeated part may affect The viewing and impact of the edited video. Therefore, the processor 20 can first identify the repeated parts of the video segments from the same initial video among the multiple video segments. When the processor 20 compares whether multiple video clips have the same repeated part, it may first obtain the video information (such as the video path) of the video clip, and find the initial video corresponding to each video clip according to the video path.
  • the processor 20 may first obtain the video information (such as the video path) of the video clip, and find the initial video corresponding to each video clip according to the video path.
  • the same video path of the video clips means that multiple video clips are selected from the same initial video, and video clips of different initial videos generally do not have overlapping parts. Therefore, in this embodiment, the processor 20 only compares and selects from the same initial video. Multiple video clips of the original video to identify repeated parts. It can be understood that the processor 20 can obtain the time information (such as the playback start time and the playback end time) of each video segment in the initial video when the video segment is selected. After the processor 20 obtains the playback start time and end time of each video segment in the initial video, it can determine whether there is an overlap between the playback start time and the playback end time of the two video segments. There is an overlap between the start time and the playback end time, which means that the two video clips have overlapping parts.
  • the time information such as the playback start time and the playback end time
  • the video segment F is selected from the 1S to 5S part of the initial video O in FIG. 17 (that is, the playback start time is the 1S and the playback end time is the 5S), and the video segment G is selected from the initial video O.
  • the part from 4S to 9S that is, the playback start time is the 4S, and the playback end time is the 9S
  • the video segment H is selected from the 7S to 13S part of the initial video O (that is, the playback start time is the 7S, playback The end time is the 13th).
  • the "1" part of the video segment F and the “2" part of the video segment G intersect, and both are the 4S to 5S video segments in the original video O, which are repeated parts; the "3" part of the video segment G and the video segment H
  • the "4" part of the cross is all the video clips from the 7th to the 9th in the initial video O, and they are also repeated parts. In this way, it is not necessary to perform image recognition to compare the video content of each video segment, and the repeated part can be quickly determined.
  • the processor 20 After determining the repetitive part, the processor 20 will have N-1 (that is, 2) of the N video segments (3 video segments in FIG. 16) of the same repetitive part, such as video segment G and video segment respectively.
  • N-1 that is, 2) of the N video segments (3 video segments in FIG. 16) of the same repetitive part, such as video segment G and video segment respectively.
  • the repetitive part in segment H) is clipped, and only the repetitive part of one of the video segments (ie, video segment F) is retained, so that after all the edited video segments are spliced together according to the rhythm point, the edited video is played ,
  • the repeat part is played only once, and the viewer’s viewing experience is better.
  • the video editing method further includes:
  • the video editing device 10 further includes a splicing module 16 and an integration module 17.
  • the splicing module 16 is used for splicing all the edited video fragments into one fusion video fragment according to the start moment and the end moment of each edited video fragment;
  • the integration module 17 is used for integrating the fusion video fragment and the predetermined music fragment to output the recommended video.
  • step 016 and step 017 can be implemented by the splicing module 16 and the integration module 17 respectively.
  • the processor 20 is further configured to splice all the clipped video clips into one fusion video clip according to the start moment and end moment of each clipped video clip, and to integrate the fusion video clip and the predetermined music clip. Output recommended videos.
  • step 016 and step 017 can be implemented by the processor 20.
  • one or more non-volatile computer-readable storage media 300 containing computer-executable instructions 302 according to the embodiments of the present application, when the computer-executable instructions 302 are executed by one or more processors 20 At this time, the processor 20 is caused to execute the video editing method of any one of the above-mentioned embodiments.
  • the processor 20 when the computer executable instruction 302 is executed by one or more processors 20, the processor 20 is caused to perform the following steps:
  • 011 Mark one or more video clips in the initial video
  • 012 Determine the video segment to be edited from one or more video segments according to the interval between two adjacent rhythm points of the preset music segment, and the duration of the video segment to be edited is greater than or equal to the interval duration;
  • 013 Edit the to-be-edited video segment to obtain the edited video segment.
  • the start time and end time of the edited video segment correspond to two rhythm points respectively.
  • the processor 20 may also perform the following steps:
  • 0138 Determine that among one or more video clips, the video clip whose duration is greater than the interval duration and the duration difference is the minimum of all duration differences is the video clip to be edited, and the duration difference is the difference between the duration of the video clip and the interval duration.

Abstract

一种视频编辑方法、视频编辑装置(10)、终端(100)和非易失性计算机可读存储介质(300)。视频编辑方法包括在初始视频中标记一个或多个视频片段;根据相邻两个节奏点之间的间隔时长,确定视频片段中待剪辑视频片段;及剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。

Description

视频编辑方法、视频编辑装置、终端和可读存储介质
优先权信息
本申请请求2019年12月17日向中国国家知识产权局提交的、专利申请号为201911298854.1的专利申请的优先权和权益,并且通过参照将其全文并入此处。
技术领域
本申请涉及消费性电子技术领域,特别涉及一种视频编辑方法、视频编辑装置、终端和非易失性计算机可读存储介质。
背景技术
目前,在进行视频编辑时,常常是通过用户将多个视频中的某些视频片段剪辑出来,然后将这些视频片段拼接在一起,然后搭配合适的背景音乐,形成一个新的剪辑视频。
发明内容
本申请的实施例提供了一种视频编辑方法、视频编辑装置、终端和非易失性计算机可读存储介质。
本申请实施方式的视频编辑方法包括在初始视频中标记一个或多个视频片段;根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;及剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
本申请实施方式的视频编辑装置包括获取模块、确定模块和编辑模块。所述获取模块用于在初始视频中标记一个或多个视频片段;所述确定模块用于根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;所述编辑模块用于剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
本申请实施方式的终端包括处理器,所述处理器用于在初始视频中标记一个或多个视频片段;根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;及剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
本申请的一种包含计算机可执行指令的非易失性计算机可读存储介质,当所述计算机可执行指令被一个或多个处理器执行时,使得所述处理器执行视频编辑方法。所述视频编辑方法包括在初始视频中标记一个或多个视频片段;根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;及剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
本申请的附加方面和优点将在下面的描述中部分给出,部分将从下面的描述中变得明显,或通过本申请的实践了解到。
附图说明
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。
图1是本申请某些实施方式的视频编辑方法的流程示意图;
图2是本申请某些实施方式的终端结构示意图;
图3是本申请某些实施方式的视频编辑装置的模块示意图;
图4是本申请某些实施方式视频编辑方法的原理示意图;
图5是本申请某些实施方式的视频编辑方法的流程示意图;
图6是本申请某些实施方式的视频编辑装置的模块示意图;
图7是本申请某些实施方式的视频编辑方法的流程示意图;
图8是本申请某些实施方式的编辑模块的模块示意图;
图9是本申请某些实施方式视频编辑方法的原理示意图;
图10是本申请某些实施方式的视频编辑方法的流程示意图;
图11是本申请某些实施方式的编辑模块的模块示意图;
图12是本申请某些实施方式的视频编辑方法的流程示意图;
图13是本申请某些实施方式的编辑模块的模块示意图;
图14是本申请某些实施方式的视频编辑方法的流程示意图;
图15是本申请某些实施方式的编辑模块的模块示意图;
图16是本申请某些实施方式视频编辑方法的原理示意图;
图17是本申请某些实施方式视频编辑方法的原理示意图
图18是本申请某些实施方式的视频编辑方法的流程示意图;
图19是本申请某些实施方式的编辑模块的模块示意图;和
图20是本申请某些实施方式的处理器和计算机可读存储介质的连接示意图。
具体实施方式
以下结合附图对本申请的实施方式作进一步说明。附图中相同或类似的标号自始至终表示相同或类似的元件或具有相同或类似功能的元件。另外,下面结合附图描述的本申请的实施方式是示例性的,仅用于解释本申请的实施方式,而不能理解为对本申请的限制。
请参阅图1,本申请实施方式的视频编辑方法包括以下步骤:
在初始视频中标记一个或多个视频片段;
根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;及
剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。
请参阅图5,在某些实施方式中,根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长包括:
确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。
请参阅图7,在某些实施方式中,剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段 的起始时刻和结束时刻分别对应两个节奏点包括:
识别待剪辑视频片段中的静音开始时刻和静音结束时刻;及在第一时长大于剪辑时长时,从起始时刻向后剪辑待剪辑视频片段,得到已剪辑视频片段,第一时长为待剪辑视频片段的起始时刻对应的静音时长,剪辑时长等于待剪辑视频片段中超过间隔时长的部分的时长;及在第二时长大于剪辑时长时,从结束时刻向前剪辑待剪辑视频片段,得到已剪辑视频片段,第二时长为待剪辑视频片段的结束时刻对应的静音时长。
请参阅图10,在某些实施方式中,剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点还包括:
在第一时长和第二时长均小于剪辑时长时,判断第一时长和第二时长之和是否大于剪辑时长;及在第一时长和第二时长之和大于剪辑时长时,分别从待剪辑视频片段的起始时刻和结束时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
请参阅图12,在某些实施方式中,剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点还包括:
在第一时长和第二时长之和小于剪辑时长时,获取静音开始时刻对应的静音时长;在静音开始时刻对应的静音时长大于剪辑时长时,从静音开始时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
请参阅图14,在某些实施方式中,视频编辑方法还包括:识别选取自同一初始视频的多个视频片段中的重复部分;及将具有相同的重复部分的N个视频片段中的任意N-1个视频片段中的重复部分剪辑掉,N为大于1的正整数。
请参阅图18,在某些实施方式中,视频编辑方法还包括:按照每个已剪辑视频片段的起始时刻和结束时刻将所有已剪辑视频片段拼接为一个融合视频片段;及整合融合视频片段和预定音乐片段以输出推荐视频。
请参阅图2,本申请实施方式的视频编辑装置10包括获取模块11、确定模块12和编辑模块13。获取模块11用于在初始视频中标记一个或多个视频片段;确定模块12用于根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;编辑模块13用于剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。
请参阅图6,在某些实施方式中,确定模块12包括确定单元121。确定单元121用于确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。
请参阅图8,在某些实施方式中,编辑模块13包括识别单元131、第一剪辑单元132和第二剪辑单元133。识别单元131用于识别待剪辑视频片段中的静音开始时刻和静音结束时刻。第一剪辑单元132用于在第一时长大于剪辑时长时,从起始时刻向后剪辑待剪辑视频片段,得到已剪辑视频片段,第一时长为待剪辑视频片段的起始时刻对应的静音时长,剪辑时长等于待剪辑视频片段中超过间隔时长的部分的时长。第二剪辑单元133用于在第二时长大于剪辑时长时,从结束时刻向前剪辑待剪辑视频片段,得到已剪辑视频片段,第二时长为待剪辑视频片段的结束时刻对应的静音时长。
请参阅图11,在某些实施方式中,编辑模块13还包括判断单元134和第三剪辑单元135。判断单元134用于在第一时长和第二时长均小于剪辑时长时,判断第一时长和第二时长之和是否大于剪辑时长。第三剪辑单元135用于在第一时长和第二时长之和大于剪辑时长时,分别从待剪辑视频片段的起始时刻和结束时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
请参阅图13,在某些实施方式中,编辑模块13还包括获取单元136和第四剪辑单元137。获取单元136用于在第一时长和第二时长之和小于剪辑时长时,获取静音开始时刻对应的静音时长;第四剪辑单元137用于在静音开始时刻对应的静音时长大于剪辑时长时,从静音开始时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
请参阅图3,本申请实施方式的终端100包括处理器20,处理器20用于在初始视频中标记一个或多个视频片段;根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;及剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。
在某些实施方式中,处理器20用于确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。
在某些实施方式中,处理器20还用于识别待剪辑视频片段中的静音开始时刻和静音结束时刻;在第一时长大于剪辑时长时,从起始时刻向后剪辑待剪辑视频片段,得到已剪辑视频片段,第一时长为待剪辑视频片段的起始时刻对应的静音时长,剪辑时长等于待剪辑视频片段中超过间隔时长的部分的时长;及在第二时长大于剪辑时长时,从结束时刻向前剪辑待剪辑视频片段,得到已剪辑视频片段,第二时长为待剪辑视频片段的结束时刻对应的静音时长。
在某些实施方式中,处理器20还用于在第一时长和第二时长均小于剪辑时长时,判断第一时长和第二时长之和是否大于剪辑时长;及在第一时长和第二时长之和大于剪辑时长时,分别从待剪辑视频片段的起始时刻和结束时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
在某些实施方式中,处理器20还用于在第一时长和第二时长之和小于剪辑时长时,获取静音开始时刻对应的静音时长、及在静音开始时刻对应的静音时长大于剪辑时长时,从静音开始时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
在某些实施方式中,处理器20还用于识别选取自同一初始视频的多个视频片段中的重复部分、及将具有相同的重复部分的N个视频片段中的任意N-1个视频片段中的重复部分剪辑掉。
在某些实施方式中,处理器20还用于按照每个已剪辑视频片段的起始时刻和结束时刻将所有已剪辑视频片段拼接为一个融合视频片段、及整合融合视频片段和预定音乐片段以输出推荐视频。
请参阅图2和图18,本申请实施方式的一种包含计算机可执行指令302的非易失性计算机可读存储介质300,当计算机可执行指令302被一个或多个处理器20执行时,使得处理器20可执行上述任一实施方式的视频编辑方法。
请参阅图1至图3,本申请实施方式的视频编辑方法包括以下步骤:
011:在初始视频中标记一个或多个视频片段;
012:根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;及
013:剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。
在某些实施方式中,视频编辑装置10包括获取模块11、确定模块12和编辑模块13。获取模块11用于在初始视频中标记一个或多个视频片段;确定模块12用于根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;编辑模块13用于剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。也即是说,步骤011、步骤012和步骤013可分别由获取模块11、 确定模块12和编辑模块13实现。
在某些实施方式中,终端100还包括处理器20,处理器20用于在初始视频中标记一个或多个视频片段;根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;及剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。也即是说,步骤011和、步骤012和步骤013可以由处理器20实现。
具体地,终端100包括壳体30和处理器20。处理器20安装在壳体30内。更具体地,终端100可以是手机、平板电脑、显示器、笔记本电脑、柜员机、闸机、智能手表、头显设备、游戏机等。本申请实施方式以终端100是手机为例进行说明,可以理解,终端100的具体形式并不限于手机。壳体30还可用于安装终端100的成像装置、供电装置、通信装置等功能模块,以使壳体30为功能模块提供防尘、防摔、防水等保护。
在用户日常使用手机拍摄视频时,处理器20会自动对拍摄的初始视频进行分类,其中,初始视频可以是即时视频,例如在聊天软件中临时视频发送给对方,该种视频无需保存到系统相册;或者,初始视频也可以是已保存的视频,例如被用户使用系统相机拍摄得到的视频,或者用户手动从网络下载保存到系统相册内的视频。
处理器20根据初始视频的内容可将初始视频分为自拍、风景、宠物、萌娃、聚会等多种类别,其中,自拍类型的初始视频定义为视频中自拍图像的帧数占整个视频的帧数的比例是否大于第一预定比例(例如第一预定比例大于1/2),自拍图像为人脸占据当前图像的比例大于第二预定比例(如第二预定比例为1/2)的图像;风景类型的初始视频定义为视频中风景图像占整个视频的帧数的比例是否大于第三预定比例(例如第三预定比例大于2/3),风景图像为不包括人物图像的图像;宠物类型的初始视频定义为视频中宠物图像占整个视频的帧数的比例是否大于第四预定比例(例如第四预定比例大于2/3),宠物图像为宠物的图像占当前图像的比例大于第五比例(例如第五预定比例大于1/3)的图像;聚会类型的初始视频定义为视频中聚会图像占整个视频的帧数的比例是否大于第六预定比例(例如第六预定比例大于3/4),聚会图像为包含多个人物的图像。如此,可通过图像识别快速确定初始视频的类型。需要指出的是,上述第一预定比例至第六预定比例的数值仅为举例说明,并不能作为对本申请的限制。
在对初始视频确定好类型后,终端100可对初始视频中的精彩片段进行标记以得到一个或多个视频片段,精彩片段为初始视频中质量较高的视频片段,例如在自拍类型的初始视频中,存在人物脸部图像的视频部分可作为该初始视频中质量较高的部分;再例如,在聚会类型的初始视频中,多个人物同时存在的视频部分可作为该初始视频中质量较高的部分;再例如,在萌宠类型的初始视频中,宠物存在的视频部分可作为该初始视频中质量较高的部分;在此不再一一列举。当然,精彩片段也可以是用户手动选取较为喜欢的部分而被标记的。
终端100可将视频片段对应的视频参数(包括视频路径、视频类型、该视频片段的在初始视频中的时间信息(播放起始时刻和播放终止时刻)等)存入视频片段库中,在需要获取视频片段时,首先从视频片段库中获取视频片段的视频参数,然后根据视频参数从对应的初始视频中将该视频片段剪辑出来,例如,根据该视频片段的视频路径找到对应的初始视频,然后根据该视频片段的时间信息从初始视频中将该视频片段剪辑出来。如此,视频片段库中不需要存储所有已被标记为精彩片段的视频片段的视频文件,而是只需要存储该视频片段对应的视频参数即可,所需的存储空间较小。
终端100可获取一个或多个视频片段,例如终端100会从同类型的多个初始视频中获取一个或多 个视频片段,在编辑初始视频时,会根据已经选取好的视频片段的视频参数,将一个或多个视频片段从对应的视频中剪辑出来,然后拼接以生成一个新的剪辑视频,例如对多个自拍类型的初始视频进行编辑以生成一个自拍类型的剪辑视频;或者,对多个风景类型的初始视频进行编辑以生成风景类型的剪辑视频;或者,对多个萌宠类型的初始视频进行编辑以生成萌宠类型的剪辑视频等。
剪辑视频包括从一个或多个初始视频中选取的一个或多个视频片段,多个视频片段可以来自同一初始视频(例如,在同一初始视频中选取3个视频片段),多个视频片段也可以分为来自多个初始视频(例如,在3个初始视频中分别选取1个、2个和3个视频片段)。在生成剪辑视频后,可仅作为临时文件进行保存,在用户对该剪辑视频满意时,可手动选择保存,而在用户对临时的剪辑视频不满意,或者在一定时间内没有对临时的剪辑视频进行操作,则删除该剪辑视频。如此,可对用户的喜好剪辑视频进行分析,如用户从不保存风景类型的剪辑视频,则可确定用户不喜欢风景类型的剪辑视频,后续可不再生成风景类型的剪辑视频,而在用户对生成的自拍类型的剪辑视频均进行保存时,则可确定用户较为喜欢自拍类型的剪辑视频,可在后续生成更多的自拍类型的剪辑视频。
为了增加剪辑视频的表现力和冲击力,处理器20可为剪辑视频配上背景音乐,背景音乐的作用在剪辑视频中的不可忽视,背景音乐可调动观看者的情绪,一个剪辑视频配上合适的背景音乐,其表现力和冲击力会得到很大的提升,其中,背景音乐为播放剪辑视频时,与剪辑视频同步播放的一段音乐。本申请实施方式中,预定音乐片段即为背景音乐,预定音乐片段会与剪辑视频一起播放,以作为剪辑视频中的背景音乐。在获取预定音乐片段(即,背景音乐时)时,可根据需要剪辑的多个初始视频的类型,从终端100或云端预设的背景音乐库中选取一段合适的音乐作为背景音乐,或者从用户下载的音乐、用户经常播放的音乐中选取一段音乐作为背景音乐。或者,用户可手动创建背景音乐库,将自己喜欢的背景音乐存在背景音乐库中。处理器20在获取背景音乐时,可从该背景音乐库中选取一段合适的音乐作为背景音乐。
视频片段中原本存在的音频片段,如音频片段包括人物的说话声、环境噪音、背景音乐等,为了防止视频片段的背景音乐和预定音乐片段同时播放,影响剪辑视频的观看效果,可通过识别视频片段中的背景音乐并将该背景音乐从音频片段中剪出来,使得视频片段仅保留说话声、环境噪音等。从而使得根据视频片段生成的剪辑视频和预定音乐片段同步播放时,仅有一个预定音乐片段作为背景音乐来播放,剪辑视频的观看效果较好。
音乐一般存在节奏点,节奏点是根据音乐起伏或者自身节拍而得出的关键时间点。例如,节奏点可以是鼓点,也可以是音乐中的节拍点,人们欣赏一些音乐时,会随着音乐的节奏打拍子,打拍子的点也可以是节奏点。再例如,节奏点可根据背景音乐的频谱,通过频域分析的方式获取背景音乐的节奏点,一般节奏点对应的位置的频谱的变化较为明显。对于背景音乐库中的音乐,可事先就确定好节奏点,处理器20在获取背景音乐时,即可一并获取节奏点信息,从而提高视频剪辑的效率。
在处理器20获取到背景音乐及节奏点信息后,会从视频片段库中确定与相邻的两个节奏点匹配的视频片段,从而确定待剪辑视频片段。可以理解,视频片段的时长越接近两个节奏点之间的时长,所需剪辑的部分就越少。因此,处理器20可从视频片段库中确定与相邻两个节奏点之间的间隔时长相匹配的视频片段作为与该两个节奏点对应的待剪辑视频片段,例如,将时长差最小的视频片段确定为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值,此时在对待剪辑视频片段进行剪辑时,只需剪辑出和间隔时长对应的部分(即,只需将超过间隔时长的部分剪辑掉),由于待剪辑视频片段的时长和间隔时长的时长差最小,相对其他视频片段作为待剪辑视频片段进行剪辑而言,被剪辑掉的视频内容较少,可尽可能的保证剪辑后得到的已剪辑视频片段(即,待剪辑视频片段剪辑掉超过间隔时 长的部分后,剩余的视频片段部分)的完整性。在图4所示的例子中,待剪辑的视频片段为3个,分别为视频片段A、视频片段B和视频片段C,视频片段A的时长为6秒(S)、视频片段B的时长为8S,视频片段C的时长为10S,背景音乐M包括四个节奏点(分别为S1、S2、S3和S4),节奏点S1和节奏点S2之间的间隔时长为5.5S,节奏点S2和节奏点S3之间的间隔时长为9.5S,节奏点S3和节奏点S4之间的间隔时长为7.5S,经过简单计算即可确定与节奏点S1和节奏点S2的间隔时长的时长差最小的视频片段为视频片段A(下称待剪辑视频片段A),与节奏点S2和节奏点S3的间隔时长的时长差最小的视频片段为视频片段C(下称待剪辑视频片段C),与节奏点S3和节奏点S4的间隔时长的时长差最小的视频片段为视频片段B(下称待剪辑视频片段B),在剪辑时,每个视频片段均只需剪辑超过对应的间隔时长的部分的视频内容,而保留和间隔时长对应的部分,如待剪辑视频片段A剪辑掉的部分为a部分,保留的部分为待剪辑视频片段A中除a部分之外的部分;待剪辑视频片段B剪辑掉的部分为b部分,保留的部分为待剪辑视频片段B中除b部分之外的部分,待剪辑视频片段C剪辑掉的部分为c部分,保留的部分为待剪辑视频片段C中除c部分之外的部分,被剪辑掉的a部分、b部分和c部分均为0.5S,被剪辑掉的视频内容较少,保留的部分则刚好和对应的间隔时长相等。
在完成剪辑后,得到的已剪辑视频片段A’(即,待剪辑视频片段A中除a部分之外的部分)的时长刚好等于节奏点S1和节奏点S2之间间隔时长;得到的已剪辑视频片段B’(即,待剪辑视频片段B中除b部分之外的部分)的时长刚好等于节奏点S2和节奏点S3之间间隔时长;得到的已剪辑视频片段C’(即,待剪辑视频片段C中除c部分之外的部分)的时长刚好等于节奏点S3和节奏点S4之间间隔时长,已剪辑视频片段的起始时刻和结束时刻分别为对应的两个节奏点,从而实现已剪辑视频片段与该两个节奏点对应的背景音乐的同步播放。
用户在人工进行视频剪辑过程中,存在需要耗费用户很多的时间和精力,剪辑效率也较低的问题。
本申请的视频编辑方法、视频编辑装置和终端100可根据背景音乐的相邻节奏点之间的间隔时长选取匹配的视频片段,使得每个视频片段能够和节奏点配合,被剪辑掉的视频内容较少,保证了每个视频片段的完整性。且无需人为进行视频片段和背景音乐的节奏点的匹配,节省了用户的时间和精力,剪辑效率也较高。
请参阅图2、图5和图6,在某些实施方式中,步骤012包括以下步骤:
0121:确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。
在某些实施方式中,确定模块12包括确定单元121。确定单元121用于确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。也即是说,步骤0121可以由确定单元121实现。
在某些实施方式中,处理器20用于确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。也即是说,步骤014可以由处理器20实现。
具体地,请参阅图4,在从视频片段库中确定与节奏点S1和节奏点S2的间隔时长匹配的视频片段时,当视频片段的时长小于节奏点S1和节奏点S2之间的间隔时长时(如视频片段D的时长为6S,小于7.5S时),若将视频片段D与节奏点S1和节奏点S2去匹配,虽然不需要对视频片段D进行剪辑,但会出现剪辑视频在播放时,存在背景音乐在播放而没有对应的视频内容显示的情况。因此,在对相邻两个节奏点匹配视频片段时,时长小于间隔时长的视频片段均不能作为待剪辑视频片段。
在视频片段的时长大于间隔时长时,为了使得视频片段被剪辑的内容最少,在时长差为所有视频 片段和间隔时长的时长差中的最小值时,即可确定该最小的时长差对应的视频片段为待剪辑视频片段,例如,在对图4中的节奏点S1和节奏点S2待剪辑视频片段时,视频片段A和节奏点S1和节奏点S2间隔时长的时长差为0.5S,视频片段B和节奏点S1和节奏点S2间隔时长的时长差为2.5S,视频片段C和节奏点S1和节奏点S2间隔时长的时长差为4.5S,此时以时长差最小的视频片段A作为节奏点S1和节奏点S2对应的待剪辑视频片段,所需剪辑的视频内容最少。当然,在视频片段的时长等于间隔时长,即可确定该视频片段为该间隔时长对应的待剪辑视频片段,此时无需剪辑,视频片段的完整性较好,不会出现背景音乐在播放而没有对应的视频内容显示的情况。
请参阅图2、图7和图8,在某些实施方式中,步骤013包括以下步骤:
0131:识别待剪辑视频片段中的静音开始时刻和静音结束时刻;及
0132:在第一时长大于剪辑时长时,从起始时刻向后剪辑待剪辑视频片段,得到已剪辑视频片段,第一时长为待剪辑视频片段的起始时刻对应的静音时长,剪辑时长等于待剪辑视频片段中超过间隔时长的部分的时长;及
0133:在第二时长大于剪辑时长时,从结束时刻向前剪辑待剪辑视频片段,得到已剪辑视频片段,第二时长为待剪辑视频片段的结束时刻对应的静音时长。
在某些实施方式中,编辑模块13包括识别单元131、第一剪辑单元132和第二剪辑单元133。识别单元131用于识别待剪辑视频片段中的静音开始时刻和静音结束时刻。第一剪辑单元132用于在第一时长大于剪辑时长时,从起始时刻向后剪辑待剪辑视频片段,得到已剪辑视频片段,第一时长为待剪辑视频片段的起始时刻对应的静音时长,剪辑时长等于待剪辑视频片段中超过间隔时长的部分的时长。第二剪辑单元133用于在第二时长大于剪辑时长时,从结束时刻向前剪辑待剪辑视频片段,得到已剪辑视频片段,第二时长为待剪辑视频片段的结束时刻对应的静音时长。也即是说,步骤0131、步骤0132和步骤0133可分别由识别单元131、第一剪辑单元132和第二剪辑单元13实现。
在某些实施方式中,处理器20还用于识别待剪辑视频片段中的静音开始时刻和静音结束时刻;在第一时长大于剪辑时长时,从起始时刻向后剪辑待剪辑视频片段,得到已剪辑视频片段,第一时长为待剪辑视频片段的起始时刻对应的静音时长,剪辑时长等于待剪辑视频片段中超过间隔时长的部分的时长;及在第二时长大于剪辑时长时,从结束时刻向前剪辑待剪辑视频片段,得到已剪辑视频片段,第二时长为待剪辑视频片段的结束时刻对应的静音时长。也即是说,步骤0131、步骤0132和步骤0133可以由处理器20实现。
具体地,可以理解,在对待剪辑视频片段进行剪辑时,将待剪辑视频片段中和间隔时长对应的部分剪辑出来(即,剪辑掉待剪辑视频片段中超过间隔时长的部分)时,剪辑的位置可能正好是用户一句话还没说完的位置,若直接剪辑的话可能导致待剪辑视频片段中一段完整的话被剪辑掉了一部分,十分影响观看体验。因此,处理器20在剪辑前,可首先识别待剪辑视频片段的静音开始时刻和静音结束时刻,其中,静音开始时刻指的是视频片段中角色开始讲话的时刻,而静音结束时刻则是指该角色讲完一段话的时刻,当然,为了防止将一段完整的语句之间的停顿位置(如犹豫、口吃、呼吸等)也作为静音结束时刻,而导致完整的语句被剪辑掉一部分,处理器20可在该角色讲完一段话的时刻后预设时长(例如预设时长为1S)内均没有用户讲话时,才确定该时刻为静音结束时刻。如此,可准确地确定静音开始时刻和静音结束时刻。
请参阅图9,在一个例子中,视频片段E包括4个静音开始时刻(分别为P0、P2、P4和P6)和3个静音结束时刻(分别为P1、P3和P5),视频片段E的剪辑时长为0.5S,在起始时刻P0对应的静音时长(即,第一时长P0P1)大于或等于剪辑时长时,表示视频片段的从起始时刻P0一直到距离起始 时刻P0最近的静音结束时刻P1均没有角色讲话,此时以起始时刻P0向后(即,向结束时刻P7方向)剪辑掉剪辑时长,被剪辑的部分均位于起始时刻P0对应的静音时长内,不会影响角色讲的一段话的完整性。若角色在起始时刻P0即开始讲话,此时起始时刻P0将被识别为静音结束时刻(即,起始时刻P0没有对应的静音时长),此时无法从起始时刻P0进行剪辑。因此,在起始时刻P0存在对应静音时长,且第一时长P0P1大于或等于剪辑时长时,处理器20可以起始时刻向后剪辑待剪辑视频片段中,超过间隔时长的部分,即从起始时刻向后剪辑掉剪辑时长的视频内容即可。
在结束时刻P7对应的静音时长(即,第二时长P6P7)大于或等于剪辑时长时,表示视频片段的结束时刻P7到距离结束时刻P7最近的静音开始时刻P6之间均没有角色讲话,此时以结束时刻P7向前(即,向起始时刻P0方向)剪辑掉剪辑时长,被剪辑的部分均位于结束时刻P7对应的静音时长内,也不会影响角色讲的一段话的完整性。若在静音开始时刻P6和结束时刻P7之间存在静音结束时刻,且静音结束时刻和结束时刻P7之间不存在静音开始时刻,则表示该静音结束时刻至结束时刻P7之间,角色一直在讲话(即,结束时刻P7没有对应的静音时长),此时无法从结束时刻P7进行剪辑。因此,在结束时刻P7存在对应静音时长,且第二时长P6P7大于或等于剪辑时长时,处理器20可以结束时刻向前剪辑待剪辑视频片段中,超过间隔时长的部分,即从结束时刻向前剪辑掉剪辑时长的视频内容即可。如此,在保证视频片段中角色讲话的完整性的前提下进行剪辑,以剪辑视频片段中超过间隔时长的部分。
请参阅图2、图10和图11,在某些实施方式中,步骤013还包括以下步骤:
0134:在第一时长和第二时长均小于剪辑时长时,判断第一时长和第二时长之和是否大于剪辑时长;及
0135:在第一时长和第二时长之和大于剪辑时长时,分别从待剪辑视频片段的起始时刻和结束时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
在某些实施方式中,编辑模块13还包括判断单元134和第三剪辑单元135。判断单元134用于在第一时长和第二时长均小于剪辑时长时,判断第一时长和第二时长之和是否大于剪辑时长。第三剪辑单元135用于在第一时长和第二时长之和大于剪辑时长时,分别从待剪辑视频片段的起始时刻和结束时刻剪辑待剪辑视频片段,得到已剪辑视频片段。也即是说,步骤0134和步骤0135可分别由判断单元134和第三剪辑单元135。
在某些实施方式中,处理器20还用于在第一时长和第二时长均小于剪辑时长时,判断第一时长和第二时长之和是否大于剪辑时长;及在第一时长和第二时长之和大于剪辑时长时,分别从待剪辑视频片段的起始时刻和结束时刻剪辑待剪辑视频片段,得到已剪辑视频片段。也即是说,步骤0134和步骤0135可以由处理器20实现。
具体地,请结合图9,在第一时长P0P1和第二时长P6P7均小于剪辑时长时(例如剪辑时长为2S,第一时长P0P1=1.5S,第二时长P6P7=1S),此时不论是从起始时刻向后剪辑剪辑时长的视频内容,还是从结束时刻向前剪辑剪辑时长的视频内容,均会导致角色讲的一段话的完整性因剪辑而破坏。因此,处理器20可在此时判断第一时长P0P1和第二时长P6P7之和是否大于或等于剪辑时长,在第一时长P0P1和第二时长P6P7之和大于或等于剪辑时长时(如图9所示的例子中,P0P1+P6P7=2.5S>2S),处理器20可分别从起始时刻P0向后剪辑小于或等于第一时长P0P1的第一剪辑时长(如第一剪辑时长为1S),从结束时刻P7向前剪辑小于或等于第二时长P6P7的第二剪辑时长(如第一剪辑时长为1S),使得第一剪辑时长和第二剪辑时长之和刚好等于剪辑时长,从而得到已剪辑视频片段。
如此,处理器20可在第一时长P0P1和第二时长P6P7均小于剪辑时长、且第一时长P0P1和第二 时长P6P7之和大于或等于剪辑时长时,分别从起始时刻P0和结束时刻P7剪辑视频片段中超过间隔时长的部分,可保证剪辑后保留下来的部分(即,已剪辑视频片段)与间隔时长对应,可保证已剪辑视频片段中的角色讲话的完整性。
请参阅图2、图12和图13,在某些实施方式中,步骤013还包括以下步骤:
0136:在第一时长和第二时长之和小于剪辑时长时,获取静音开始时刻对应的静音时长;
0137:在静音开始时刻对应的静音时长大于剪辑时长时,从静音开始时刻剪辑待剪辑视频片段,得到已剪辑视频片段。
在某些实施方式中,编辑模块13还包括获取单元136和第四剪辑单元137。获取单元136用于在第一时长和第二时长之和小于剪辑时长时,获取静音开始时刻对应的静音时长;第四剪辑单元137用于在静音开始时刻对应的静音时长大于剪辑时长时,从静音开始时刻剪辑待剪辑视频片段,得到已剪辑视频片段。也即是说,步骤0136和步骤0137可分别由获取单元136和第四剪辑单元137实现。
在某些实施方式中,处理器20还用于在第一时长和第二时长之和小于剪辑时长时,获取静音开始时刻对应的静音时长、及在静音开始时刻对应的静音时长大于剪辑时长时,从静音开始时刻剪辑待剪辑视频片段,得到已剪辑视频片段。也即是说,步骤0136和步骤0137可以由处理器20实现。
具体地,请结合图9,例如此时的剪辑时长为3S,当第一时长P0P1和第二时长P6P7均小于剪辑时长、且第一时长P0P1和第二时长P6P7之和小于剪辑时长时,此时从起始时刻P0向后剪辑、从结束时刻P7向前剪辑、或分别时从起始时刻P0向后剪辑和从结束时刻P7向前剪辑均会使得剪辑后的视频片段中的角色说话的完整性破坏,因此,处理器20首先可先获取每个静音开始时刻(分别为P0、P2、P4和P6)对应的静音时长(即,静音开始时刻至对应的静音结束时刻之间的时长,如静音时长P0P1=1.5S,静音时长P2P3=3.5S,静音时长P4P5=1.4S,静音时长P6P7=1S),然后在选取任一大于剪辑时长的静音时长(如静音时长P2P3),从静音时长P2P3的静音开始时刻P2开始剪辑,剪辑掉剪辑时长(如剪辑时长为3S)的视频内容,此时被剪辑的部分均处于静音时长内,所以剪辑的部分不会影响剪辑后的视频片段中角色讲话的完整性。最后处理器20将剪辑后剩下的两部分的视频片段拼接成一个已剪辑视频片段即可,该已剪辑视频片段的时长等于对应的两个节奏点之间的间隔时长。
请参阅图2、图14和图15,在某些实施方式中,视频编辑方法还包括:
014:识别选取自同一初始视频的多个视频片段中的重复部分;及
015:将具有相同的重复部分的N个视频片段中的任意N-1个视频片段中的重复部分剪辑掉,N为大于1的正整数。
在某些实施方式中,视频编辑装置10还包括识别模块14和剪辑模块15。识别模块14用于识别选取自同一初始视频的多个视频片段中的重复部分。剪辑模块15用于将具有相同的重复部分的N个视频片段中的任意N-1个视频片段中的重复部分剪辑掉。也即是说,步骤014和步骤015可分别由识别模块14和剪辑模块15实现。
在某些实施方式中,处理器20还用于识别选取自同一初始视频的多个视频片段中的重复部分、及将具有相同的重复部分的N个视频片段中的任意N-1个视频片段中的重复部分剪辑掉。也即是说,步骤014和步骤015可以由处理器20实现。
具体地,在剪辑前,处理器20获取的多个视频片段可能有来自同一初始视频的多个视频片段,选取自同一初始视频的多个视频片段可能存在重复部分,若剪辑视频存在多个具有相同重复部分的视频片段,则剪辑视频在播放时会多次播放该重复部分,一般的,对于重复的部分,观众并不想第二次看到,因此,重复部分的多次播放可能会影响剪辑视频的观赏性和冲击力。因此,处理器20可以识别首 先识别多个视频片段中来自同一初始视频的视频片段的重复部分。处理器20在比对多个视频片段是否存在相同的重复部分时,可首先获取该视频片段的视频信息(如视频路径),根据视频的路径可找到每个视频片段对应的初始视频,多个视频片段的视频路径相同即表示多个视频片段选取自同一个初始视频,而不同初始视频的视频片段一般不存在重复部分,因此,本实施方式中,处理器20仅比对选取自同一初始视频的多个视频片段以识别重复部分。可以理解,在选取视频片段时处理器20就可获取到每个视频片段在初始视频中的时间信息(如播放起始时刻和播放终止时刻)。处理器20在获取到每个视频片段在初始视频中的播放起始时刻和终止时刻后,可判断两个视频片段的播放起始时刻和播放终止时刻是否存在交叉,若两个视频片段的播放起始时刻和播放终止时刻存在交叉,则表示该两个视频片段存在重复部分。例如图16中,视频片段F选自图17的初始视频O第1S到第5S的部分(即播放起始时刻为第1S,播放终止时刻为第5S),视频片段G选自初始视频O第4S到第9S的部分(即播放起始时刻为第4S,播放终止时刻为第9S),视频片段H选自初始视频O第7S到第13S的部分(即播放起始时刻为第7S,播放终止时刻为第13S)。视频片段F的“1”部分和视频片段G的“2”部分交叉,均为初始视频O中第4S到第5S的视频片段,为重复部分;视频片段G的“3”部分和视频片段H的“4”部分交叉,均为初始视频O中第7S到第9S的视频片段,也为重复部分。如此,无需进行图像识别来比对每个视频片段的视频内容,即可快速确定重复部分。
在确定重复部分后,处理器20将存在同样的重复部分的N个视频片段(如图16中的3个视频片段)中的N-1个(即2个,如分别为视频片段G和视频片段H)中的重复部分剪辑掉,而仅仅保留其中一个视频片段(即,视频片段F)的重复部分,这样在后续将所有剪辑后的视频片段根据节奏点拼接起来后,剪辑视频在播放时,重复部分仅播放一次,观众的观赏体验较好。
请参阅图2、图18和图19,在某些实施方式中,视频编辑方法还包括:
016:按照每个已剪辑视频片段的起始时刻和结束时刻将所有已剪辑视频片段拼接为一个融合视频片段;及
017:整合融合视频片段和预定音乐片段以输出推荐视频。
在某些实施方式中,视频编辑装置10还包括拼接模块16和整合模块17。拼接模块16用于按照每个已剪辑视频片段的起始时刻和结束时刻将所有已剪辑视频片段拼接为一个融合视频片段;整合模块17用于整合融合视频片段和预定音乐片段以输出推荐视频。也即是说,步骤016和步骤017可分别由拼接模块16和整合模块17实现。
在某些实施方式中,处理器20还用于按照每个已剪辑视频片段的起始时刻和结束时刻将所有已剪辑视频片段拼接为一个融合视频片段、及整合融合视频片段和预定音乐片段以输出推荐视频。也即是说,步骤016和步骤017可以由处理器20实现。
具体地,如图4所示,已剪辑视频片段A’的起始时刻和结束时刻分别对应节奏点S1和节奏点S2,已剪辑视频片段B’起始时刻和结束时刻分别对应节奏点S2和节奏点S3,已剪辑视频片段C’起始时刻和结束时刻分别对应节奏点S3和节奏点S4,处理器20可按照已剪辑视频片段的起始时刻和结束时刻的顺序(即,对应的节奏点的顺序),将剪辑后的已剪辑视频片段A’、已剪辑视频片段B’和已剪辑视频片段C’拼接形成一个新的融合视频片段V1,然后将融合视频片段V1和背景音乐M整合成一个推荐视频V2,在播放推荐视频V2时,已剪辑视频片段A’和背景音乐M中处于节奏点S1和S2之间的音乐同时开始播放且同时结束播放,已剪辑视频片段B’和背景音乐M中处于节奏点S3和S4之间的音乐同时开始播放且同时结束播放,已剪辑视频片段C’和背景音乐M中处于节奏点S2和S3之间的音乐同时开始播放且同时结束播放。如此,使得拼接后的已剪辑视频片段和背景音乐能够很 好的配合,从而提高推荐视频V2的冲击力和表现力。
请参阅图2和图18,本申请实施方式的一个或多个包含计算机可执行指令302的非易失性计算机可读存储介质300,当计算机可执行指令302被一个或多个处理器20执行时,使得处理器20可执行上述任一实施方式的视频编辑方法。
例如,请结合图1,当计算机可执行指令302被一个或多个处理器20执行时,使得处理器20执行以下步骤:
011:在初始视频中标记一个或多个视频片段;
012:根据预设音乐片段的相邻两个节奏点之间的间隔时长,从一个或多个视频片段中确定待剪辑视频片段,待剪辑视频片段的时长大于或者等于间隔时长;及
013:剪辑待剪辑视频片段,得到已剪辑视频片段,以已剪辑视频片段的起始时刻和结束时刻分别对应两个节奏点。
再例如,请结合图5,当计算机可执行指令302被一个或多个处理器20执行时,处理器20还可以执行以下步骤:
0138:确定一个或多个视频片段中,时长大于间隔时长、且时长差为所有时长差中的最小值的视频片段为待剪辑视频片段,时长差为视频片段的时长和间隔时长的差值。
尽管上面已经示出和描述了本申请的实施方式,可以理解的是,上述实施方式是示例性的,不能理解为对本申请的限制,本领域的普通技术人员在本申请的范围内可以对上述实施方式进行变化、修改、替换和变型,本申请的范围由权利要求及其等同物限定。

Claims (20)

  1. 一种视频编辑方法,其特征在于,所述视频编辑方法包括:
    在初始视频中标记一个或多个视频片段;
    根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;及
    剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
  2. 根据权利要求1所述的视频编辑方法,其特征在于,所述根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,包括:
    确定所述一个或多个视频片段中,时长大于所述间隔时长、且时长差为所有时长差中的最小值的视频片段为所述待剪辑视频片段,所述时长差为所述视频片段的时长和所述间隔时长的差值。
  3. 根据权利要求1所述的视频编辑方法,其特征在于,所述剪辑所述待剪辑视频片段,得到已剪辑视频片段,包括:
    识别所述待剪辑视频片段中的静音开始时刻和静音结束时刻;
    在第一时长大于剪辑时长时,从所述起始时刻向后剪辑所述待剪辑视频片段,得到所述已剪辑视频片段,所述第一时长为所述待剪辑视频片段的起始时刻对应的静音时长,所述剪辑时长等于所述待剪辑视频片段中超过所述间隔时长的部分的时长;及
    在第二时长大于所述剪辑时长时,从所述结束时刻向前剪辑所述待剪辑视频片段,得到所述已剪辑视频片段,所述第二时长为所述待剪辑视频片段的结束时刻对应的所述静音时长。
  4. 根据权利要求3所述的视频编辑方法,其特征在于,所述剪辑所述待剪辑视频片段,得到已剪辑视频片段,还包括:
    在所述第一时长和所述第二时长均小于所述剪辑时长时,判断所述第一时长和所述第二时长之和是否大于所述剪辑时长;及
    在所述第一时长和所述第二时长之和大于所述剪辑时长时,分别从所述待剪辑视频片段的起始时刻和结束时刻剪辑所述待剪辑视频片段,得到所述已剪辑视频片段。
  5. 根据权利要求4所述的视频编辑方法,其特征在于,所述剪辑所述待剪辑视频片段,得到已剪辑视频片段,还包括:
    在所述第一时长和所述第二时长之和小于所述剪辑时长时,获取所述静音开始时刻对应的所述静音时长;及
    在所述静音开始时刻对应的所述静音时长大于所述剪辑时长时,从所述静音开始时刻剪辑所述待剪辑视频片段,得到所述已剪辑视频片段。
  6. 根据权利要求1所述的视频编辑方法,其特征在于,所述视频编辑方法还包括:
    识别选取自同一所述初始视频的多个视频片段中的重复部分;及
    将具有相同的所述重复部分的N个视频片段中的任意N-1个视频片段中的所述重复部分剪辑掉, 所述N为大于1的正整数。
  7. 根据权利要求1所述的视频编辑方法,其特征在于,所述视频编辑方法还包括:
    按照每个所述已剪辑视频片段的起始时刻和结束时刻将所有所述已剪辑视频片段拼接为一个融合视频片段;及
    整合所述融合视频片段和所述预定音乐片段以输出推荐视频。
  8. 一种视频的视频编辑装置,其特征在于,所述视频编辑装置包括:
    获取模块,用于在初始视频中标记一个或多个视频片段;
    确定模块,用于根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;及
    编辑模块,用于剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
  9. 根据权利要求8所述的视频编辑装置,其特征在于,所述确定模块包括确定单元,所述确定单元用于确定所述一个或多个视频片段中,时长大于所述间隔时长、且时长差为所有时长差中的最小值的视频片段为所述待剪辑视频片段,所述时长差为所述视频片段的时长和所述间隔时长的差值。
  10. 根据权利要求8所述的视频编辑装置,其特征在于,所述编辑模块包括识别单元、第一剪辑单元和第二剪辑单元,所述识别单元用于识别所述待剪辑视频片段中的静音开始时刻和静音结束时刻;所述第一剪辑单元用于在第一时长大于剪辑时长时,从所述起始时刻向后剪辑所述待剪辑视频片段,得到所述已剪辑视频片段,所述第一时长为所述待剪辑视频片段的起始时刻对应的静音时长,所述剪辑时长等于所述待剪辑视频片段中超过所述间隔时长的部分的时长;所述第二剪辑单元用于在第二时长大于所述剪辑时长时,从所述结束时刻向前剪辑所述待剪辑视频片段,得到所述已剪辑视频片段,所述第二时长为所述待剪辑视频片段的结束时刻对应的所述静音时长。
  11. 根据权利要求10所述的视频编辑装置,其特征在于,所述编辑模块还包括判断单元和第三剪辑单元,所述判断单元用于在所述第一时长和所述第二时长均小于所述剪辑时长时,判断所述第一时长和所述第二时长之和是否大于所述剪辑时长;所述第三剪辑单元用于在所述第一时长和所述第二时长之和大于所述剪辑时长时,分别从所述待剪辑视频片段的起始时刻和结束时刻剪辑所述待剪辑视频片段,得到所述已剪辑视频片段。
  12. 根据权利要求11所述的视频编辑装置,其特征在于,所述编辑模块还包括获取单元和第四剪辑单元,所述获取单元用于在所述第一时长和所述第二时长之和小于所述剪辑时长时,获取所述静音开始时刻对应的所述静音时长;;所述第四剪辑单元用于在所述静音开始时刻对应的所述静音时长大于所述剪辑时长时,从所述静音开始时刻剪辑所述待剪辑视频片段,得到所述已剪辑视频片段。
  13. 一种终端,其特征在于,所述终端包括处理器,所述处理器用于:
    在初始视频中标记一个或多个视频片段;
    根据预设音乐片段的相邻两个节奏点之间的间隔时长,从所述一个或多个视频片段中确定待剪辑视频片段,所述待剪辑视频片段的时长大于或者等于所述间隔时长;及
    剪辑所述待剪辑视频片段,得到已剪辑视频片段,以所述已剪辑视频片段的起始时刻和结束时刻分别对应两个所述节奏点。
  14. 根据权利要求13所述的终端,其特征在于,所述处理器还用于:
    确定所述一个或多个视频片段中,时长大于所述间隔时长、且时长差为所有时长差中的最小值的视频片段为所述待剪辑视频片段,所述时长差为所述视频片段的时长和所述间隔时长的差值。
  15. 根据权利要求13所述的终端,其特征在于,所述处理器还用于:
    识别所述待剪辑视频片段中的静音开始时刻和静音结束时刻;
    在第一时长大于剪辑时长时,从所述起始时刻向后剪辑所述待剪辑视频片段,得到所述已剪辑视频片段,所述第一时长为所述待剪辑视频片段的起始时刻对应的静音时长,所述剪辑时长等于所述待剪辑视频片段中超过所述间隔时长的部分的时长;及
    在第二时长大于所述剪辑时长时,从所述结束时刻向前剪辑所述待剪辑视频片段,得到所述已剪辑视频片段,所述第二时长为所述待剪辑视频片段的结束时刻对应的所述静音时长。
  16. 根据权利要求15所述的终端,其特征在于,所述处理器还用于:
    在所述第一时长和所述第二时长均小于所述剪辑时长时,判断所述第一时长和所述第二时长之和是否大于所述剪辑时长;及
    在所述第一时长和所述第二时长之和大于所述剪辑时长时,分别从所述待剪辑视频片段的起始时刻和结束时刻剪辑所述待剪辑视频片段,得到所述已剪辑视频片段。
  17. 根据权利要求16所述的终端,其特征在于,所述处理器还用于:
    在所述第一时长和所述第二时长之和小于所述剪辑时长时,获取所述静音开始时刻对应的所述静音时长;及
    在所述静音开始时刻对应的所述静音时长大于所述剪辑时长时,从所述静音开始时刻剪辑所述待剪辑视频片段,得到所述已剪辑视频片段。
  18. 根据权利要求13所述的终端,其特征在于,所述处理器还用于:
    识别选取自同一所述初始视频的多个视频片段中的重复部分;及
    将具有相同的所述重复部分的N个视频片段中的任意N-1个视频片段中的所述重复部分剪辑掉,所述N为大于1的正整数。
  19. 根据权利要求13所述的终端,其特征在于,所述处理器还用于:
    按照每个所述已剪辑视频片段的起始时刻和结束时刻将所有所述已剪辑视频片段拼接为一个融合视频片段;及
    整合所述融合视频片段和所述预定音乐片段以输出推荐视频。
  20. 一种包含计算机可执行指令的非易失性计算机可读存储介质,当所述计算机可执行指令被一个或多个处理器执行时,使得所述处理器执行如权利要求1至7中任一项所述的视频编辑方法。
PCT/CN2020/133047 2019-12-17 2020-12-01 视频编辑方法、视频编辑装置、终端和可读存储介质 WO2021121023A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP20903904.9A EP4068793A4 (en) 2019-12-17 2020-12-01 VIDEO EDITING METHOD, VIDEO EDITING APPARATUS, TERMINAL AND READABLE STORAGE MEDIA
US17/841,641 US20220312048A1 (en) 2019-12-17 2022-06-15 Video editing method, terminal and readable storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911298854.1A CN110992993B (zh) 2019-12-17 2019-12-17 视频编辑方法、视频编辑装置、终端和可读存储介质
CN201911298854.1 2019-12-17

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/841,641 Continuation US20220312048A1 (en) 2019-12-17 2022-06-15 Video editing method, terminal and readable storage medium

Publications (1)

Publication Number Publication Date
WO2021121023A1 true WO2021121023A1 (zh) 2021-06-24

Family

ID=70094431

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/133047 WO2021121023A1 (zh) 2019-12-17 2020-12-01 视频编辑方法、视频编辑装置、终端和可读存储介质

Country Status (4)

Country Link
US (1) US20220312048A1 (zh)
EP (1) EP4068793A4 (zh)
CN (1) CN110992993B (zh)
WO (1) WO2021121023A1 (zh)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110992993B (zh) * 2019-12-17 2022-12-09 Oppo广东移动通信有限公司 视频编辑方法、视频编辑装置、终端和可读存储介质
CN111918146B (zh) * 2020-07-28 2021-06-01 广州筷子信息科技有限公司 一种视频合成方法和系统
CN111901626B (zh) * 2020-08-05 2021-12-14 腾讯科技(深圳)有限公司 背景音频确定方法、视频剪辑方法、装置和计算机设备
CN115396691A (zh) * 2021-05-21 2022-11-25 北京金山云网络技术有限公司 一种数据流处理方法、装置及电子设备
CN113392206A (zh) * 2021-06-17 2021-09-14 李元烈 民俗文化热点内容智能编辑方法
CN113727038B (zh) * 2021-07-28 2023-09-05 北京达佳互联信息技术有限公司 一种视频处理方法、装置、电子设备及存储介质
CN113596579B (zh) * 2021-07-29 2023-04-07 北京字节跳动网络技术有限公司 视频生成方法、装置、介质及电子设备
CN113365147B (zh) * 2021-08-11 2021-11-19 腾讯科技(深圳)有限公司 基于音乐卡点的视频剪辑方法、装置、设备及存储介质

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107393569A (zh) * 2017-08-16 2017-11-24 成都品果科技有限公司 音视频剪辑方法及装置
CN107483843A (zh) * 2017-08-16 2017-12-15 成都品果科技有限公司 音视频匹配剪辑方法及装置
US20180167698A1 (en) * 2016-12-12 2018-06-14 Cooler Technologies, Inc. Instant clip creation based on media content recognition
US20180295427A1 (en) * 2017-04-07 2018-10-11 David Leiberman Systems and methods for creating composite videos
CN109168084A (zh) * 2018-10-24 2019-01-08 麒麟合盛网络技术股份有限公司 一种视频剪辑的方法和装置
CN109257545A (zh) * 2018-08-27 2019-01-22 咪咕文化科技有限公司 一种多源视频剪辑方法、装置及存储介质
CN110992993A (zh) * 2019-12-17 2020-04-10 Oppo广东移动通信有限公司 视频编辑方法、视频编辑装置、终端和可读存储介质

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7027124B2 (en) * 2002-02-28 2006-04-11 Fuji Xerox Co., Ltd. Method for automatically producing music videos
JP4117616B2 (ja) * 2003-07-28 2008-07-16 ソニー株式会社 編集システム、その制御方法及び編集装置
EP1750586A4 (en) * 2004-04-26 2009-07-29 U Systems Inc VERSATILE CHEST ULTRASOUND METAL PROCESS
US8128568B2 (en) * 2006-05-02 2012-03-06 U-Systems, Inc. Handheld volumetric ultrasound scanning device
US20080119733A1 (en) * 2006-11-22 2008-05-22 Wei Zhang Selectably compounding and displaying breast ultrasound images
US9747949B2 (en) * 2014-02-10 2017-08-29 Google Inc. Providing video transitions
CN110519638B (zh) * 2019-09-06 2023-05-16 Oppo广东移动通信有限公司 处理方法、处理装置、电子装置和存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180167698A1 (en) * 2016-12-12 2018-06-14 Cooler Technologies, Inc. Instant clip creation based on media content recognition
US20180295427A1 (en) * 2017-04-07 2018-10-11 David Leiberman Systems and methods for creating composite videos
CN107393569A (zh) * 2017-08-16 2017-11-24 成都品果科技有限公司 音视频剪辑方法及装置
CN107483843A (zh) * 2017-08-16 2017-12-15 成都品果科技有限公司 音视频匹配剪辑方法及装置
CN109257545A (zh) * 2018-08-27 2019-01-22 咪咕文化科技有限公司 一种多源视频剪辑方法、装置及存储介质
CN109168084A (zh) * 2018-10-24 2019-01-08 麒麟合盛网络技术股份有限公司 一种视频剪辑的方法和装置
CN110992993A (zh) * 2019-12-17 2020-04-10 Oppo广东移动通信有限公司 视频编辑方法、视频编辑装置、终端和可读存储介质

Also Published As

Publication number Publication date
EP4068793A4 (en) 2023-01-04
CN110992993A (zh) 2020-04-10
CN110992993B (zh) 2022-12-09
EP4068793A1 (en) 2022-10-05
US20220312048A1 (en) 2022-09-29

Similar Documents

Publication Publication Date Title
WO2021121023A1 (zh) 视频编辑方法、视频编辑装置、终端和可读存储介质
US20200302179A1 (en) Method for labeling performance segment, video playing method, apparaus and system
CN107396177B (zh) 视频播放方法、装置及存储介质
WO2021042933A1 (zh) 处理方法、处理装置、电子装置和存储介质
CN111050201B (zh) 数据处理方法、装置、电子设备及存储介质
US9213705B1 (en) Presenting content related to primary audio content
WO2021042926A1 (zh) 匹配方法、终端和可读存储介质
US9667773B2 (en) Audio file management for automated synchronization of an audio track with external video playback
US20200126559A1 (en) Creating multi-media from transcript-aligned media recordings
WO2019015613A1 (zh) 电子书语音播放方法、装置及终端设备
US20220084313A1 (en) Video processing methods and apparatuses, electronic devices, storage mediums and computer programs
WO2016119370A1 (zh) 一种实现录音的方法、装置和移动终端
CN112653902B (zh) 说话人识别方法、装置及电子设备
JP2005341015A (ja) 議事録作成支援機能を有するテレビ会議システム
WO2018130173A1 (zh) 配音方法、终端设备、服务器及存储介质
US20200143839A1 (en) Automatic video editing using beat matching detection
US10062130B2 (en) Generating authenticated instruments for oral agreements
JP2024513640A (ja) 仮想対象のアクション処理方法およびその装置、コンピュータプログラム
CN111246289A (zh) 视频生成方法及装置、电子设备、存储介质
CN113886612A (zh) 一种多媒体浏览方法、装置、设备及介质
US20230109852A1 (en) Data processing method and apparatus, device, and medium
CN111161710A (zh) 同声传译方法、装置、电子设备及存储介质
WO2022217944A1 (zh) 字幕与音源的绑定方法及装置
CN114339391A (zh) 视频数据处理方法、装置、计算机设备以及存储介质
WO2015131700A1 (zh) 文件存储方法及装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20903904

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2020903904

Country of ref document: EP

Effective date: 20220628

NENP Non-entry into the national phase

Ref country code: DE