WO2022127839A1 - 视频处理方法、装置、设备、存储介质及计算机程序产品 - Google Patents

视频处理方法、装置、设备、存储介质及计算机程序产品 Download PDF

Info

Publication number
WO2022127839A1
WO2022127839A1 PCT/CN2021/138507 CN2021138507W WO2022127839A1 WO 2022127839 A1 WO2022127839 A1 WO 2022127839A1 CN 2021138507 W CN2021138507 W CN 2021138507W WO 2022127839 A1 WO2022127839 A1 WO 2022127839A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
target
shooting
audio
speed
Prior art date
Application number
PCT/CN2021/138507
Other languages
English (en)
French (fr)
Inventor
叶欣靖
王林林
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Priority to EP21905765.0A priority Critical patent/EP4207783A4/en
Priority to JP2023537263A priority patent/JP2023554470A/ja
Publication of WO2022127839A1 publication Critical patent/WO2022127839A1/zh
Priority to US18/147,940 priority patent/US12003884B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording
    • H04N5/78Television signal recording using magnetic recording
    • H04N5/782Television signal recording using magnetic recording on tape
    • H04N5/783Adaptations for reproducing at a rate different from the recording rate
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/49Segmenting video sequences, i.e. computational techniques such as parsing or cutting the sequence, low-level clustering or determining units such as shots or scenes
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B20/00Signal processing not specific to the method of recording or reproducing; Circuits therefor
    • G11B20/10Digital recording or reproducing
    • G11B20/10527Audio or video recording; Data buffering arrangements
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/005Reproducing at a different information rate from the information rate of recording
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8549Creating video summaries, e.g. movie trailer

Definitions

  • the embodiments of the present disclosure relate to the field of computer technologies, and in particular, to a video processing method, apparatus, device, storage medium, and computer program product.
  • the video recording function is a common function of the client's video recorder or video recording application. With the development of science and technology, the application of video recording function is more and more extensive.
  • the video recording function can realize the production of videos in various ways, such as making slow-motion videos or fast-motion videos.
  • the page performs corresponding slow-play processing or fast-play processing, and then the production of slow-motion video or fast-motion video playback can be completed.
  • the video variable speed recording method in the prior art has cumbersome operation steps and complicated processing procedures, and cannot provide users with convenient operations, thereby affecting user experience.
  • Embodiments of the present disclosure provide a video processing method, apparatus, device, storage medium, and computer program product, which can provide users with convenient operations, and the processing process is simple to meet user needs and improve user operation experience.
  • an embodiment of the present disclosure provides a video processing method, the method comprising:
  • the first part in the audio indicates that in the captured video image Identifying the unplayed part of the audio when the target image presents the target state;
  • the video image captured before the target object presents the target state is played at a normal speed
  • the video image captured after the target object presents the target state is played at a second speed change mode
  • Perform variable speed playback and the audio is played at a normal speed
  • first shifting mode and the second shifting mode have opposite shifting directions and the same shifting speed.
  • an embodiment of the present disclosure provides a video processing apparatus, the apparatus comprising:
  • the shooting module is used to play audio during the process of shooting video
  • the processing module is configured to play the first part of the audio in variable speed according to the first variable speed mode in response to identifying that the target object presents the target state in the captured video image; When it is recognized in the obtained video image that the target image presents the target state, the unplayed part of the audio;
  • the processing module is further configured to obtain a target video according to the captured video image and the audio;
  • a playback module configured to play the video images captured before the target object presents the target state in the target video at a normal speed, and the video images captured after the target object presents the target state Perform variable speed playback according to the second variable speed mode, and the audio is played according to a normal speed;
  • first shifting mode and the second shifting mode have opposite shifting directions and the same shifting speed.
  • embodiments of the present disclosure provide an electronic device
  • processor including: at least one processor; and memory;
  • the memory stores computer-executable instructions
  • the at least one processor executes the computer-executable instructions stored in the memory, so that the at least one processor executes the video processing method according to any one of the first aspect above.
  • embodiments of the present disclosure provide a computer-readable storage medium, where computer-executable instructions are stored in the computer-readable storage medium, and when a processor executes the computer-executable instructions, any one of the above first aspects is implemented the video processing method.
  • embodiments of the present disclosure provide a computer program product, including a computer program, characterized in that, when the computer program is executed by a processor, the video processing method described in the first aspect above is implemented.
  • an embodiment of the present disclosure provides a computer program, which, when executed by a processor, is used to implement the video processing method described in the first aspect.
  • the video processing method, device, device, storage medium, and computer program product provided by the embodiments of the present disclosure play audio during the process of shooting a video, and then respond to identifying a target object in the captured video image and present a target state, according to
  • the first variable speed mode performs variable speed playback of the first part of the audio, where the first part represents the unplayed part of the audio when the target object is identified in the captured video image and presents the target state;
  • the audio obtains the target video, in the target video, the video image captured before the target object presents the target state is played at a constant speed, and the video image captured after the target object presents the target state is played at a variable speed according to the second variable speed mode, and
  • the audio is played at a constant speed, wherein the first variable speed mode and the second variable speed mode are in opposite directions and at the same speed, ensuring that the audio in the target video is played at a constant speed, and the video image is played at a constant speed and played at a variable speed.
  • variable-speed recording and playback of shooting video the operation of controlling the variable-speed video recording can be started directly by recognizing the target state presented by the target object, and there is no need to import the captured video into the video editing software or editing page for corresponding slow-play or fast-play processing.
  • the played audio is played in the first variable-speed mode.
  • the video shot is played at a variable speed according to the second variable speed mode, while the audio is played at a constant speed, which meets the individual needs of users.
  • the processing process is simple and the operation steps are simplified, so that users can realize the video through convenient operations. change processing, thereby improving the user experience.
  • FIG. 1 is a schematic diagram of a shooting interface provided by an embodiment of the present disclosure
  • FIG. 2 is a schematic diagram of a scene of a video processing method provided by an embodiment of the present disclosure
  • FIG. 3 is a schematic diagram of a scene of a video processing method provided by another embodiment of the present disclosure.
  • FIG. 4 is a schematic flowchart of a video processing method provided by another embodiment of the present disclosure.
  • FIG. 5 is a schematic flowchart of a video processing method provided by an embodiment of the present disclosure.
  • FIG. 6 is a schematic scene diagram of a video processing method provided by still another embodiment of the present disclosure.
  • FIG. 7 is a structural block diagram of a video processing apparatus provided by an embodiment of the present disclosure.
  • FIG. 8 is a schematic diagram of a hardware structure of an electronic device according to an embodiment of the present disclosure.
  • the term “including” and variations thereof are open-ended inclusions, ie, "including but not limited to”.
  • the term “based on” is “based at least in part on.”
  • the term “one embodiment” means “at least one embodiment”; the term “another embodiment” means “at least one additional embodiment”; the term “some embodiments” means “at least some embodiments”. Relevant definitions of other terms will be given in the description below.
  • the video recording function can realize the production of videos in various ways, such as making slow-motion videos or fast-motion videos.
  • the page performs corresponding slow-play processing or fast-play processing, and then the production of slow-motion video or fast-motion video playback can be completed.
  • there are also platforms in the prior art that include fast and slow functions in the shooting function but the fast and slow speeds here are only processed in real-time during the shooting process and presented to the user. The requirements for real-time processing are difficult and the recording time is short. Usually 1s. Therefore, the video recording method in the prior art has cumbersome operation steps and complicated processing procedures, and cannot provide users with convenient operations, thereby affecting user experience.
  • the technical idea of the present disclosure is to configure a variable speed trigger operation, and by identifying the variable speed trigger operation and the acquired second variable speed recorded by the variable speed, firstly, based on the second variable speed mode, obtain the matching second variable speed mode.
  • the first shifting mode that is, the shifting direction of the first shifting mode and the second shifting mode are opposite and the shifting speed is the same, and then the video image captured after the target object identified in the captured video image presents the target state is in accordance with the second shifting mode.
  • variable speed playback method is used to prompt the user to turn on the variable speed processing, and then part of the video images in the target video generated by the video image and audio is played at the normal speed and the other is played at the second variable speed mode, while ensuring that the audio in the target video is at a constant speed. Play, complete the variable speed processing of the video, the processing process is simple and the operation steps are simplified, so that the user can realize the variable speed recording through convenient operations, thereby improving the user experience.
  • the execution body of the embodiment of the present disclosure may be a terminal device, such as a mobile terminal, a computer device (eg, a desktop computer, a notebook computer, an all-in-one computer, etc.), etc.
  • the mobile terminal may include a smart phone, a handheld computer, a tablet computer, etc. , mobile devices with playback functions, such as wearable devices with display screens.
  • the displayed user interface such as the shooting interface, can be used to shoot or record the target object, where the target object can be a person, scenery, etc., and the playback interface can be used to play the target video, and the target video can be recorded with variable speed.
  • the processed audio and video are not specifically limited in this disclosure.
  • a terminal device may display a shooting interface 10 for shooting a target object, and the shooting interface 10 is provided with a shooting button 101 and a shooting duration.
  • the switch button 102 , the special effect editing button 103 , the music selection button 104 , the album button 105 , and other shooting function display areas 106 may also be included.
  • the shooting operation of the target object can be triggered by clicking the shooting button 101 .
  • the location and size of the shooting button 101 , the shooting duration switching button 102 , the special effect editing button 103 , the music selection button 104 and other shooting function display areas can be set according to actual needs.
  • the shooting button 101 is used to provide the user with a shooting start function;
  • the shooting duration switching button 102 is used to provide the user with the shooting duration switching function;
  • the special effect editing effect can identify the target state of the shooting object, that is, the target object, and then trigger the variable speed processing operation; selecting the music button 104 can provide the user with a music library; the album button 105 can provide the user with shooting materials.
  • the video processing method can be applied to a shooting or recording scene, and specifically can be a variable-speed video recording scene, for example, supports triggering of any special effect editing effect configured, and the special effect editing effect is used to support the function of controlling the variable-speed video recording , you can identify whether there is a target state in the captured video image. If the target state is identified during the shooting or recording process, that is, when a trigger operation for controlling video variable speed recording is detected, variable speed processing can be implemented, such as for Variable-speed processing of video images or variable-speed playback of audio during shooting, etc., do not require cumbersome operation steps.
  • the present disclosure does not limit application scenarios.
  • the terminal device can display multiple functions in the function display area in the shooting interface 10, such as flipping, filters, beautification, countdown, special effects editing effects, photo albums, music selection, shooting duration, etc., among which, the special effects editing effects here
  • It can include special effects editing effects to support the variable speed recording function of the video, such as predefined stickers.
  • the special effects in the predefined stickers can contain the speed that requires variable speed processing. Recording at the required speed can realize variable speed recording, and then identify the target by identifying the target.
  • the object presents the target state, and obtains the speed change mode (including the speed change direction and speed change speed) configured by the special effect editing effect; the selectable speed change processing speed and speed change direction (fast or slow) can also be provided to the user through the shooting interface, and then Based on the speed change mode selected by the user and the identifiable speed change trigger mode of the special effect configuration in the special effect editing effect, the target state presented by the target object is recognized.
  • the speed change mode including the speed change direction and speed change speed
  • speed change direction fast or slow
  • the special effects editing effects configured in the embodiments of the present disclosure can control variable-speed recording by identifying a trigger action or a specific object.
  • a sticker as an example, you can configure the sticker as follows: this sticker is reset when shooting; this sticker is only valid for recording, but not for preview; the corresponding variable speed of this sticker is configured in the resource package; variable playback events, variable speed, It needs to be transmitted to the Video Expression (VE) software development kit (SDK) and client in real time; a new variable-speed recording message interface is added, and the slowest control is required to be within 10 frames; when the sticker is made, it should be properly accelerated , and add frames appropriately. After entering the editing page, it will become normal speed after slow playback.
  • VE Video Expression
  • SDK software development kit
  • the shooting object is a character
  • the special effect editing effect itself is configured with a variable speed trigger special effect and a variable speed mode (it can be that one special effect editing effect corresponds to one variable speed trigger special effect, and one special effect corresponds to one variable speed mode; it can also be one special effect editing effect corresponding to Multiple variable speed trigger effects, one special effect corresponds to one variable speed mode; one special effect editing effect can also correspond to multiple variable speed trigger effects, and multiple special effects correspond to one variable speed mode; it can also be multiple special effect editing effects corresponding to one variable speed trigger effect, one The special effect corresponds to a speed change mode, etc., and there is a mapping relationship between the special effect and the speed change mode, which is not specifically limited here) as an example, the character is displayed on the shooting interface 10 .
  • the user can select the shooting time, such as 15s shown in FIG. 1, which means that the current shooting interface allows the longest shooting time of 15s.
  • a sticker such as sticker 1 replace the position of the special effect editing button with sticker 1 (see (c) in Figure 2), and display the background image corresponding to sticker 1 on the shooting interface; you can also click on the other functions mentioned above at the same time, such as Click the select music button 104 (refer to Fig. 3(a)) to display the music selection display interface 1041 (refer to Fig. 3(b)), the user can search the target music from the search bar and confirm the use, such as music 1 (see (c) in Figure 3).
  • the shooting function When the shooting function is turned on, The terminal device starts to shoot the target object, the current shooting interface displays the progress bar 110 of the shooting time, the end shooting button 107, and the function module is hidden, and the shooting button 101 is updated to the shooting state at the same time, that is, the shooting button 101 is switched to the shooting pause button 1011 ( See (b) in Figure 4), when it is detected that a special effect (such as a smile) appears during the shooting process, the variable speed recording function can be triggered (take slow recording as an example, see (c) in Figure 4), in During the shooting process, the video image is played at a normal speed, and at the same time, the audio is played at a rate in the opposite direction of the rate required by the special effect. The required variable speed is played, and the audio is played at a normal speed, thereby ensuring that the audio played when the video is played back is at a normal speed.
  • variable-speed recording can It is non-perceptible.
  • the video image and audio can be played at normal speed, or in order to meet the user experience, for variable speed recording, during the shooting process, the captured video image is played at normal speed.
  • the unplayed part is played according to the second variable speed mode matching the first variable speed mode required by the special effect.
  • the audio and video recorded before the special effect is played at the normal rate, and after the special effect trigger operation (ie, the variable speed trigger operation) is detected, the video is played at the rate required by the special effect and the audio is played at the normal rate, so as to realize the video variable speed recording process, which is
  • the user is provided with a convenient operation, and the processing process is simple, thereby satisfying the user's needs and improving the user's operation experience.
  • the constant speed of the video here refers to the shooting speed
  • the constant speed of the audio refers to its own speed.
  • the present disclosure does not limit application scenarios.
  • triggering operation may include (but is not limited to): click triggering, that is, the user triggers by clicking a corresponding button or any position in the interface; gesture triggering, the user makes a trigger by making Corresponding gestures are used to trigger operations; voice triggering, that is, users can trigger operations by issuing corresponding voice commands; expression triggering, that is, triggering operations by making corresponding expressions, etc.
  • click triggering that is, the user triggers by clicking a corresponding button or any position in the interface
  • gesture triggering the user makes a trigger by making Corresponding gestures are used to trigger operations
  • voice triggering that is, users can trigger operations by issuing corresponding voice commands
  • expression triggering that is, triggering operations by making corresponding expressions, etc.
  • the present disclosure does not limit trigger operations.
  • the realization of the video processing method can be realized in the following ways:
  • FIG. 5 is a schematic flowchart of a video processing method provided by an embodiment of the present disclosure.
  • the video processing method may include:
  • the first speed change mode may match the recognized target state. For example, if a smile triggers a speed change processing operation, and the speed change mode is 2x slow playback, the first speed change mode is 2 times faster; The input is selected before shooting video, and there is no specific limitation here.
  • the method of this embodiment of the present disclosure can be applied to a terminal device, and the terminal device can display a user interface for shooting a target object, that is, a shooting interface.
  • the shooting interface includes a shooting area, a function display area, and a shooting button.
  • the special effect editing effect module can be displayed. By clicking the special effect editing button corresponding to the special effect editing effect module, multiple special effect editing effects can be displayed to the user on the shooting interface.
  • the multiple special effect editing effects can include special effects editing for supporting the video variable speed recording function. Effect.
  • the display manner of multiple special effects editing effects is not specifically limited.
  • the user clicks the shooting button through the terminal device to enable the shooting function, and plays the default audio or the audio of the special effect editing effect or the audio selected by the user during the process of shooting the video.
  • the terminal device detects in real time whether there is an operation for controlling the video speed change processing during the shooting process, that is, recognizes whether the target object presents the target state, and if detected, responds to the recognized target state.
  • the first variable speed mode may match the recording mode required by special effects For example, if the recording and playback mode required by the special effect is 2x slow playback, the first variable speed mode is 2x fast playback; it can also be a user input method.
  • the target state may include interactive actions and/or preset objects included in the captured video, wherein the interactive actions may include facial and limb actions, such as smiling, blinking, raising a hand, raising a leg, etc.
  • the recognized target object is, for example, Preset objects such as clothes, cars, buildings, etc., are not limited in this disclosure.
  • the shooting interface needs to be displayed and the shooting function is turned on. Before that, the user needs to select the corresponding special effects of variable speed recording in advance, such as a special effect editing effect.
  • the terminal device can display the shooting interface and enable the shooting function through the following steps:
  • Step a1 Display the shooting interface for shooting the target object, where the shooting interface includes a special effect editing button for supporting the control of variable-speed video recording.
  • Step a2 in response to the first trigger operation acting on the special effect editing button, when receiving an instruction for instructing to perform the shooting operation on the shooting interface, perform the first shooting operation, and shoot the target object in the shooting interface.
  • the target state includes a state in which the target object presents a target action and/or a preset object is presented in the target object.
  • Step a3 If it is recognized that the target object is in the target state during the process of shooting the video, the operation of controlling the variable speed recording of the video is performed.
  • the terminal device when a user who shoots/records video (hereinafter referred to as a user) opens the application, the terminal device can display multiple switching interfaces, and the user can click the switching button matching the shooting interface to jump to display the shooting interface 10 .
  • the shooting interface may include special effects editing buttons for supporting variable-speed video recording, which may specifically be stickers, which may be a sticker corresponding to a recording function of one speed; or a sticker corresponding to a recording function of multiple speeds. Therefore, according to the action of the target object or what the target object is, the recording function of the speed can be triggered accordingly. For example, smile corresponds to slow 2x recording, waving corresponds to fast 2x recording, etc.
  • variable speed special effects can also be the special effect action in the sticker to trigger the variable speed processing operation.
  • the user can select other special effects to edit effects such as variable speed special effects
  • For editing effects during the shooting process, when the presented target state is recognized, corresponding speed change processing is performed according to the obtained speed change mode input by the user. The present disclosure does not limit this.
  • the user can click the special effect editing button to select from the special effect editing effect display area to support the control of video variable speed recording. If there are multiple special effects editing effects used to support variable-speed video recording, you can select the target special effect editing effect corresponding to the speed you want to shoot from the special effect editing effects, and then make a special effect editing effect based on the special effects in the special effect editing effect.
  • the corresponding target trigger operation can be generated; if there is only one special effect editing effect used to support the control of the variable speed recording of the video, the corresponding target trigger operation can be made based on the identifiable special effects in the special effect editing effect.
  • the terminal device responds to the trigger operation acting on the special effect editing button, that is, the first trigger operation, and displays the background environment of the special effect editing effect on the shooting interface.
  • the shooting button When the user clicks the shooting button to start shooting , an instruction for instructing to perform a shooting operation on the shooting interface is generated, and the terminal device receives the instruction and starts to perform the shooting operation, that is, the first shooting operation.
  • real-time detection of whether there is a target-triggered operation in the captured video such as smiling, waving, or interactive actions such as a target building and/or a target object, if there is, the client in the terminal device will trigger the operation with the target.
  • the corresponding variable speed recording message is sent to the audio and video terminals in the terminal device, so as to perform variable speed recording and save the current music.
  • the video image captured before the target object presents the target state is played at a normal speed
  • the video image captured after the target object presents the target state is played according to the second Variable speed playback is performed in a variable speed mode, and the audio is played at a normal speed
  • first shifting mode and the second shifting mode have opposite shifting directions and the same shifting speed.
  • the captured video image and audio are synthesized to obtain the audio and video to be played, that is, the target video.
  • the terminal device plays the target video.
  • the video image captured before the target object presents the target state is played at a constant speed
  • the video image captured by the target object is played at a variable speed according to the second variable speed mode.
  • the video image captured after the target state, and the audio is played at the normal speed.
  • the constant speed of video here refers to the speed of shooting
  • the constant speed of audio refers to its own speed.
  • the second speed change mode here is opposite to the speed change direction of the first speed change mode. For example, if the first speed change mode is twice the slow speed, the second speed change mode is twice the fast speed.
  • the purpose may be to play the target video after the shooting ends. , the music is played at a normal rate, which improves the user's viewing experience. In order to ensure that the audio played after the shooting is completed is not disturbed by the original sound played and/or randomly appeared before the variable speed recording, the original sound of the shooting is not recorded at the same time as the audio processing.
  • the second shifting mode is determined in response to a user's setting operation, and the first shifting mode is based on the Determined by the second speed change mode.
  • the target state that can be recognized by the special effect editing effect can determine its corresponding The type and/or speed of the shifting. It is also possible to identify special effects through special effects editing effects, and different actions in the special effects trigger different types of shifting, or even shifting at different speeds, and then determine the type and/or speed of shifting by the target state.
  • variable-speed playback of the unplayed part of the audio according to the first variable-speed mode in response to recognizing that the target object presents the target state in the captured video image can be achieved by the following steps:
  • Step b1 In response to identifying in the captured video image that the target object presents a target state, determine a first shifting mode that has a mapping relationship with the target state.
  • Step b2 according to the first variable speed mode, perform variable speed playback on the first part of the audio.
  • a first variable speed mode matching the target state is obtained through a mapping relationship, and then the first part of the audio is played in variable speed based on the first variable speed mode.
  • the second shifting mode is determined in response to the user's setting operation, and the first shifting mode is determined based on the second shifting mode.
  • the user may push the slider to set different speeds, and the user may select different types of speed through buttons. Shift is ok.
  • the shifting modes include the type and/or speed of shifting.
  • the second speed change mode may be the speed change mode required by the special effect.
  • the speed change mode required by the special effect is 2x slow playback
  • the first speed change mode is fast 2x playback
  • the second speed change mode is slow 2x playback ; can also be the way the user enters it.
  • the playback interface is displayed by jumping, and the video with the variable speed processing, that is, the target video, is played through the playback interface.
  • the target video here includes the video image taken before the target object presents the target state to change speed and the video image played according to the first variable speed mode.
  • the video image captured after the target object presents the target state and the audio played at constant speed.
  • the video images played in the first variable speed mode may be processed by adding or subtracting frames.
  • the terminal device when the shooting function is turned on, the terminal device starts shooting the target object and plays the currently shot video.
  • the content played here includes the video image when shooting the target object and the music played during the shooting process, until the variable speed is triggered, and the variable speed starts.
  • the first part of the audio is played in a variable speed manner according to the first variable speed mode, where the first part of the audio indicates that when the target object is identified in the captured video image and presents the target state, the audio Unplayed part.
  • the start time of the first part in the audio is the time when the shooting interface captures the target object in response to the recognition that the target object presents the target state in the captured video image (that is, the time when the variable speed processing is turned on);
  • the termination time of the first part is the time in response to the end of the shooting, which may be triggered by the user or may be the time when the shooting time arrives.
  • the start time of the first part in the audio is the time when the shooting is turned on, it means that the target state is detected immediately when the shooting function is turned on. Therefore, the video image played after the shooting is completed is a variable speed video image from the beginning. .
  • FIG. 6 is a schematic diagram of a scenario provided by yet another embodiment of the present disclosure.
  • the method may further include:
  • the playback interface is displayed in a jump.
  • the user can click the end shooting button 107 (see (a) in FIG. 6 ), and the terminal device jumps from the shooting interface 10 in response to the click operation, that is, the second trigger operation to the playback interface 20 and displayed (refer to (b) in FIG. 6 ). If the operation of the user triggering the end shooting button 107 is not detected during the shooting process, when the shooting time in the shooting interface ends, for example, the shooting progress bar on the shooting interface reaches the end position (see (c) in FIG. 6 ). , automatically jump from the shooting interface 10 to the playback interface 20 and display it (see (b) in FIG. 6 ).
  • the present disclosure does not specifically limit the way of switching and jumping between the shooting interface and the playing interface.
  • the present disclosure does not require the user to operate the screen, and only needs to smile at the camera to control the shooting speed. At the same time, there is no need to import the captured video into other editing pages or software, the processing process is simple, and the user experience is improved. It should be noted that the recognition of a smile is taken as an example to clearly illustrate the variable speed recording triggering and processing process, but the present disclosure does not limit interactive behaviors, including raising a hand, gagging, shaking his head, and so on.
  • audio is played during the process of shooting a video, and then in response to a target object being identified in the captured video image showing a target state, the first part of the audio is processed in a first variable speed manner. Perform variable-speed playback.
  • the first part here represents the part of the audio that is not played when the target object is identified in the captured video image and presents the target state; then the target video is obtained according to the captured video image and audio.
  • the video image captured before the target object presents the target state is played at a constant speed
  • the video image captured after the target object presents the target state is played at a variable speed according to the second variable speed mode
  • the audio is played at a constant speed
  • the first The speed change mode is opposite to the second speed change mode and the speed change speed is the same, which ensures that the audio in the target video is played at a constant speed
  • the video image is played at a constant speed and played at a variable speed, realizing the variable speed recording and playback of the captured video.
  • the operation of controlling the variable-speed video recording can be started directly by recognizing the target state presented by the target object, and there is no need to import the captured video into the video editing software or editing page for corresponding slow-play or fast-play processing.
  • the played audio is played in the first variable-speed mode.
  • the video shot is played at a variable speed according to the second variable speed mode, while the audio is played at a constant speed, which meets the individual needs of users.
  • the processing process is simple and the operation steps are simplified, so that users can realize the video through convenient operations. change processing, thereby improving the user experience.
  • how to obtain the target video can be implemented in at least two ways as follows:
  • Mode 1 Perform variable speed processing on the video image and then synthesize the target video with the audio played at normal speed.
  • Step c1 Record the captured video image at a normal speed to obtain a first video.
  • Step c2 Divide the first video into a first video segment and a second video segment; the first video segment represents a segment of the first video before the target object presents the target state; the The second video segment represents a segment of the second video after the target object exhibits the target state.
  • Step c3 Perform variable speed processing on the second video clip according to the second variable speed mode to obtain a third video clip.
  • Step c4 synthesizing the audio in the spliced video of the first video clip and the third video clip to obtain the target video.
  • the captured video images are first recorded at a normal speed to obtain a first video, and then the first video is divided into segments before the target object presents the target state, that is, the first video segment and the The segment after the target object presents the target state is the second video segment.
  • splicing the saved first video clip and the third video clip obtained by variable speed processing to obtain a spliced video, and at the same time synthesizing the audio recorded and saved during normal-speed playback and the spliced video to obtain a target video. Because it is background processing, not real-time processing, it can support variable-speed processing that is different from the prior art, and the operation is simple, no need to import other editing software, and the user experience is improved.
  • Mode 2 The target video is generated by splicing the variable-speed processed video image and the variable-speed processed audio.
  • Step d1 in the process of shooting video, record the played video image and audio to obtain a second video; in the second video, before the target object presents the target state, the audio is at a normal speed. Play, after the target object presents the target state, the audio is played at a variable speed according to the first variable speed mode, and the captured video image is played at a normal speed.
  • Step d2 dividing the second video into a third video segment and a fourth video segment;
  • the third video segment represents a segment of the second video before the target object presents the target state;
  • the The fourth video segment represents a segment of the second video after the target object exhibits the target state.
  • Step d3 performing variable speed processing on the fourth video clip according to the second variable speed mode to obtain a fifth video clip; and splicing the third video clip and the fifth video clip to obtain the target video.
  • the played video image and audio are recorded to obtain a second video, and the recorded audio is played at a normal speed before the target object presents the target state, and the recorded audio is presented on the target object
  • the audio is played at a variable speed according to the first variable speed mode, and the recorded video image is played at a normal speed.
  • the second video is divided into a segment before the target object presents the target state, that is, a third video segment, and a segment after the target object presents the target state, that is, a fourth video segment.
  • the third video clip is saved, the fourth video clip is subjected to variable speed processing according to the second variable speed mode, that is, frame addition or frame reduction is performed on the image and video in the fourth video clip, and the audio in the fourth video clip is processed by adding or subtracting frames.
  • the audio recorded by the variable speed playback in the first variable speed mode is played according to the second variable speed mode, so that the played audio is played at a constant speed.
  • the target video is obtained by splicing the saved third video clip with the fifth video clip subjected to variable speed processing. Because it is background processing, not real-time processing, it can support variable-speed processing that is different from the prior art, and the operation is simple, no need to import other editing software, and the user experience is improved.
  • the audio processed by the first variable speed mode can be obtained by performing variable speed recording first through a second variable speed mode with the same speed and opposite to that of the first variable speed mode. For example, if the first shifting speed is twice the slow speed, the second shifting speed is twice the fast speed. Then, the audio processed by the first variable speed mode is subjected to variable speed processing according to the second variable speed mode to obtain the audio played at a constant speed, that is, the audio played in the target video.
  • the terminal device acquires the target music currently playing when the variable-speed recording is triggered during the shooting process, and then performs variable-speed recording on the audio after the target state is displayed according to the determined second variable-speed mode, and generates the variable-speed audio that is not included in the shooting interface when shooting. of the original audio.
  • the starting time of the audio after the variable speed recording in the first variable speed mode is the time when the target state is recognized in response to the recognition (that is, the time when the recognized target object presents the target state);
  • the termination time of the recorded audio is the time corresponding to the progress of the music playing when the shooting interface shoots the target object at the end of the shooting (ie, the time when the variable-speed recording ends).
  • the speed of the speed change includes slow speed, fast speed and normal speed, and the speed is any value between 0.333-3, and the speed change can be expressed by the speed multiplier.
  • the first variable speed as 2 times the slow speed as an example, for example, the video image recorded at the normal rate is 30 frames/s, and at the rate of 2 times the slow speed, the corresponding frame reduction operation is performed on the video image, such as recording.
  • the video image is recorded 30 frames in 2s, which makes the playback rate of each frame of video image slow down during playback.
  • the video processing method may further include:
  • At least one of the following is reset: the first speed change mode, the second speed change mode, the state of the special effect editing effect, and the state of the shooting interface.
  • the user can click the end button in the shooting interface for indicating the end of shooting, and when the terminal device detects the trigger operation acting on the end button, that is, the second trigger operation, it sends a reset in response to the second trigger operation. Recording speed information, control VE reset status, client reset status, and then update shooting status. Specifically, the first shifting mode and/or the second shifting mode and/or the state of the special effect editing effect and/or the state of the shooting interface are reset.
  • the specific process for the user to click to start shooting/stop the special effect editing effect is as follows: the terminal device first determines whether to start shooting, if so, it updates the shooting status and ends the shooting status; When shooting ends, send a message to reset the recording speed, control the VE reset state and the client reset state, and then update the shooting state and end the shooting state.
  • a video processing method is described in detail. After jumping to display the playback interface, the video processing method can also be implemented by the following steps:
  • Step e1 In response to the first return trigger operation, control the playback interface to return to the shooting interface containing the target object to be shot.
  • Step e2 if an instruction for instructing re-shooting is received, then according to the instruction for instructing re-shooting, reset the content of the shooting interface containing the target object to be shot and exit the shooting interface .
  • the recorded video that is, the target video
  • the play interface When the user wants to return to the shooting interface to re-edit, re-shoot, or continue shooting, the user can click on the play interface.
  • the first return button of the terminal device in response to the first return trigger operation acting on the first return button, controls the playback interface to return to the photographing interface containing the target object to be photographed, that is, jumping to display the photographing interface in the paused photographing state.
  • the user can choose to cancel the current shooting content on the current shooting interface, that is, the user can click the second return button on the current shooting interface to generate an instruction for re-shooting and send it to the terminal device. This command resets the content of the shooting interface and exits the current shooting interface. It is convenient for users to re-shoot, and the interactive operation is simple.
  • the video processing method may further include the following steps:
  • Step f1 after the control of the playback interface to return to the shooting interface containing the target object to be shot or when receiving an instruction for instructing to suspend the shooting when shooting the target object on the shooting interface, detect the current shooting interface. Whether the progress bar on the shooting time has reached the end position.
  • Step f2 if the end position is not reached, when an instruction for instructing to perform a shooting operation on the shooting interface is received, control to continue to perform the shooting operation on the shooting interface.
  • the user may continue to continue shooting on the current shooting interface following the original shooting content. This can be achieved through at least two of the following scenarios:
  • Scenario 1 The current interface is in the play interface.
  • the user can click the first return button in the play interface to make the terminal device return the play interface to the shooting interface in the state of paused shooting.
  • the shooting interface in this state displays the shooting interface.
  • the progress bar of the time by detecting whether the progress bar of the shooting time on the shooting interface in this state has reached the end position to determine whether the shooting can continue. If the end position is not reached, the user can click the shooting on the shooting interface in this state. button, the terminal device receives an instruction for instructing to perform a shooting operation on the shooting interface, and continues to perform the shooting operation on the shooting interface according to the instruction.
  • Scenario 2 The current interface is in a shooting interface, and the shooting interface is in a shooting pause state in response to an instruction input by the user for instructing to pause shooting.
  • a progress bar of the shooting time is displayed on the shooting interface in this state. Whether the shooting can continue can be determined by detecting whether the progress bar of the shooting time on the shooting interface in this state has reached the end position. If it does not reach the end position, the user can Clicking the shooting button on the shooting interface in this state, the terminal device receives an instruction for instructing to perform a shooting operation on the shooting interface, and continues to perform the shooting operation on the shooting interface according to the instruction.
  • users in order to lower the threshold for users to shoot "slow motion transformation” videos, users can record slow motion videos for more than 1s or fast motion through special effects editing effects, and solve the problem of audio and video mismatch. , adding the logic of "recognizing a trigger condition and letting the recording speed up”, and transmits the variable speed playback event to the audio and video terminal and the client in real time through special effects, and performs variable speed recording on the audio and video terminal, which can help users produce more easily High-quality slow-motion video or fast-motion video for improved user experience.
  • FIG. 7 is a structural block diagram of the video processing apparatus provided by the embodiments of the present disclosure.
  • the video processing apparatus may be a terminal device.
  • the video processing apparatus 70 may include: a shooting module 701, a processing module 702 and a playing module 703; a shooting module for playing audio in the process of shooting a video; a processing module 702 for responding to the The target object is identified in the video image of the target object, and the first part of the audio is played at a variable speed according to the first variable speed mode; the first part of the audio indicates that the target is recognized in the captured video image.
  • the processing module is also used to obtain the target video according to the captured video image and the audio;
  • the playback module is used to display the target video , the video image captured before the target object presents the target state is played at a constant speed, and the video image captured after the target object presents the target state is played at a variable speed according to the second variable speed mode, so The audio is played at a constant speed; wherein, the first and second speed-change modes have opposite speed-change directions and the same speed-change.
  • the shooting module 701 , the processing module 702 , and the playing module 703 provided by the embodiments of the present disclosure play the video during the process of shooting the video, and then respond to the recognition that the target object presents the target state in the captured video image, according to the first variable speed
  • the first part of the audio is played at a variable speed, and the first part here represents the part of the audio that is not played when the target object is identified in the captured video image and presents the target state; then the target is obtained according to the captured video image and audio.
  • Video, in the target video the video image captured before the target object presents the target state is played at a constant speed, and the video image captured after the target object presents the target state is played at a variable speed according to the second variable speed mode, and the audio is played at a normal speed.
  • the first speed change mode and the second speed change mode are in opposite directions and at the same speed, which ensures that the audio in the target video is played at a constant speed, and the video image is played at a constant speed and played at a variable speed.
  • Variable speed recording and playback of captured video can be started directly by recognizing the target state presented by the target object, and there is no need to import the captured video into the video editing software or editing page for corresponding slow-play or fast-play processing. In order to improve the user experience , so that the user has a strong perception of variable-speed recording. After the variable-speed recording is triggered during the shooting of the video, the played audio is played in the first variable-speed mode.
  • the video shot is played at a variable speed according to the second variable speed mode, while the audio is played at a constant speed, which meets the individual needs of users.
  • the processing process is simple and the operation steps are simplified, so that users can realize the video through convenient operations. change processing, thereby improving the user experience.
  • the apparatus provided by the embodiment of the present disclosure can be used to execute the technical solution of the video processing method embodiment described in any one of the first aspect above, and the implementation principle and technical effect thereof are similar, and are not repeated in the embodiment of the present disclosure.
  • the embodiment of the present disclosure describes the video processing apparatus in detail on the basis of the above-mentioned disclosed embodiments.
  • the target state includes a state in which the target object presents a target action and/or a preset object is presented in the target object.
  • the processing module 702 is specifically configured to: record the captured video image at a normal speed to obtain a first video; divide the first video into a first video segment and a first video Two video clips; the first video clip represents the clip of the first video before the target object presents the target state; the second video clip represents the second video when the target object presents The segment after the target state; performing variable speed processing on the second video segment according to the second variable speed mode to obtain a third video segment; in the spliced video of the first video segment and the third video segment Synthesize the audio to obtain the target video.
  • the processing module 702 is further specifically configured to: in the process of shooting a video, record the played video image and audio to obtain a second video; in the second video, in the Before the target object presents the target state, the audio is played at a constant speed, after the target object presents the target state, the audio is played at a variable speed according to the first variable speed mode, and the captured video image is played at a constant speed playing; dividing the second video into a third video segment and a fourth video segment; the third video segment represents a segment of the second video before the target object presents the target state; the third video segment represents a segment of the second video before the target object presents the target state; The four video clips represent the clips of the second video after the target object presents the target state; the fourth video clip is subjected to variable speed processing according to the second variable speed mode to obtain a fifth video clip; The third video clip and the fifth video clip are spliced to obtain the target video.
  • the second shifting manner is determined in response to a user's setting operation, and the first shifting manner is determined based on the second shifting manner.
  • the processing module 702 is further configured to reset at least one of the following: the first shifting mode, the first Second, the speed change mode, the status of special effects editing effects, and the status of the shooting interface.
  • the playing module 703 is further configured to: before playing the target video, in response to a second trigger operation for instructing to end the shooting, jump to display a playing interface, where the playing interface is used for Play the target video.
  • the processing module 702 is further configured to: after the jumping to display the playback interface, in response to a first return trigger operation, control the playback interface to return to the playback interface containing the shooting.
  • the shooting interface of the target object when receiving an instruction for instructing re-shooting, reset the content of the shooting interface containing the target object to be shot according to the instruction for instructing re-shooting and exit the shooting interface.
  • the processing module 702 is further configured to: after the controlling the playback interface to return to the shooting interface containing the photographed target object or when the target object is photographed on the photographing interface When receiving an instruction to instruct to suspend shooting, detect whether the progress bar of the shooting time on the current shooting interface has reached the end position; , and control to continue to perform the shooting operation on the shooting interface.
  • the playing module 703 is further configured to: before playing the target video, when the shooting time in the shooting interface ends, jump to display the playing interface.
  • modules may be implemented as software components executing on one or more general-purpose processors, or as hardware, such as programmable logic devices and/or application specific integrated circuits, that perform certain functions or combinations thereof.
  • the modules may be embodied in the form of a software product that may be stored in non-volatile storage media including a computer device (eg, a personal computer, a server, a network device, mobile terminal, etc.) to implement the method described in the embodiments of the present invention.
  • a computer device eg, a personal computer, a server, a network device, mobile terminal, etc.
  • the above-mentioned modules can also be implemented on a single device, or can be distributed on multiple devices. The functions of these modules can be combined with each other or further split into multiple sub-modules.
  • an embodiment of the present application further provides an electronic device, including a processor and a memory;
  • the memory is used to store programs for executing the methods described in the above method embodiments; the processor is configured to execute the programs stored in the memory.
  • the electronic device may be a terminal device.
  • the electronic device includes a processor and a memory; the memory is used for storing a program for executing the methods described in the above method embodiments; the processor is configured to execute the program stored in the memory.
  • a processor may be a processing unit that includes a central processing unit (CPU) or other form of processing unit having data processing capabilities and/or instruction execution capabilities, and may control other components in an electronic device to perform desired functions.
  • the memory may include one or more computer program products, which may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory.
  • the volatile memory may include, for example, random access memory (RAM) and/or cache memory, or the like.
  • the non-volatile memory may include, for example, read only memory (ROM), hard disk, flash memory, and the like.
  • ROM read only memory
  • One or more computer program instructions may be stored on the computer-readable storage medium, which may be executed by a processor to implement the functions of the embodiments of the present disclosure described above and/or other desired functions.
  • Terminal devices may include, but are not limited to, such as mobile phones, notebook computers, digital broadcast receivers, personal digital assistants (Personal Digital Assistant, referred to as PDA), tablet computers, portable multimedia players (Portable Media Player, referred to as PMP), vehicle terminal ( Mobile terminals such as in-vehicle navigation terminals), wearable electronic devices, etc., and stationary terminals such as digital TVs, desktop computers, and the like.
  • PDA Personal Digital Assistant
  • PMP portable multimedia players
  • vehicle terminal Mobile terminals such as in-vehicle navigation terminals
  • wearable electronic devices etc.
  • stationary terminals such as digital TVs, desktop computers, and the like.
  • the electronic device shown in FIG. 8 is only an example, and should not impose any limitation on the function and scope of use of the embodiments of the present disclosure.
  • the electronic device may include a processing device (eg, a central processing unit, a graphics processing unit, etc.) 801 , which Various appropriate actions and processing can be performed according to a program stored in a read-only memory (Read Only Memory, referred to as ROM) 802 or a program loaded from a storage device 808 into a random access memory (Random Access Memory, referred to as RAM) 803 .
  • ROM Read Only Memory
  • RAM Random Access Memory
  • the processing device 801, the ROM 802, and the RAM 803 are connected to each other through a bus 804.
  • An input/output (I/O) interface 805 is also connected to bus 804 .
  • an input device 806 including, for example, a touch screen, a touch pad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, etc.; including, for example, a Liquid Crystal Display (LCD for short) ), speaker, vibrator, etc. output device 807; storage device 808 including, eg, magnetic tape, hard disk, etc.; and communication device 809.
  • Communication means 809 may allow the electronic device to communicate wirelessly or by wire with other devices to exchange data. While FIG. 8 illustrates an electronic device having various means, it should be understood that not all of the illustrated means are required to be implemented or available. More or fewer devices may alternatively be implemented or provided.
  • embodiments of the present disclosure include a computer program product comprising a computer program carried on a computer-readable medium, the computer program containing program code for performing the method illustrated in the flowchart.
  • the computer program may be downloaded and installed from the network via the communication device 809, or from the storage device 808, or from the ROM 802.
  • the processing device 801 the above-mentioned functions defined in the methods of the embodiments of the present disclosure are executed.
  • Embodiments of the present disclosure also include a computer program, which, when executed by a processor, is configured to perform the above-mentioned functions defined in the methods of the embodiments of the present disclosure.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two.
  • the computer-readable storage medium can be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or a combination of any of the above. More specific examples of computer readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Programmable Read Only Memory (EPROM), flash memory, optical fiber, portable Compact Disc-Read Only Memory (CD-ROM), optical storage device, magnetic storage device, or any suitable of the above The combination.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave with computer-readable program code embodied thereon. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device .
  • the program code embodied on the computer readable medium may be transmitted by any suitable medium, including but not limited to: electric wire, optical fiber cable, radio frequency (RF), etc., or any suitable combination of the above.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; or may exist alone without being assembled into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, causes the electronic device to execute the methods shown in the above disclosed embodiments.
  • Computer program code for carrying out operations of the present disclosure may be written in one or more programming languages, including object-oriented programming languages—such as Java, Smalltalk, C++, but also conventional Procedural programming language - such as the "C" language or similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network—including a Local Area Network (LAN) or a Wide Area Network (WAN)—or, can be connected to an external A computer (eg using an internet service provider to connect via the internet).
  • LAN Local Area Network
  • WAN Wide Area Network
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of code that contains one or more logical functions for implementing the specified functions executable instructions.
  • the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or operations , or can be implemented in a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments of the present disclosure may be implemented in a software manner, and may also be implemented in a hardware manner.
  • the name of the unit does not constitute a limitation of the unit itself under certain circumstances, for example, the first obtaining unit may also be described as "a unit that obtains at least two Internet Protocol addresses".
  • exemplary types of hardware logic components include: Field Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (Application Specific Standard Products) Standard Product, ASSP), system on chip (System On Chip, SOC), complex programmable logic device (Complex Programmable Logic Device, CPLD) and so on.
  • FPGAs Field Programmable Gate Arrays
  • ASICs Application Specific Integrated Circuits
  • ASSP Application Specific Standard Products
  • ASOC System On Chip
  • complex programmable logic device Complex Programmable Logic Device, CPLD
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with the instruction execution system, apparatus or device.
  • the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM), flash memory, optical fiber, portable compact disk read only memory (CD-ROM), optical storage devices, magnetic storage devices, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM erasable programmable read only memory
  • flash memory optical fiber
  • CD-ROM portable compact disk read only memory
  • optical storage devices magnetic storage devices, or any suitable combination of the foregoing.
  • Embodiments of the present disclosure provide a computer program product, including a computer program, characterized in that, when the computer program is executed by a processor, the video processing method described in the first aspect above is implemented.
  • an embodiment of the present disclosure provides a video processing method, including:
  • variable speed playback In response to recognizing that the target object presents the target state in the captured video image, perform variable speed playback on the first part of the audio according to the first variable speed mode; the first part in the audio indicates that in the captured video image Identifying the unplayed part of the audio when the target object presents the target state;
  • the video image captured before the target object presents the target state is played at a normal speed
  • the video image captured after the target object presents the target state is played at a second speed change mode
  • Perform variable speed playback and the audio is played at a normal speed
  • first shifting mode and the second shifting mode have opposite shifting directions and the same shifting speed.
  • variable-speed playback on the unplayed part of the audio according to the first variable-speed mode including:
  • determining a first shifting mode that has a mapping relationship with the target state In response to recognizing that the target object presents a target state in the captured video image, determining a first shifting mode that has a mapping relationship with the target state;
  • variable speed playback is performed on the first part of the audio.
  • the target state includes a state in which the target object presents a target action and/or a preset object is presented in the target object.
  • obtaining the target video according to the captured video image and the audio includes:
  • the first video is divided into a first video segment and a second video segment; the first video segment represents a segment of the first video before the target object presents the target state; the second video A segment represents a segment of the second video after the target object presents the target state;
  • the audio is synthesized in the spliced video of the first video clip and the third video clip to obtain the target video.
  • obtaining the target video according to the captured video image and the audio includes:
  • the played video image and audio are recorded to obtain a second video; in the second video, the audio is played at a normal speed before the target object presents the target state, and in the second video After the target object presents the target state, the audio is played at a variable speed according to the first variable speed mode, and the captured video image is played at a normal speed;
  • the second video is divided into a third video segment and a fourth video segment;
  • the third video segment represents a segment of the second video before the target object presents the target state;
  • the fourth video A segment represents a segment of the second video after the target object presents the target state;
  • the fourth video clip is subjected to variable speed processing according to the second variable speed mode to obtain a fifth video clip; and the target video is obtained by splicing the third video clip and the fifth video clip.
  • the second shifting manner is determined in response to a user's setting operation, and the first shifting manner is determined based on the second shifting manner.
  • an embodiment of the present disclosure provides a video processing apparatus, the apparatus comprising:
  • the shooting module is used to play audio during the process of shooting video
  • the processing module is configured to play the first part of the audio in variable speed according to the first variable speed mode in response to identifying that the target object presents the target state in the captured video image; When it is recognized in the obtained video image that the target image presents the target state, the unplayed part of the audio;
  • the processing module is further configured to obtain a target video according to the captured video image and the audio;
  • a playback module configured to play the video images captured before the target object presents the target state in the target video at a normal speed, and the video images captured after the target object presents the target state Perform variable speed playback according to the second variable speed mode, and the audio is played according to a normal speed;
  • first shifting mode and the second shifting mode have opposite shifting directions and the same shifting speed.
  • mapping relationship between the target state and the first shifting mode there is a mapping relationship between the target state and the first shifting mode; the processing module is specifically configured to: in response to recognizing that the target object presents the target state in the captured video image, A first variable speed mode having a mapping relationship with the target state is determined; according to the first variable speed mode, the first part of the audio is played in variable speed.
  • the target state includes a state in which the target object presents a target action and/or a preset object is presented in the target object.
  • the processing module 702 is specifically configured to: record the captured video image at a normal speed to obtain a first video; and divide the first video into first video segments and a second video clip; the first video clip represents a clip of the first video before the target object presents the target state; the second video clip represents, the second video is in the target state The object presents the segment after the target state; the second video segment is subjected to variable speed processing according to the second variable speed mode to obtain a third video segment; after the splicing of the first video segment and the third video segment The audio is synthesized in the video to obtain the target video.
  • the processing module 702 is further specifically configured to: in the process of shooting a video, record the played video image and audio to obtain a second video; in the second video, Before the target object presents the target state, the audio is played at a constant speed, after the target object presents the target state, the audio is played at a variable speed according to the first variable speed mode, and the captured video image is played according to normal speed playback; dividing the second video into a third video segment and a fourth video segment; the third video segment represents a segment of the second video before the target object presents the target state; the The fourth video clip represents a clip of the second video after the target object presents the target state; the fourth video clip is subjected to variable speed processing according to the second variable speed mode to obtain a fifth video clip; The target video is obtained by splicing the third video clip and the fifth video clip.
  • the second shifting manner is determined in response to a user's setting operation, and the first shifting manner is determined based on the second shifting manner.
  • embodiments of the present disclosure provide an electronic device
  • the memory stores computer-executable instructions
  • the at least one processor executes the computer-executable instructions stored in the memory, so that the at least one processor executes the video processing method according to any one of the first aspect above.
  • embodiments of the present disclosure provide a computer-readable storage medium, where computer-executable instructions are stored in the computer-readable storage medium, and when a processor executes the computer-executable instructions, any one of the above first aspects is implemented the video processing method.
  • embodiments of the present disclosure provide a computer program product, including a computer program, characterized in that, when the computer program is executed by a processor, the video processing method described in the first aspect above is implemented.
  • an embodiment of the present disclosure provides a computer program, which, when executed by a processor, is used to implement the video processing method described in the first aspect.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Computing Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Television Signal Processing For Recording (AREA)
  • User Interface Of Digital Computer (AREA)
  • Studio Devices (AREA)

Abstract

本公开实施例提供一种视频处理方法、装置、设备、存储介质及计算机程序产品,该方法包括:在拍摄视频的过程中播放音频;响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;根据拍摄到的视频图像和音频,得到目标视频;在目标视频中,在目标对象呈现目标状态之前拍摄到的视频图像按照常速进行播放,在目标对象呈现目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,音频按照常速进行播放;其中,第一变速方式与第二变速方式的变速方向相反且变速速度相同。本公开实施例能够为用户提供便捷的操作,并且处理过程简单进而满足了用户需求,提高了用户的操作体验。

Description

视频处理方法、装置、设备、存储介质及计算机程序产品
本申请要求于2020年12月18日提交的申请号为202011508607.2、名称为“视频处理方法、装置、设备、存储介质及计算机程序产品”的中国专利申请的优先权,此申请的内容通过引用并入本文。
技术领域
本公开实施例涉及计算机技术领域,尤其涉及一种视频处理方法、装置、设备、存储介质及计算机程序产品。
背景技术
视频录制功能是客户端的视频拍摄器或视频录制应用程序的常用功能。随着科技的发展,视频录制功能的应用越来越广泛。
目前,视频录制功能可以实现制作多种方式的视频,比如制作慢动作的视频或是快动作的视频等,然而实现多种方式的视频的制作是基于一段拍摄的视频导入视频编辑软件或是编辑页进行相应的慢放处理或是快放处理,然后才能完成制作实现慢动作的视频或是快动作的视频播放。
因此,现有技术的视频变速录制方式操作步骤繁琐且处理过程复杂,无法为用户提供便捷的操作,进而影响用户体验。
发明内容
本公开实施例提供一种视频处理方法、装置、设备、存储介质及计算机程序产品,能够为用户提供便捷的操作,并且处理过程简单进而满足了用户需求,提高了用户的操作体验。
第一方面,本公开实施例提供一种视频处理方法,所述方法包括:
在拍摄视频的过程中播放音频;
响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标图像呈现所述目标状态时,所述音频中未播放的部分;
根据拍摄到的视频图像和所述音频,得到目标视频;
在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
第二方面,本公开实施例提供一种视频处理装置,所述装置包括:
拍摄模块,用于在拍摄视频的过程中播放音频;
处理模块,用于响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标图像呈现所述目标状态时,所述音频中未播放的部分;
所述处理模块,还用于根据拍摄到的视频图像和所述音频,得到目标视频;
播放模块,用于在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
第三方面,本公开实施例提供一种电子设备,
包括:至少一个处理器;以及存储器;
所述存储器存储计算机执行指令;
所述至少一个处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如上述第一方面任一项所述的视频处理方法。
第四方面,本公开实施例提供一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上第一方面任一项所述的视频处理方法。
第五方面,本公开实施例提供一种计算机程序产品,包括计算机程序,其特征在于,该计算机程序被处理器执行时实现如上述第一方面所述的视频处理方法。
第六方面,本公开实施例提供一种计算机程序,所述计算机程序被处理器执行时,用于实现如上述第一方面所述的视频处理方法。
本公开实施例提供的视频处理方法、装置、设备、存储介质及计算机程序产品,通过在拍摄视频的过程中播放音频,然后响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对该音频中的第一部分进行变速播放,这里的第一部分表示在拍摄到的视频图像中识别到目标对象呈现目标状态时音频中未播放的部分;再根据拍摄到的视频图像和音频得到目标视频,在目标视频中,在目标对象呈现目标状态之前拍摄到的视频图像按照常速进行播放,在目标对象呈现目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,且音频按照常速进行播放,其中,第一变速方式与第二变速方式的变速方向相反且变速速度相同,保证在目标视频中音频是常速播放,而视频图像是一部分常速播放一部分变速播放,实现了对拍摄视频的变速录制以及播放。本申请直接通过识别目标对象呈现的目标状态即可开启控制视频变速录制的操作,无需将拍摄的视频导入视频编辑软件或是编辑页进行相应的慢放处理或是快放处理,为了提高用户体验,使得用户对变速录制有较强的感知,在拍摄视频过程中变速录制触发后,播放的音频是按照第一变速方式变速播放,而在拍摄完成后生成的目标视频播放过程中,变速触发后拍摄的视频是按照第二变速方式变速播放视频图像,而音频是按照常速播放的,满足了用户的个性化需求,处理过程简单且简化了操作步骤,使得用户通过便捷的操作即可实现视频的变处理,进而提高了用户体验。
附图说明
为了更清楚地说明本公开实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本公开的一些 实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本公开实施例提供的拍摄界面的示意图;
图2为本公开实施例提供的视频处理方法的场景示意图;
图3为本公开又一实施例提供的视频处理方法的场景示意图;
图4为本公开另一实施例提供的视频处理方法的流程示意图;
图5为本公开实施例提供的视频处理方法的流程示意图;
图6为本公开再一实施例提供的视频处理方法的场景示意图;
图7为本公开实施例提供的视频处理装置的结构框图;
图8为本公开实施例提供的电子设备的硬件结构示意图。
具体实施方式
下面将参照附图更详细地描述本公开的实施例。虽然附图中显示了本公开的某些实施例,然而应当理解的是,本公开可以通过各种形式来实现,而且不应该被解释为限于这里阐述的实施例,相反提供这些实施例是为了更加透彻和完整地理解本公开。应当理解的是,本公开的附图及实施例仅用于示例性作用,并非用于限制本公开的保护范围。
应当理解,本公开的方法实施方式中记载的各个步骤可以按照不同的顺序执行,和/或并行执行。此外,方法实施方式可以包括附加的步骤和/或省略执行示出的步骤。本公开的范围在此方面不受限制。
本文使用的术语“包括”及其变形是开放性包括,即“包括但不限于”。术语“基于”是“至少部分地基于”。术语“一个实施例”表示“至少一个实施例”;术语“另一实施例”表示“至少一个另外的实施例”;术语“一些实施例”表示“至少一些实施例”。其他术语的相关定义将在下文描述中给出。
需要注意,本公开中提及的“第一”、“第二”等概念仅用于对不同的装置、模块或单元进行区分,并非用于限定这些装置、模块或单元所执行的功能的顺序或者相互依存关系。
需要注意,本公开中提及的“一个”、“多个”的修饰是示意性而非限制性的,本领域技术人员应当理解,除非在上下文另有明确指出,否则应该理解为“一个或多个”。
本公开实施方式中的多个装置或模块之间所交互的消息或者信息的名称仅用于说明性的目的,而并不是用于对这些消息或信息的范围进行限制。
目前,视频录制功能可以实现制作多种方式的视频,比如制作慢动作的视频或是快动作的视频等,然而实现多种方式的视频的制作是基于一段拍摄的视频导入视频编辑软件或是编辑页进行相应的慢放处理或是快放处理,然后才能完成制作实现慢动作的视频或是快动作的视频播放。此外,现有技术中也有拍摄功能中含有快慢速功能的平台,但是这里的快慢速仅仅是拍摄过程中实时处理并呈现给用户的,实时处理的要求难度较大且录制时间较短,通常为1s。因此,现有技术的视频录制方式操作步骤繁琐且处理过程复杂,无法为用户提供便捷的操作,进而影响用户体验。
为了解决上述问题,本公开的技术构思是可以配置变速触发操作,通过识别变速触发操作,以及获取到的变速录制的第二变速方式,首先基于第二变速方式,获取与第二变速方式匹配的第一变速方式即第一变速方式与第二变速方式的变速方向相反且变速速度相同,然后 将在拍摄到的视频图像中识别到的目标对象呈现目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,用以提示用户开启了变速处理,然后将视频图像和音频生成的目标视频中的一部分视频图像按照常速播放一部分按照第二变速方式播放,同时保证目标视频中的音频常速播放,完成视频的变速处理,处理过程简单且简化了操作步骤,使得用户通过便捷的操作即可实现变速录制,进而提高了用户体验。
下面以具体地实施例对本公开的技术方案进行详细说明。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例不再赘述。
在实际应用中,本公开实施例的执行主体可以是终端设备,比如移动终端、计算机设备(如,台式机、笔记本电脑、一体机等)等,移动终端可以包括智能手机、掌上电脑、平板电脑、带显示屏的可穿戴设备等具有播放功能的移动设备。在实际应用中,显示的用户界面,比如拍摄界面可以用于拍摄或录制目标对象,这里的目标对象可以为人物、景色等,播放界面可以用于播放目标视频,该目标视频可以是经过变速录制处理过的音视频,本公开对此不做具体限定。
在一种场景下,参考图1,图1为本公开实施例提供的拍摄界面的示意图,终端设备可以显示用于拍摄目标对象的拍摄界面10,该拍摄界面10设置有拍摄按钮101、拍摄时长切换按钮102、特效编辑按钮103、选择音乐按钮104、相册按钮105,还可以包括其他拍摄功能展示区域106。其中,可以通过点击拍摄按钮101触发目标对象的拍摄操作。本领域技术人员应当理解,拍摄按钮101、拍摄时长切换按钮102、特效编辑按钮103、选择音乐按钮104以及其他拍摄功能展示区域所在位置和大小可以根据实际需求设置。其中,拍摄按钮101用于为用户提供拍摄启动功能;拍摄时长切换按钮102用于为用户提供拍摄时长切换功能;特效编辑按钮103用于为用户提供特效,其中特效中可以自带默认音乐,通过特效编辑效果可以识别拍摄对象即目标对象呈现的目标状态,即可触发变速处理操作;选择音乐按钮104可以为用户提供音乐库;相册按钮105用于为用户提供拍摄素材。
根据本公开实施例的视频处理方法可以应用于拍摄或录制场景,具体可以为视频变速录制场景,例如,支持触发配置好的任意特效编辑效果,该特效编辑效果用于支持控制视频变速录制的功能,可以通过识别拍摄到的视频图像中是否呈现有目标状态,若在拍摄或录制过程中识别到该目标状态即检测到用于控制视频变速录制的触发操作时,即可实现变速处理,比如对视频图像的变速处理或者拍摄时音频的变速播放等,无需繁琐的操作步骤。本公开不对应用场景进行限制。
下面以具体地实施例对本公开的技术方案进行详细说明。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例不再赘述。终端设备可以在该拍摄界面10中的功能展示区域中展示多个功能,比如翻转、滤镜、美化、倒计时、特效编辑效果、相册、选择音乐、拍摄时长等等,其中,这里的特效编辑效果可以包括用于支持控制视频变速录制功能的特效编辑效果,比如预定义贴纸,该预定义贴纸中特效可以包含需求变速处理的速率,以要求的速率进行录制可以实现变速录制,然后通过识别出目标对象呈现目标状态,获取该特效编辑效果配置的变速方式(包括变速方向和变速的速度);也可以通过拍摄界面向用户提供可选择的变速处理的速度以及变速方向(快速或慢速),然后基于用户选择的变速方式结合特效编辑效果中特效配置的可识别的变速触发方式即识别目标对象呈现的目标状态。
其中,本公开实施例配置的特效编辑效果,可以通过识别触发动作或是具体物体控制变速录制。以贴纸为例,可以对该贴纸进行如下配置:此贴纸开拍重置;此贴纸只能录制有效,预览无效;此贴纸相应的变速的速率配置在资源包内;变速播放事件、变速的速率,要实时传给视频表示(Video Expression,VE)软件开发工具包(Software Development Kit,SDK)和客户端;新增变速录制消息接口,要求最慢控制在10帧以内;贴纸在制作时,适当加速,并适当加帧,进入编辑页之后慢放之后变成正常速度,如果需要跟着画面变慢的效果,配置正常速度的贴纸即可,粒子效果、物理效果(布料、羽毛等)只能跟着变慢;变速播放事件、变速的速率,要实时传给VE和客户端,这里可能需要新开接口;其中,此贴纸禁止合拍;此贴纸不录原声;此贴纸需要禁用调整速率的功能。本公开不对此进行限制。
具体地,以拍摄对象为人物,且特效编辑效果本身配置有变速触发特效以及变速方式(可以是一个特效编辑效果对应一个变速触发特效,一个特效对应一个变速方式;也可以是一个特效编辑效果对应多个变速触发特效,一个特效对应一个变速方式;还可以是一个特效编辑效果对应多个变速触发特效,多个特效对应一个变速方式;还可以是多个特效编辑效果对应一个变速触发特效,一个特效对应一个变速方式等等,特效与变速方式存在映射关系,在此不做具体限定)为例,在拍摄界面10显示该人物。在拍摄界面10用户可以选择拍摄时长,比如图1中所示的15s,表示当前的拍摄界面允许最长拍摄15s。用户通过点击特效编辑按钮103(参见图2中(a)所示),显示特效编辑效果展示栏1031,包括多个贴纸(参见图2中(b)所示),选择特效编辑效果中的预定义贴纸,比如贴纸1,特效编辑按钮所在位置替换为贴纸1(参见图2中(c)所示),并且在拍摄界面显示该贴纸1对应的背景图像;还可以同时点击上述其他功能,比如点击选择音乐按钮104(参见图3中(a)所示),显示音乐选择展示界面1041(参见图3中(b)所示),用户可以从搜索栏中搜索目标音乐并确认使用,比如音乐1(参见图3中(c)所示)。
其中,需要说明的是,图2中(a)所示与图3中(a)所示的操作顺序不做具体限定。在图2中(c)和图3中(c)的基础上,用户可以通过点击拍摄界面中的拍摄按钮101开启拍摄功能(参见图4中(a)所示),当拍摄功能开启后,终端设备开始拍摄目标对象,当前的拍摄界面上显示拍摄时间的进度条110、结束拍摄按钮107,并且功能模块隐藏,同时拍摄按钮101更新为拍摄状态即由拍摄按钮101切换为拍摄暂停按钮1011(参见图4中(b)所示),当检测到拍摄过程中特效出现(比如微笑),即可触发变速录制功能(以慢速录制为例,参见图4中(c)所示),在拍摄过程中视频图像是按照常速播放的,与此同时音频以特效要求的速率相反方向的速率进行播放,并且在拍摄完成后播放视频时,通过对获取的触发变速操作之后拍摄视频图像按照特性要求的变速方式进行播放,音频按照常速播放,进而保证在播放拍摄完成的视频时播放的音频是正常速率。
以特效编辑效果的特效为微笑为例,在拍摄过程中若检测到微笑的动作,则触发用于控制视频变速录制的操作,获取该微笑特效要求变速录制的速率,比如慢速2倍,同时将以快速2倍的速率对特效触发后的音频进行后台(这里指音视频端)快速2倍的变速处理并播放,此时为了保证在录制完成后播放时播放的是音频不受变速录制之前播放和/或随机出现的原声的干扰,在对音频处理时不录制拍摄的原声,比如环境声音以及播放的音乐原声等,需要说明的是,在拍摄过程中,对于用户来说,变速录制可以是无感知的,播放视频图像以及音频可以是正常速度,也可以为了满足用户体验,针对变速录制,在拍摄过程中,拍摄到的视频 图像是常速播放,对音频中的触发变速操作开启后未播放的部分按照与特效要求的第一变速方式匹配的第二变速方式进行播放。在拍摄结束后按照正常速率播放特效出现前录制的音视频以及在检测到特效触发操作(即变速触发操作)后按照特效要求的速率播放视频以及按照正常速率播放音频,实现视频变速录制处理,为用户提供便捷的操作,并且处理过程简单进而满足了用户需求,提高了用户的操作体验。
其中,这里的视频的常速指的是拍摄的速度,音频的常速是指它本身的速度。本公开不对应用场景进行限制。
此外,本领域技术人员应当理解的是,“触发”操作可以包括(但不限于):点击触发,即用户通过点击对应的按钮或者界面中的任一位置实现触发;手势触发,用户通过做出相应的手势从而实现触发操作;语音触发,即用户通过发出相应的语音指令实现触发操作;表情触发,即通过做出相应的表情从而实现触发操作,等等,本公开不对触发操作进行限制。
在一个实施例中,视频处理方法的实现可以用过以下方式实现:
参考图5,图5为本公开实施例提供的视频处理方法的流程示意图。该视频处理方法可以包括:
S101、在拍摄视频的过程中播放音频;
S102、响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标对象呈现所述目标状态时,所述音频中未播放的部分。
其中,第一变速方式可以是与识别到的目标状态匹配的,比如,微笑触发变速处理操作,且变速方式为慢速2倍播放,则第一变速方式是快速2倍;也可以是由用户在拍摄视频之前选择输入的,在此不做具体限定。
本公开实施例的方法可以应用在终端设备中,在终端设备上可以显示用于拍摄目标对象的用户界面即拍摄界面,该拍摄界面上包括拍摄区域、功能展示区域以及拍摄按钮,在功能展示区可以展示特效编辑效果模块,通过点击特效编辑效果模块对应的特效编辑按钮可以在拍摄界面向用户展示多个特效编辑效果,这里多个特效编辑效果中可以包括用于支持视频变速录制功能的特效编辑效果。其中,多个特效编辑效果的展示方式不做具体限定。
具体地,用户通过终端设备点击拍摄按钮即可开启拍摄功能,在拍摄视频的过程中播放默认音频或是特效编辑效果自带音频或是用户选择的音频等等。终端设备在拍摄过程中实时检测是否存在用于控制视频变速处理的操作即识别目标对象是否呈现目标状态,如果检测到,则响应于该识别到的目标状态。首先获取在拍摄到的视频图像中识别到目标对象呈现目标状态时音频中未播放的部分所需变速播放的变速方式即第一变速方式,该第一变速方式可以是与特效要求的录制方式匹配的方式,比如特效要求的录制播放方式为慢速2倍播放,则第一变速方式为快速2倍播放;也可以是用户输入的方式。该目标状态可以包括交互动作和/或拍摄视频中包含的预设物体,其中,交互动作可以包括脸部,四肢动作,如微笑,眨眼,举手,抬腿等,识别到的目标对象是比如衣服,汽车,建筑等预设物体,本公开不对此进行限定。
在实际应用中,在变速触发之前,需要显示拍摄界面并开启拍摄功能,在此之前需要用户提前选取变速录制相应的特效比如某一特效编辑效果。具体地,终端设备可以通过以下步骤实现拍摄界面的显示以及拍摄功能的开启:
步骤a1、显示用于拍摄目标对象的所述拍摄界面,所述拍摄界面包括用于支持控制视频变速录制的特效编辑按钮。
步骤a2、响应作用于所述特效编辑按钮的第一触发操作,在接收到用于指示在拍摄界面执行拍摄操作的指令时,执行第一拍摄操作,并在所述拍摄界面中拍摄目标对象的同时,检测目标对象是否呈现目标状态,所述目标状态包括所述目标对象呈现目标动作的状态和/或所述目标对象中呈现有预设物体。
步骤a3、若在拍摄视频过程中识别到所述目标对象呈现目标状态,则执行控制视频变速录制的操作。
本公开实施例中,当拍摄/录制视频的用户(以下简称为用户)打开应用程序时,终端设备可以显示多个切换界面,用户可以点击与拍摄界面匹配的切换按键跳转显示拍摄界面10。其中,拍摄界面可以包括用于支持控制视频变速录制的特效编辑按钮,具体可以为贴纸,可以是一个贴纸对应开启一个速度的录制功能;也可以是一个贴纸对应多个速度的录制功能。因此,可以根据目标对象的动作或目标对象具体是什么,就相应的触发什么速度的录制功能。比如,微笑对应慢速2倍录制,挥手对应快速2倍录制等等,也可以是贴纸中的特效动作用于触发变速处理操作,用户在拍摄视频之前可以通过选择其他特效编辑效果比如变速方式特效编辑效果,在拍摄过程中,识别到呈现的目标状态时,依据获取到的用户输入的变速方式进行相应的变速处理。本公开不对此进行限定。
以特效编辑效果中的特效配置有变速方式为例,在拍摄界面中,拍摄开启之前,用户可以通过点击特效编辑按钮,从特效编辑效果的展示区域中选择用于支持控制视频变速录制的特效编辑效果,其中,若用于支持控制视频变速录制的特效编辑效果为多个,则可以从特效编辑效果中选择想要拍摄的速率对应的目标特效编辑效果,然后基于该特效编辑效果中的特效做出相应的目标触发操作;若用于支持控制视频变速录制的特效编辑效果为一个,则可以基于特效编辑效果中可识别的特效做出相应的目标触发操作。
具体地,用户选中特效编辑效果后,终端设备响应作用于该特效编辑按钮上的触发操作即第一触发操作,在拍摄界面显示该特效编辑效果的背景环境,当用户点击开启拍摄的拍摄按钮时,生成用于指示在拍摄界面执行拍摄操作的指令,终端设备接收该指令,开始执行拍摄操作即第一拍摄操作。在拍摄过程中,实时检测拍摄的视频中是否存在目标触发操作,比如微笑、挥手或是目标建筑等交互动作和/或目标物体,若存在,则终端设备中的客户端将与该目标触发操作相应的变速录制消息发送至终端设备中的音视频端,用以对当前音乐进行变速录制并保存。
S103、根据拍摄到的视频图像和所述音频,得到目标视频。
S104、在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
本实施例中,将拍摄到的视频图像与音频进行合成,得到待播放的音视频即目标视频。然后拍摄结束后,终端设备进行目标视频的播放,在目标视频播放过程中,按照常速播放在目标对象呈现所述目标状态之前拍摄到的视频图像以及按照第二变速方式变速播放在目标对象呈现目标状态之后拍摄到的视频图像,同时按照常速播放音频。
这里的视频的常速指的是拍摄的速度,音频的常速是指它本身的速度。这里的第二变速方式与第一变速方式的变速方向相反,比如,第一变速方式是慢速2倍,则第二变速方式为快速2倍,目的可以是为了在拍摄结束后播放目标视频时,音乐的播放是以正常速率播放的,提高用户的观看体验。为了保证在拍摄完成后播放时播放的是音频不受变速录制之前播放和/或随机出现的原声的干扰,在对音频处理的同时不录制拍摄的原声。
在一种可能的设计中,所述目标状态与所述第一变速方式存在映射关系;或者,所述第二变速方式是响应于用户的设置操作确定的,所述第一变速方式是基于所述第二变速方式确定的。
本实施例中,针对目标状态与第一变速方式存在映射关系,可以是特效编辑效果本身默认了变速的类型和/或速度,因此,特效编辑效果所能识别到的目标状态即可确定其对应的变速的类型和/或速度。也可以是通过特效编辑效果识别特效,特效中不同的动作触发不同类型的变速,甚至不同速度的变速,进而通过目标状态确定变速的类型和/或速度。
其中,如何响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中未播放的部分进行变速播放,可以通过以下步骤实现:
步骤b1、响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,确定与所述目标状态具有映射关系的第一变速方式。
步骤b2、按照所述第一变速方式,对所述音频中的第一部分进行变速播放。
本实施例中,响应于识别到的目标状态,通过映射关系,获取与目标状态匹配的第一变速方式,然后基于第一变速方式对音频中的第一部分进行变速播放。
针对第二变速方式是响应于用户的设置操作确定的,所述第一变速方式是基于所述第二变速方式确定的,可以是用户通过滑竿推动设置不同的速度,用户通过按钮选择不同类型的变速确定的。在此不具体限定用户以何种方式选择或设置不同的变速方式,变速方式包括变速的类型和/或速度。
具体地,该第二变速方式可以是特效要求的变速方式,比如特效要求的变速方式为慢速2倍播放,则第一变速方式为快速2倍播放,第二变速方式为慢速2倍播放;也可以是用户输入的方式。具体地,拍摄结束后跳转显示播放界面,通过播放界面播放变速处理的视频即目标视频,这里的目标视频中包括在目标对象呈现目标状态变速之前拍摄的视频图像以及按照第一变速方式播放的在目标对象呈现目标状态之后拍摄到的视频图像以及常速播放的音频。其中,按照第一变速方式播放的视频图像可以通过加帧或是减帧操作处理的。
此外,当拍摄功能开启时,终端设备开始拍摄目标对象,并对当前拍摄的视频进行播放,这里播放的内容包括拍摄目标对象时的视频图像以及拍摄过程中播放的音乐,直到变速触发,开始变速处理,按照第一变速方式对音频中的第一部分进行变速播放,这里的音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标对象呈现所述目标状态时,所述音频中未播放的部分。其中,该音频中的第一部分的起始时间为响应于在拍摄到的视频图像中识别到目标对象呈现目标状态时,拍摄界面拍摄目标对象的时间(即开启变速处理的时刻);音频中的第一部分的终止时间为响应于拍摄结束的时间,该拍摄结束的时间可以是用户触发的也可以是拍摄时间到达时刻。示例性的,当音频中的第一部分的起始时间为拍摄开启的时间时,说明开启拍摄功能时立即检测出目标状态,因此,在完成拍摄后播放的视频图像一开始即为变速的视频图像。
其中,如何显示播放界面,可以参见图6所示,图6为本公开再一实施例提供的场景示意图。在播放目标视频之前,所述方法还可以包括:
响应于用于指示结束拍摄的第二触发操作,跳转显示所述播放界面,所述播放界面用于播放所述目标视频;或者,
当所述拍摄界面中的拍摄时间结束后,跳转显示所述播放界面。
本公开实施例中,用户在拍摄过程中,可以通过点击结束拍摄按钮107(参见图6中(a)所示),终端设备响应于该点击操作即第二触发操作,从拍摄界面10跳转至播放界面20并显示(参见图6中(b)所示)。若在拍摄过程中未检测到用户触发结束拍摄按钮107的操作,当拍摄界面中的拍摄时间结束后,比如,拍摄界面上的拍摄进度条到达结束位置(参见图6中(c)所示),从拍摄界面10自动跳转至播放界面20并显示(参见图6中(b)所示)。本公开对拍摄界面与播放界面切换跳转的方式不做具体限定。
以微笑动作作为目标触发操作(即目标状态),以微笑动作表示2倍慢速录制的触发动作为例,首先添加微笑动作对应的特效编辑效果,同时可以选择音乐也可以将该微笑动作对应的特效编辑效果的背景音乐作为拍摄的音乐,终端设备开始检测目标触发操作。具体地,是否有微笑动作,若有微笑动作,则确定当前微笑动作的出现是否在拍摄中,若是在拍摄中,则确定是否触发变速状态,若是,则将该微笑动作要求的速率对应的消息即按照2倍慢速录制的消息发送至音视频端VE,VE将默认音乐(拍摄界面当前播放的音乐)快速录制保存(这里的快速录制可以是按照2倍快速录制),同时不录制原声,然后将录制后的目标音频发送至客户端,客户端按照2倍慢放速度播放,缩短总时间,结束录制以及播放。
本公开无需用户操作屏幕,只需要对着摄像头完成微笑动作即可控制拍摄速度。同时,无需将拍摄的视频再导入其他编辑页或是软件中,处理过程简单,提高了用户体验。需要说明的是,以识别微笑为例,是为了清楚说明变速录制触发以及处理过程,但是本公开对交互的行为不限制,包括举手,堵嘴,摇头等等。
本公开实施例提供的视频处理方法,通过在拍摄视频的过程中播放音频,然后响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对该音频中的第一部分进行变速播放,这里的第一部分表示在拍摄到的视频图像中识别到目标对象呈现目标状态时音频中未播放的部分;再根据拍摄到的视频图像和音频得到目标视频,在目标视频中,在目标对象呈现目标状态之前拍摄到的视频图像按照常速进行播放,在目标对象呈现目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,且音频按照常速进行播放,其中,第一变速方式与第二变速方式的变速方向相反且变速速度相同,保证在目标视频中音频是常速播放,而视频图像是一部分常速播放一部分变速播放,实现了对拍摄视频的变速录制以及播放。本申请直接通过识别目标对象呈现的目标状态即可开启控制视频变速录制的操作,无需将拍摄的视频导入视频编辑软件或是编辑页进行相应的慢放处理或是快放处理,为了提高用户体验,使得用户对变速录制有较强的感知,在拍摄视频过程中变速录制触发后,播放的音频是按照第一变速方式变速播放,而在拍摄完成后生成的目标视频播放过程中,变速触发后拍摄的视频是按照第二变速方式变速播放视频图像,而音频是按照常速播放的,满足了用户的个性化需求,处理过程简单且简化了操作步骤,使得用户通过便捷的操作即可实现视频的变处理,进而提高了用户体验。
在本公开的一个实施例中,在上述实施例的基础上,对如何得到目标视频,可以通过以下至少两种方式实现:
方式1、将视频图像进行变速处理后再与常速播放的音频合成目标视频。
步骤c1、按照常速对所述拍摄到的视频图像进行录制,得到第一视频。
步骤c2、将所述第一视频划分成第一视频片段和第二视频片段;所述第一视频片段表示,所述第一视频在所述目标对象呈现所述目标状态之前的片段;所述第二视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段。
步骤c3、按照所述第二变速方式对所述第二视频片段进行变速处理,得到第三视频片段。
步骤c4、在所述第一视频片段和所述第三视频片段的拼接视频中合成所述音频,得到所述目标视频。
本实施例中,首先按照常速对拍摄到的视频图像进行录制,得到第一视频,然后将第一视频划分成在所述目标对象呈现所述目标状态之前的片段即第一视频片段和在所述目标对象呈现所述目标状态之后的片段即第二视频片段。保存第一视频片段不变,针对第二视频片段,按照第二变速方式对第二视频片段进行加帧或减帧处理,得到第三视频片段,这里的第三视频片段为变速处理后的视频图像。然后将保存的第一视频片段与变速处理得到的第三视频片段进行拼接得到拼接视频,同时将常速播放录制保存的音频与该拼接视频进行合成,得到目标视频。由于是后台处理,不是实时处理,因此可以支持区别于现有技术时长的变速处理,并且操作简单,无需导入其他编辑软件,提高了用户体验。
方式2、将经过变速处理的视频图像与经过变速处理后的音频进行拼接处理生成目标视频。
步骤d1、在拍摄视频的过程,对播放出的视频图像和音频进行录制,得到第二视频;在所述第二视频中,在所述目标对象呈现所述目标状态之前所述音频按照常速播放,在所述目标对象呈现所述目标状态之后所述音频按照第一变速方式进行变速播放,所述拍摄到的视频图像按照常速播放。
步骤d2、将所述第二视频划分成第三视频片段和第四视频片段;所述第三视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之前的片段;所述第四视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段。
步骤d3、按照所述第二变速方式对所述第四视频片段进行变速处理,得到第五视频片段;将所述第三视频片段和所述第五视频片段进行拼接,得到所述目标视频。
本实施例中,在拍摄视频的过程,对播放出的视频图像和音频进行录制,得到第二视频,并且,录制的在目标对象呈现目标状态之前音频按照常速播放,录制的在目标对象呈现目标状态之后音频按照第一变速方式变速播放,录制的视频图像按照常速播放。其中,将第二视频划分成在所述目标对象呈现所述目标状态之前的片段即第三视频片段和在所述目标对象呈现所述目标状态之后的片段即第四视频片段。
具体地,针对第三视频片段进行保存,针对第四视频片段按照第二变速方式变速处理即对第四视频片段中的图像视频进行加帧或减帧处理,对第四视频片段中的音频即经过第一变速方式变速播放录制的音频按照第二变速方式播放,使得播放的音频是常速播放。将保存的第三视频片段与经过变速处理的第五视频片段进行拼接,得到目标视频。由于是后台处理,不是实时处理,因此可以支持区别于现有技术时长的变速处理,并且操作简单,无需导入其他编辑软件,提高了用户体验。
其中,为了保证经过第一变速方式播放的音频为正常速度,则可以通过与第一变速方式的变速方向相反且速度相同的第二变速方式先进行变速录制得到第一变速方式处理后的音频,比如,第一变速速度是慢速2倍的速率,则第二变速速度为快速2倍的速率。然后再对第一变速方式处理后的音频按照第二变速方式进行变速处理,得到常速播放的音频即目标视频中播放的音频。
具体地,终端设备获取拍摄过程中触发变速录制时当前播放的目标音乐,然后按照确定的第二变速方式对呈现目标状态之后的音频进行变速录制,生成变速后的不包含在拍摄界面中拍摄时的原声的音频。
其中,经过第一变速方式变速录制后的音频的起始时间为响应于识别到呈现目标状态的时间(即响应于识别到的目标对象呈现目标状态时的时刻);该经过第一变速方式变速录制后的音频的终止时间为拍摄结束时,拍摄界面拍摄目标对象时播放音乐的进度对应的时间(即结束变速录制的时刻)。
其中,变速的速率包含了慢速、快速、正常速度,速度为0.333-3之间的任意数值,可以用速度倍率表示变速。示例性的,以第一变速速度为2倍慢速为例,比如正常速率录制的视频图像是30帧/s,按照2倍慢速的速率,对视频图像进行相应的减帧操作,比如录制的视频图像是2s录制30帧,使得在播放时,每帧视频图像播放的速率变慢。
在本公开的一个实施例中,在上述实施例的基础上,在拍摄功能开启后,该视频处理方法还可以包括:
响应于用于指示结束拍摄的第二触发操作,重置下述至少一项:所述第一变速方式、所述第二变速方式、特效编辑效果的状态、拍摄界面的状态。
本公开实施例中,用户可以点击拍摄界面中用于指示结束拍摄的结束按钮,终端设备检测到作用于结束按钮的触发操作即第二触发操作时,响应于该第二触发操作,发送重置录制速度的消息,控制VE重置状态、客户端重置状态,进而更新拍摄状态。具体地,重置第一变速方式和/或所述第二变速方式和/或特效编辑效果的状态和/或拍摄界面的状态。
具体地,用户点击开拍/停止特效编辑效果的具体流程为:终端设备首先判断是否开始拍摄,若开始拍摄,则更新拍摄状态,结束开始拍摄状态;若不是开始拍摄,则判断是否结束拍摄,若结束拍摄,则发送重置录制速度的消息,控制VE重置状态、客户端重置状态,进而更新拍摄状态,结束拍摄状态。
在本公开的一个实施例中,在上述实施例的基础上,对视频处理方法进行了详细说明。在跳转显示所述播放界面之后,视频处理方法还可以通过以下步骤实现:
步骤e1、响应于第一返回触发操作,控制所述播放界面返回至含有拍摄的所述目标对象的所述拍摄界面。
步骤e2、若接收到用于指示重新拍摄的指令,则根据所述用于指示重新拍摄的指令,重置所述含有拍摄的所述目标对象的所述拍摄界面的内容并退出所述拍摄界面。
本公开实施例中,在跳转显示播放界面之后,在播放界面播放录制的视频即目标视频,当用户想要返回拍摄界面重新编辑或是重新拍摄或是继续拍摄时,用户可以点击播放界面上的第一返回按钮,终端设备响应作用于该第一返回按钮的第一返回触发操作,控制播放界面返回至含有拍摄的目标对象的拍摄界面即跳转显示处于暂停拍摄状态的拍摄界面。用户可以在当前的拍摄界面选择取消当前拍摄的内容即用户可以点击当前的拍摄界面上的第二返回按 钮,生成用于指示重新拍摄的指令并发送至终端设备,终端设备接收该指令,并根据该指令重置该拍摄界面的内容并退出当前的拍摄界面。便于用户重新拍摄,且交互操作简单。
在本公开的一个实施例中,在上述实施例的基础上,对视频处理方法进行了详细说明。该视频处理方法还可以包括以下步骤:
步骤f1、在所述控制所述播放界面返回至含有拍摄的所述目标对象的所述拍摄界面之后或者在拍摄界面拍摄目标对象时接收到用于指示暂停拍摄的指令时,检测当前的拍摄界面上的拍摄时间的进度条是否达到结束位置。
步骤f2、若未到达结束位置,则当接收到用于指示在拍摄界面执行拍摄操作的指令时,控制在所述拍摄界面上继续执行拍摄操作。
本公开实施例中,如果拍摄界面中的拍摄时间还未结束,用户可以继续在当前的拍摄界面接着原来拍摄的内容继续拍摄。可以通过以下至少两种场景实现:
场景1:当前的界面处于播放界面,用户可以通过点击播放界面中的第一返回按钮,使得终端设备将播放界面返回至处于暂停拍摄状态下的拍摄界面,该状态下的拍摄界面上显示有拍摄时间的进度条,通过检测该状态下的拍摄界面上的拍摄时间的进度条是否到达结束位置来确定是否能继续拍摄,若未到达结束位置,则用户可以点击该状态下的拍摄界面上的拍摄按钮,终端设备接收用于指示在拍摄界面执行拍摄操作的指令,并根据该指令在拍摄界面上继续执行拍摄操作。
场景2:当前的界面处于拍摄界面,且拍摄界面通过响应于用户输入的用于指示暂停拍摄的指令处于暂停拍摄状态。该状态下的拍摄界面上显示有拍摄时间的进度条,通过检测该状态下的拍摄界面上的拍摄时间的进度条是否到达结束位置来确定是否能继续拍摄,若未到达结束位置,则用户可以点击该状态下的拍摄界面上的拍摄按钮,终端设备接收用于指示在拍摄界面执行拍摄操作的指令,并根据该指令在拍摄界面上继续执行拍摄操作。
因此,本公开为了降低用户拍摄“慢动作变身”类视频的门槛,使用户通过特效编辑效果录慢动作视频超过1s或录快动作,解决音画不匹配的问题,同在在特效包内,加入了“识别某个触发条件,让录制变速”的逻辑,且通过特效将变速播放事件实时传给音视频端和客户端,在音视频端进行变速录制,能够帮助用户更简单地生产更高质量的慢动作视频或快动作视频,提高了用户体验。
对应于上文公开实施例的视频处理方法,图7为本公开实施例提供的视频处理装置的结构框图。其中视频处理装置可以是终端设备。为了便于说明,仅示出了与本公开实施例相关的部分。参照图7,所述视频处理装置70可以包括:拍摄模块701、处理模块702和播放模块703;拍摄模块,用于在拍摄视频的过程中播放音频;处理模块702,用于响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标图像呈现所述目标状态时,所述音频中未播放的部分;所述处理模块,还用于根据拍摄到的视频图像和所述音频,得到目标视频;播放模块,用于在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
本公开实施例提供的拍摄模块701、处理模块702和播放模块703,通过在拍摄视频的过程中播放视频,然后响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对该音频中的第一部分进行变速播放,这里的第一部分表示在拍摄到的视频图像中识别到目标对象呈现目标状态时音频中未播放的部分;再根据拍摄到的视频图像和音频得到目标视频,在目标视频中,在目标对象呈现目标状态之前拍摄到的视频图像按照常速进行播放,在目标对象呈现目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,且音频按照常速进行播放,其中,第一变速方式与第二变速方式的变速方向相反且变速速度相同,保证在目标视频中音频是常速播放,而视频图像是一部分常速播放一部分变速播放,实现了对拍摄视频的变速录制以及播放。本申请直接通过识别目标对象呈现的目标状态即可开启控制视频变速录制的操作,无需将拍摄的视频导入视频编辑软件或是编辑页进行相应的慢放处理或是快放处理,为了提高用户体验,使得用户对变速录制有较强的感知,在拍摄视频过程中变速录制触发后,播放的音频是按照第一变速方式变速播放,而在拍摄完成后生成的目标视频播放过程中,变速触发后拍摄的视频是按照第二变速方式变速播放视频图像,而音频是按照常速播放的,满足了用户的个性化需求,处理过程简单且简化了操作步骤,使得用户通过便捷的操作即可实现视频的变处理,进而提高了用户体验。
本公开实施例提供的装置,可用于执行上文第一方面任一项所述的视频处理方法实施例的技术方案,其实现原理和技术效果类似,本公开实施例此处不再赘述。
在本公开的一个实施例中,本公开实施例在上述公开实施例的基础上,对视频处理装置进行了详细说明。其中,所述目标状态与所述第一变速方式存在映射关系;处理模块702,具体用于:响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,确定与所述目标状态具有映射关系的第一变速方式;按照所述第一变速方式,对所述音频中的第一部分进行变速播放。
在本公开的一个实施例中,所述目标状态包括所述目标对象呈现目标动作的状态和/或所述目标对象中呈现有预设物体。
在本公开的一个实施例中,处理模块702,具体用于:按照常速对所述拍摄到的视频图像进行录制,得到第一视频;将所述第一视频划分成第一视频片段和第二视频片段;所述第一视频片段表示,所述第一视频在所述目标对象呈现所述目标状态之前的片段;所述第二视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;按照所述第二变速方式对所述第二视频片段进行变速处理,得到第三视频片段;在所述第一视频片段和所述第三视频片段的拼接视频中合成所述音频,得到所述目标视频。
在本公开的一个实施例中,处理模块702,还具体用于:在拍摄视频的过程,对播放出的视频图像和音频进行录制,得到第二视频;在所述第二视频中,在所述目标对象呈现所述目标状态之前所述音频按照常速播放,在所述目标对象呈现所述目标状态之后所述音频按照第一变速方式进行变速播放,所述拍摄到的视频图像按照常速播放;将所述第二视频划分成第三视频片段和第四视频片段;所述第三视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之前的片段;所述第四视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;按照所述第二变速方式对所述第四视频片段进行变速处理,得到第五视频片段;将所述第三视频片段和所述第五视频片段进行拼接,得到所述目标视频。
在本公开的一个实施例中,所述第二变速方式是响应于用户的设置操作确定的,所述第一变速方式是基于所述第二变速方式确定的。
在本公开的一个实施例中,所述处理模块702,还用于:响应于用于指示结束拍摄的第二触发操作,重置下述至少一项:所述第一变速方式、所述第二变速方式、特效编辑效果的状态、拍摄界面的状态。
在本公开的一个实施例中,所述播放模块703,还用于:在播放目标视频之前,响应于用于指示结束拍摄的第二触发操作,跳转显示播放界面,所述播放界面用于播放所述目标视频。
在本公开的一个实施例中,所述处理模块702,还用于:在所述跳转显示所述播放界面之后,响应于第一返回触发操作,控制所述播放界面返回至含有拍摄的所述目标对象的所述拍摄界面;在接收到用于指示重新拍摄的指令时,根据所述用于指示重新拍摄的指令,重置所述含有拍摄的所述目标对象的所述拍摄界面的内容并退出所述拍摄界面。
在本公开的一个实施例中,所述处理模块702,还用于:在所述控制所述播放界面返回至含有拍摄的所述目标对象的所述拍摄界面之后或者在拍摄界面拍摄目标对象时接收到用于指示暂停拍摄的指令时,检测当前的拍摄界面上的拍摄时间的进度条是否到达结束位置;若未到达结束位置,则当接收到用于指示在拍摄界面执行拍摄操作的指令时,控制在所述拍摄界面上继续执行拍摄操作。
在本公开的一个实施例中,播放模块703,还用于:在播放目标视频之前,当所述拍摄界面中的拍摄时间结束后,跳转显示所述播放界面。
上述模块可以被实现为在一个或多个通用处理器上执行的软件组件,也可以被实现为诸如执行某些功能或其组合的硬件,诸如可编程逻辑设备和/或专用集成电路。在一些实施例中,这些模块可以体现为软件产品的形式,该软件产品可以存储在非易失性存储介质中,这些非易失性存储介质中包括使得计算机设备(例如个人计算机、服务器、网络设备、移动终端等)实现本发明实施例中描述的方法。在一个实施例中,上述模块还可以在单个设备上实现,也可以分布在多个设备上。这些模块的功能可以相互合并,也可以进一步拆分为多个子模块。
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的信息处理装置的模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。
基于与方法同样的发明构思,本申请实施例还提供一种电子设备,包括处理器和存储器;
存储器用于存储执行上述各个方法实施例所述方法的程序;处理器被配置为执行存储器中存储的程序。
该电子设备可以为终端设备。该电子设备,包括处理器和存储器;存储器用于存储执行上述各个方法实施例所述方法的程序;处理器被配置为执行存储器中存储的程序。处理器可以是包括中央处理单元(CPU)或者具有数据处理能力和/或指令执行能力的其它形式的处理单元,并且可以控制电子设备中的其它组件以执行期望的功能。存储器可以包括一个或多个计算机程序产品,所述计算机程序产品可以包括各种形式的计算机可读存储介质,例如易失性存储器和/或非易失性存储器。所述易失性存储器例如可以包括随机存取存储器(RAM)和/或高速缓冲存储器(cache)等。所述非易失性存储器例如可以包括只读存储器(ROM)、硬盘、闪存等。在所述计算机可读存储介质上可以存储一个或多个计算机程序指令,处理器可以运行所述程序指令,以实现上文所述的本公开的实施例的功能以及/或者其它期望的功能。
终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、个人数字助理(Personal Digital Assistant,简称PDA)、平板电脑、便携式多媒体播放器(Portable Media Player,简称PMP)、车载终端(例如车载导航终端)、可穿戴电子设备等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图8示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
参考图8,其示出了适于用来实现本公开实施例的电子设备的结构示意图,如图8所示,电子设备可以包括处理装置(例如中央处理器、图形处理器等)801,其可以根据存储在只读存储器(Read Only Memory,简称ROM)802中的程序或者从存储装置808加载到随机访问存储器(Random Access Memory,简称RAM)803中的程序而执行各种适当的动作和处理。在RAM 803中,还存储有电子设备操作所需的各种程序和数据。处理装置801、ROM 802以及RAM 803通过总线804彼此相连。输入/输出(I/O)接口805也连接至总线804。
通常,以下装置可以连接至I/O接口805:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置806;包括例如液晶显示器(Liquid Crystal Display,简称LCD)、扬声器、振动器等的输出装置807;包括例如磁带、硬盘等的存储装置808;以及通信装置809。通信装置809可以允许电子设备与其他设备进行无线或有线通信以交换数据。虽然图8示出了具有各种装置的电子设备,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置809从网络上被下载和安装,或者从存储装置808被安装,或者从ROM 802被安装。在该计算机程序被处理装置801执行时,执行本公开实施例的方法中限定的上述功能。本公开的实施例还包括一种计算机程序,所述计算机程序被处理器执行时,用于执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(Electrically Programmable Read Only Memory,EPROM)、闪存、光纤、便携式紧凑磁盘只读存储器(Compact Disc-Read Only Memory,CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何 适当的介质传输,包括但不限于:电线、光缆、射频(Radio Frequency,RF)等等,或者上述的任意合适的组合。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备执行上述公开实施例所示的方法。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(Local Area Network,简称LAN)或广域网(Wide Area Network,简称WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(Field Programmable Gate Array,FPGA)、专用集成电路(Application Specific Integrated Circuit,ASIC)、专用标准产品(Application Specific Standard Product,ASSP)、片上系统(System On Chip,SOC)、复杂可编程逻辑设备(Complex Programmable Logic Device,CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM)、快闪存储器、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
本公开实施例提供一种计算机程序产品,包括计算机程序,其特征在于,该计算机程序被处理器执行时实现如上述第一方面所述的视频处理方法。
第一方面,本公开实施例提供一种视频处理方法,包括:
在拍摄视频的过程中播放音频;
响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标对象呈现所述目标状态时,所述音频中未播放的部分;
根据拍摄到的视频图像和所述音频,得到目标视频;
在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
根据本公开的一个或多个实施例,所述目标状态与所述第一变速方式存在映射关系;
所述响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中未播放的部分进行变速播放,包括:
响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,确定与所述目标状态具有映射关系的第一变速方式;
按照所述第一变速方式,对所述音频中的第一部分进行变速播放。
根据本公开的一个或多个实施例,所述目标状态包括所述目标对象呈现目标动作的状态和/或所述目标对象中呈现有预设物体。
根据本公开的一个或多个实施例,所述根据拍摄到的视频图像和所述音频,得到目标视频,包括:
按照常速对所述拍摄到的视频图像进行录制,得到第一视频;
将所述第一视频划分成第一视频片段和第二视频片段;所述第一视频片段表示,所述第一视频在所述目标对象呈现所述目标状态之前的片段;所述第二视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;
按照所述第二变速方式对所述第二视频片段进行变速处理,得到第三视频片段;
在所述第一视频片段和所述第三视频片段的拼接视频中合成所述音频,得到所述目标视频。
根据本公开的一个或多个实施例,所述根据拍摄到的视频图像和所述音频,得到目标视频,包括:
在拍摄视频的过程,对播放出的视频图像和音频进行录制,得到第二视频;在所述第二视频中,在所述目标对象呈现所述目标状态之前所述音频按照常速播放,在所述目标对象呈现所述目标状态之后所述音频按照第一变速方式进行变速播放,所述拍摄到的视频图像按照常速播放;
将所述第二视频划分成第三视频片段和第四视频片段;所述第三视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之前的片段;所述第四视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;
按照所述第二变速方式对所述第四视频片段进行变速处理,得到第五视频片段;将所述第三视频片段和所述第五视频片段进行拼接,得到所述目标视频。
根据本公开的一个或多个实施例,所述第二变速方式是响应于用户的设置操作确定的,所述第一变速方式是基于所述第二变速方式确定的。
第二方面,本公开实施例提供一种视频处理装置,所述装置包括:
拍摄模块,用于在拍摄视频的过程中播放音频;
处理模块,用于响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标图像呈现所述目标状态时,所述音频中未播放的部分;
所述处理模块,还用于根据拍摄到的视频图像和所述音频,得到目标视频;
播放模块,用于在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
根据本公开的一个或多个实施例,所述目标状态与所述第一变速方式存在映射关系;处理模块,具体用于:响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,确定与所述目标状态具有映射关系的第一变速方式;按照所述第一变速方式,对所述音频中的第一部分进行变速播放。
根据本公开的一个或多个实施例,所述目标状态包括所述目标对象呈现目标动作的状态和/或所述目标对象中呈现有预设物体。
根据本公开的一个或多个实施例,处理模块702,具体用于:按照常速对所述拍摄到的视频图像进行录制,得到第一视频;将所述第一视频划分成第一视频片段和第二视频片段;所述第一视频片段表示,所述第一视频在所述目标对象呈现所述目标状态之前的片段;所述第二视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;按照所述第二变速方式对所述第二视频片段进行变速处理,得到第三视频片段;在所述第一视频片段和所述第三视频片段的拼接视频中合成所述音频,得到所述目标视频。
根据本公开的一个或多个实施例,处理模块702,还具体用于:在拍摄视频的过程,对播放出的视频图像和音频进行录制,得到第二视频;在所述第二视频中,在所述目标对象呈现所述目标状态之前所述音频按照常速播放,在所述目标对象呈现所述目标状态之后所述音频按照第一变速方式进行变速播放,所述拍摄到的视频图像按照常速播放;将所述第二视频划分成第三视频片段和第四视频片段;所述第三视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之前的片段;所述第四视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;按照所述第二变速方式对所述第四视频片段进行变速处理,得到第五视频片段;将所述第三视频片段和所述第五视频片段进行拼接,得到所述目标视频。
根据本公开的一个或多个实施例,所述第二变速方式是响应于用户的设置操作确定的,所述第一变速方式是基于所述第二变速方式确定的。
第三方面,本公开实施例提供一种电子设备,
包括:至少一个处理器、存储器;
所述存储器存储计算机执行指令;
所述至少一个处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如上述第一方面任一项所述的视频处理方法。
第四方面,本公开实施例提供一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如上第一方面任一项所述的视频处理方法。
第五方面,本公开实施例提供一种计算机程序产品,包括计算机程序,其特征在于,该计算机程序被处理器执行时实现如上述第一方面所述的视频处理方法。
第六方面,本公开实施例提供一种计算机程序,所述计算机程序被处理器执行时,用于实现如上述第一方面所述的视频处理方法。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (14)

  1. 一种视频处理方法,其特征在于,所述方法包括:
    在拍摄视频的过程中播放音频;
    响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在所述拍摄到的视频图像中识别到所述目标对象呈现所述目标状态时,所述音频中未播放的部分;
    根据所述拍摄到的视频图像和所述音频,得到目标视频;
    在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
    其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
  2. 根据权利要求1所述的方法,其特征在于,所述目标状态与所述第一变速方式存在映射关系;
    所述响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中未播放的部分进行变速播放,包括:
    响应于在所述拍摄到的视频图像中识别到所述目标对象呈现所述目标状态,确定与所述目标状态具有映射关系的所述第一变速方式;
    按照所述第一变速方式,对所述音频中的第一部分进行变速播放。
  3. 根据权利要求1或2所述的方法,其特征在于,所述目标状态包括所述目标对象呈现目标动作的状态和/或所述目标对象中呈现有预设物体。
  4. 根据权利要求1-3任一项所述的方法,其特征在于,所述根据所述拍摄到的视频图像和所述音频,得到目标视频,包括:
    按照常速对所述拍摄到的视频图像进行录制,得到第一视频;
    将所述第一视频划分成第一视频片段和第二视频片段;所述第一视频片段表示,所述第一视频在所述目标对象呈现所述目标状态之前的片段;所述第二视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;
    按照所述第二变速方式对所述第二视频片段进行变速处理,得到第三视频片段;
    在所述第一视频片段和所述第三视频片段的拼接视频中合成所述音频,得到所述目标视频。
  5. 根据权利要求1-3任一项所述的方法,其特征在于,所述根据所述拍摄到的视频图像和所述音频,得到目标视频,包括:
    在拍摄视频的过程,对播放出的视频图像和音频进行录制,得到第二视频;在所述第二视频中,在所述目标对象呈现所述目标状态之前所述音频按照常速播放,在所述目标对象呈现所述目标状态之后所述音频按照第一变速方式进行变速播放,所述拍摄到的视频图像按照常速播放;
    将所述第二视频划分成第三视频片段和第四视频片段;所述第三视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之前的片段;所述第四视频片段表示,所述第二视频在所述目标对象呈现所述目标状态之后的片段;
    按照所述第二变速方式对所述第四视频片段进行变速处理,得到第五视频片段;将所述第三视频片段和所述第五视频片段进行拼接,得到所述目标视频。
  6. 根据权利要求1-5任一项所述的方法,其特征在于,所述第二变速方式是响应于用户的设置操作确定的,所述第一变速方式是基于所述第二变速方式确定的。
  7. 根据权利要求1-6任一项所述的方法,其特征在于,在所述拍摄视频的过程之前,所述方法包括显示拍摄界面并开启拍摄功能,具体包括:
    显示用于拍摄所述目标对象的所述拍摄界面,所述拍摄界面包括用于支持控制视频变速录制的特效编辑按钮;
    响应作用于所述特效编辑按钮的第一触发操作,在接收到用于指示在所述拍摄界面执行拍摄操作的指令时,执行第一拍摄操作,并在所述拍摄界面中拍摄所述目标对象的同时,识别所述目标对象是否呈现所述目标状态。
  8. 根据权利要求1-7任一项所述的方法,其特征在于,所述方法还包括:
    响应于用于指示结束拍摄的第二触发操作,跳转显示播放界面,所述播放界面用于播放所述目标视频;或者,
    当所述拍摄界面中的拍摄时间结束后,跳转显示所述播放界面;
    播放所述目标视频。
  9. 根据权利要求8所述的方法,其特征在于,在所述拍摄功能开启后,所述方法还包括:
    响应于所述用于指示结束拍摄的第二触发操作,重置下述至少一项:所述第一变速方式、所述第二变速方式、特效编辑效果的状态、拍摄界面的状态。
  10. 根据权利要求8或9所述的方法,其特征在于,在所述跳转显示所述播放界面之后,所述方法还包括:
    响应于第一返回触发操作,控制所述播放界面返回至含有拍摄的所述目标对象的所述拍摄界面。
  11. 一种视频处理装置,其特征在于,所述装置包括:
    拍摄模块,用于在拍摄视频的过程中播放音频;
    处理模块,用于响应于在拍摄到的视频图像中识别到目标对象呈现目标状态,按照第一变速方式对所述音频中的第一部分进行变速播放;所述音频中的第一部分表示,在拍摄到的视频图像中识别到所述目标图像呈现所述目标状态时,所述音频中未播放的部分;
    所述处理模块,还用于根据所述拍摄到的视频图像和所述音频,得到目标视频;
    播放模块,用于在所述目标视频中,在所述目标对象呈现所述目标状态之前拍摄到的视频图像按照常速进行播放,在所述目标对象呈现所述目标状态之后拍摄到的视频图像按照第二变速方式进行变速播放,所述音频按照常速进行播放;
    其中,所述第一变速方式与所述第二变速方式的变速方向相反且变速速度相同。
  12. 一种电子设备,其特征在于,包括:
    至少一个处理器;以及
    存储器;
    所述存储器存储计算机执行指令;
    所述至少一个处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如权利要求1至10任一项所述的视频处理方法。
  13. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如权利要求1至10任一项所述的视频处理方法。
  14. 一种计算机程序产品,包括计算机程序,其特征在于,该计算机程序被处理器执行时实现权利要求1-10任一项所述的视频处理方法。
PCT/CN2021/138507 2020-12-18 2021-12-15 视频处理方法、装置、设备、存储介质及计算机程序产品 WO2022127839A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP21905765.0A EP4207783A4 (en) 2020-12-18 2021-12-15 VIDEO PROCESSING METHOD AND APPARATUS, DEVICE, STORAGE MEDIUM AND COMPUTER PROGRAM PRODUCT
JP2023537263A JP2023554470A (ja) 2020-12-18 2021-12-15 ビデオ処理方法、装置、機器、記憶媒体、及びコンピュータプログラム製品
US18/147,940 US12003884B2 (en) 2020-12-18 2022-12-29 Video processing method and apparatus, device, storage medium and computer program product

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011508607.2 2020-12-18
CN202011508607.2A CN112653920B (zh) 2020-12-18 2020-12-18 视频处理方法、装置、设备及存储介质

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US18/147,940 Continuation US12003884B2 (en) 2020-12-18 2022-12-29 Video processing method and apparatus, device, storage medium and computer program product

Publications (1)

Publication Number Publication Date
WO2022127839A1 true WO2022127839A1 (zh) 2022-06-23

Family

ID=75355345

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/138507 WO2022127839A1 (zh) 2020-12-18 2021-12-15 视频处理方法、装置、设备、存储介质及计算机程序产品

Country Status (4)

Country Link
EP (1) EP4207783A4 (zh)
JP (1) JP2023554470A (zh)
CN (1) CN112653920B (zh)
WO (1) WO2022127839A1 (zh)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112653920B (zh) * 2020-12-18 2022-05-24 北京字跳网络技术有限公司 视频处理方法、装置、设备及存储介质
CN113726949B (zh) * 2021-05-31 2022-08-26 荣耀终端有限公司 一种视频处理方法、电子设备及存储介质
CN115442661B (zh) * 2021-06-01 2024-03-19 北京字跳网络技术有限公司 视频处理方法、设备、存储介质及计算机程序产品
CN113873319A (zh) * 2021-09-27 2021-12-31 维沃移动通信有限公司 视频处理方法、装置、电子设备及存储介质

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120011552A1 (en) * 2010-07-08 2012-01-12 Echostar Broadcasting Corporation Apparatus, Systems and Methods for Quick Speed Presentation of Media Content
CN107396019A (zh) * 2017-08-11 2017-11-24 维沃移动通信有限公司 一种慢动作视频录制方法及移动终端
CN108401124A (zh) * 2018-03-16 2018-08-14 广州酷狗计算机科技有限公司 视频录制的方法和装置
CN109376266A (zh) * 2018-09-18 2019-02-22 北京小米移动软件有限公司 慢动作视频的处理方法、装置、电子设备和存储介质
CN110086905A (zh) * 2018-03-26 2019-08-02 华为技术有限公司 一种录像方法及电子设备
CN112422863A (zh) * 2019-08-22 2021-02-26 华为技术有限公司 一种智能录像方法和装置
CN112653920A (zh) * 2020-12-18 2021-04-13 北京字跳网络技术有限公司 视频处理方法、装置、设备、存储介质及计算机程序产品
CN113067994A (zh) * 2021-03-31 2021-07-02 联想(北京)有限公司 一种视频录制方法及电子设备

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6424789B1 (en) * 1999-08-17 2002-07-23 Koninklijke Philips Electronics N.V. System and method for performing fast forward and slow motion speed changes in a video stream based on video content
CN104967900B (zh) * 2015-05-04 2018-08-07 腾讯科技(深圳)有限公司 一种生成视频的方法和装置
KR102537210B1 (ko) * 2016-07-29 2023-05-26 삼성전자주식회사 동영상 컨텐츠 제공 방법 및 이를 지원하는 전자 장치
WO2018085982A1 (zh) * 2016-11-08 2018-05-17 深圳市大疆创新科技有限公司 视频录制方法、装置及拍摄设备
US11317028B2 (en) * 2017-01-06 2022-04-26 Appsure Inc. Capture and display device
CN106851385B (zh) * 2017-02-20 2019-12-27 北京乐我无限科技有限责任公司 视频录制方法、装置和电子设备
JP2019020530A (ja) * 2017-07-13 2019-02-07 キヤノン株式会社 信号処理装置、制御方法、及びプログラム
CN108184165B (zh) * 2017-12-28 2020-08-07 Oppo广东移动通信有限公司 视频播放方法、电子装置和计算机可读存储介质
CN108616696B (zh) * 2018-07-19 2020-04-14 北京微播视界科技有限公司 一种视频拍摄方法、装置、终端设备及存储介质
WO2020019212A1 (zh) * 2018-07-25 2020-01-30 深圳市大疆创新科技有限公司 视频播放速度控制方法及系统、控制终端和可移动平台
CN113163133A (zh) * 2018-10-15 2021-07-23 华为技术有限公司 一种图像处理方法、装置与设备
WO2020172826A1 (zh) * 2019-02-27 2020-09-03 华为技术有限公司 一种视频处理方法和移动设备
CN109862193B (zh) * 2019-04-12 2020-10-02 珠海天燕科技有限公司 一种终端中来电视频的控制方法及装置
CN110337009A (zh) * 2019-07-01 2019-10-15 百度在线网络技术(北京)有限公司 视频播放的控制方法、装置、设备及存储介质
CN111722775A (zh) * 2020-06-24 2020-09-29 维沃移动通信(杭州)有限公司 图像处理方法、装置、设备及可读存储介质

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120011552A1 (en) * 2010-07-08 2012-01-12 Echostar Broadcasting Corporation Apparatus, Systems and Methods for Quick Speed Presentation of Media Content
CN107396019A (zh) * 2017-08-11 2017-11-24 维沃移动通信有限公司 一种慢动作视频录制方法及移动终端
CN108401124A (zh) * 2018-03-16 2018-08-14 广州酷狗计算机科技有限公司 视频录制的方法和装置
CN110086905A (zh) * 2018-03-26 2019-08-02 华为技术有限公司 一种录像方法及电子设备
CN109376266A (zh) * 2018-09-18 2019-02-22 北京小米移动软件有限公司 慢动作视频的处理方法、装置、电子设备和存储介质
CN112422863A (zh) * 2019-08-22 2021-02-26 华为技术有限公司 一种智能录像方法和装置
CN112653920A (zh) * 2020-12-18 2021-04-13 北京字跳网络技术有限公司 视频处理方法、装置、设备、存储介质及计算机程序产品
CN113067994A (zh) * 2021-03-31 2021-07-02 联想(北京)有限公司 一种视频录制方法及电子设备

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP4207783A4

Also Published As

Publication number Publication date
US20230133163A1 (en) 2023-05-04
CN112653920A (zh) 2021-04-13
EP4207783A1 (en) 2023-07-05
JP2023554470A (ja) 2023-12-27
CN112653920B (zh) 2022-05-24
EP4207783A4 (en) 2024-02-28

Similar Documents

Publication Publication Date Title
WO2020015333A1 (zh) 视频拍摄方法、装置、终端设备及存储介质
WO2022127839A1 (zh) 视频处理方法、装置、设备、存储介质及计算机程序产品
US11284005B2 (en) Video processing method and apparatus, terminal device, and storage medium
CN108900902B (zh) 确定视频背景音乐的方法、装置、终端设备及存储介质
WO2020010814A1 (zh) 选择背景音乐拍摄视频的方法、装置、终端设备及介质
WO2020077856A1 (zh) 视频拍摄方法、装置、电子设备及计算机可读存储介质
US11670339B2 (en) Video acquisition method and device, terminal and medium
US11632584B2 (en) Video switching during music playback
WO2022007724A1 (zh) 视频处理方法、装置、设备及存储介质
GB2593005A (en) Video generation method and device, electronic device and computer storage medium
WO2022007722A1 (zh) 显示方法、装置、设备及存储介质
WO2022089178A1 (zh) 视频处理方法及设备
WO2020015331A1 (zh) 视频拍摄方法、装置、终端设备和存储介质
US8837912B2 (en) Information processing apparatus, information processing method and program
WO2018085982A1 (zh) 视频录制方法、装置及拍摄设备
US20230307004A1 (en) Audio data processing method and apparatus, and device and storage medium
WO2021083145A1 (zh) 视频处理的方法、装置、终端及存储介质
US20240129427A1 (en) Video processing method and apparatus, and terminal and storage medium
WO2024046360A1 (zh) 媒体内容处理方法、装置、设备、可读存储介质及产品
WO2024032635A1 (zh) 媒体内容获取方法、装置、设备、可读存储介质及产品
WO2023088484A1 (zh) 用于多媒体资源剪辑场景的方法、装置、设备及存储介质
WO2023072280A1 (zh) 媒体内容发送方法、装置、设备、可读存储介质及产品
US12003884B2 (en) Video processing method and apparatus, device, storage medium and computer program product
WO2023185968A1 (zh) 相机功能页面切换方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21905765

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2021905765

Country of ref document: EP

Effective date: 20230331

WWE Wipo information: entry into national phase

Ref document number: 2023537263

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE