CN114520875A - Video processing method and device and electronic equipment - Google Patents

Video processing method and device and electronic equipment Download PDF

Info

Publication number
CN114520875A
CN114520875A CN202210109250.3A CN202210109250A CN114520875A CN 114520875 A CN114520875 A CN 114520875A CN 202210109250 A CN202210109250 A CN 202210109250A CN 114520875 A CN114520875 A CN 114520875A
Authority
CN
China
Prior art keywords
video
input
frames
sub
stream
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210109250.3A
Other languages
Chinese (zh)
Other versions
CN114520875B (en
Inventor
陈露兰
杨渊明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xi'an Weiwo Software Technology Co ltd
Original Assignee
Xi'an Weiwo Software Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xi'an Weiwo Software Technology Co ltd filed Critical Xi'an Weiwo Software Technology Co ltd
Priority to CN202210109250.3A priority Critical patent/CN114520875B/en
Publication of CN114520875A publication Critical patent/CN114520875A/en
Application granted granted Critical
Publication of CN114520875B publication Critical patent/CN114520875B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72439User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for image or video messaging
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/61Control of cameras or camera modules based on recognised objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2621Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/76Television signal recording

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Studio Devices (AREA)

Abstract

The application discloses a video processing method and device and electronic equipment, and belongs to the technical field of camera shooting. Wherein, the method comprises the following steps: displaying the first video stream and the second video stream; receiving a first input of a user; and responding to the first input, and synthesizing the N first video frames and the M second video frames selected by the first input to obtain the target video.

Description

Video processing method and device and electronic equipment
Technical Field
The application belongs to the technical field of camera shooting, and particularly relates to a video processing method and device and electronic equipment.
Background
Generally, in a scene where a user performs video shooting, the user may first trigger the electronic device to shoot to obtain a video file, then trigger the electronic device to start a video processing application, and add a plurality of special effects required for the video file in the video processing application to obtain a video file that the user is satisfied with.
However, after the video file is obtained by shooting with the electronic device, the user needs to trigger the electronic device to start the video processing application, and perform multiple operations to add a special effect required by the user to the video file, so that the user operation is cumbersome and time-consuming in the process of obtaining the video file satisfied by the user by shooting.
As such, the efficiency of the electronic device capturing the video file is low.
Disclosure of Invention
The embodiment of the application aims to provide a video processing method, a video processing device and electronic equipment, which can simplify user operation in the process of shooting a video file satisfied by a user, reduce time consumption and improve the efficiency of shooting the video file by the electronic equipment.
In a first aspect, an embodiment of the present application provides a video processing method, where the method includes: displaying the first video stream and the second video stream; receiving a first input of a user; responding to the first input, and carrying out image fusion on N first video frames and M second video frames selected by the first input to obtain a target video; the N first video frames are video frames in a first video stream, the M second video frames are video frames in a second video stream, and N, M are positive integers.
In a second aspect, an embodiment of the present application provides a video processing apparatus, including: the device comprises a display module, a receiving module and a processing module. The display module is used for displaying the first video stream and the second video stream. The receiving module is used for receiving a first input of a user. The processing module is used for responding to the first input received by the receiving module and carrying out image fusion on N first video frames and M second video frames selected by the first input to obtain a target video; the N first video frames are video frames in a first video stream, the M second video frames are video frames in a second video stream, and N, M are positive integers.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a processor, a memory, and a program or instructions stored in the memory and executable on the processor, and when executed by the processor, the program or instructions implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In the embodiment of the application, the electronic device may display the first video stream and the second video stream, and perform image fusion on N first video frames in the first video stream selected by the first input and M second video frames in the second video stream according to the first input of the user to obtain the target video. The electronic equipment can display the first video frame of the first video stream and the second video frame of the second video stream firstly, so that a user can input the first video frame and the second video frame once according to the requirement of the user, the electronic equipment can perform image fusion on the N first video frames and the M second video frames required by the user to obtain the target video with the special effect required by the user, namely, the user can customize the video frames required to perform image fusion, the electronic equipment can obtain the video with the double-exposure special effect required by the user, and the user does not need to perform multiple operations, so that the operation of the user in the process of shooting the video file satisfied by the user can be simplified, the consumed time is reduced, and the efficiency of shooting the video file by the electronic equipment can be improved.
Drawings
Fig. 1 is a schematic diagram of a video processing method provided in an embodiment of the present application;
fig. 2 is one of schematic diagrams of an example of an interface of a mobile phone according to an embodiment of the present disclosure;
fig. 3A is a second schematic diagram of an example of an interface of a mobile phone according to an embodiment of the present disclosure;
fig. 3B is a third schematic diagram of an example of an interface of a mobile phone according to the embodiment of the present application;
fig. 3C is a fourth schematic view of an example interface of a mobile phone according to an embodiment of the present application;
fig. 4A is a fifth schematic view of an example of an interface of a mobile phone according to an embodiment of the present application;
fig. 4B is a sixth schematic view of an example of an interface of a mobile phone according to an embodiment of the present application;
fig. 5 is a seventh schematic diagram of an example of an interface of a mobile phone according to an embodiment of the present disclosure;
fig. 6 is an eighth schematic diagram of an example of an interface of a mobile phone according to an embodiment of the present disclosure;
fig. 7 is a ninth schematic diagram of an example of an interface of a mobile phone according to an embodiment of the present application;
FIG. 8 is a tenth illustration of an example of an interface of a mobile phone provided by an embodiment of the present application;
fig. 9 is a second schematic diagram of a video processing method according to an embodiment of the present application;
fig. 10 is an eleventh schematic diagram of an example of an interface of a mobile phone according to an embodiment of the present application;
fig. 11 is a schematic structural diagram of a video processing apparatus according to an embodiment of the present application;
fig. 12 is a schematic structural diagram of an electronic device provided in an embodiment of the present application;
fig. 13 is a hardware schematic diagram of an electronic device provided in an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
The following describes in detail a video processing method, a video processing apparatus, and an electronic device according to embodiments of the present application with reference to the accompanying drawings.
The embodiment of the application can be applied to video shooting scenes.
Assuming that a user wants to trigger the electronic device to capture a video file with an overlapping (ghost) special effect (i.e., a double-exposure special effect), in the related art, the user may first trigger the electronic device to capture the video file, and then select a video material in a video processing application of the electronic device, and adjust capturing parameters of the video material, so that the electronic device may obtain the video file with the overlapping (ghost) special effect according to the adjusted video material and the video file through the video processing application, so as to obtain the video file that is satisfactory to the user. However, the user needs to perform multiple operations to trigger the electronic device to capture a video file that is satisfactory to the user, so the user operation is complicated.
However, in this embodiment of the present application, a user may first trigger the electronic device to display the video stream 1 and the video stream 2, so that the user may perform one input according to a user requirement, so that the electronic device may perform image fusion on a part of video frames in the video stream 1 and a part of video frames in the video stream 2 selected by the user to obtain a video file with an overlapping (ghost) special effect. It can be understood that the user can trigger the electronic device to shoot the video file satisfied by the user by one-time input, so that the operation of the user can be simplified.
Fig. 1 shows a flowchart of video processing provided by an embodiment of the present application. As shown in fig. 1, a video processing method provided in an embodiment of the present application may include steps 101 to 103 described below.
Step 101, the video processing device displays a first video stream and a second video stream.
Optionally, in this embodiment, the video processing apparatus may display the first video stream and the second video stream in the video processing interface.
Further optionally, in this embodiment of the application, in a case that the video processing apparatus displays a desktop, the video processing apparatus may open a shooting type application in the desktop and display a shooting interface according to a click input of a user to the shooting type application, open a "double exposure recording" mode according to a click input of the user to a "double exposure recording" control in the shooting interface, and acquire a first video stream and a second video stream according to an input of the user, so that the video processing apparatus may update the shooting interface to a video processing interface, where the video processing interface includes the first video stream and the second video stream.
Optionally, in this embodiment of the present application, the first video stream may be: the video processing device comprises a video processing device and a control device, wherein the video processing device comprises a camera, a video processing device and a control device, the camera of the video processing device is used for acquiring a video stream, or a video stream pre-stored in the video processing device, or a video stream downloaded by the video processing device from other equipment. Wherein the first video stream may comprise at least one video frame.
Optionally, in this embodiment of the application, the second video stream may be: the video processing device comprises a video processing device and a control device, wherein the video processing device comprises a camera, a video processing device and a control device, the camera of the video processing device is used for acquiring a video stream, or a video stream pre-stored in the video processing device, or a video stream downloaded by the video processing device from other equipment. Wherein the second video stream may comprise at least one video frame. It should be noted that the "video stream" in the present application may be understood as continuous video data collected by the camera.
Optionally, in this embodiment of the application, in a case that the first video stream and the second video stream are both video streams captured by a camera of the video processing apparatus, the first video stream and the second video stream may be video streams captured by the same camera (or different cameras).
The number of video frames in the second video stream may be the same as or different from the number of video frames in the first video stream; shooting parameters of the second video stream may be different from the shooting parameters of the first video stream; the image content of the video frames in the second video stream may be the same or different from the image content of the video frames in the first video stream.
Wherein, the shooting parameters may include at least one of the following: focus parameters, Auto Exposure (AE) parameters, filter parameters, beauty parameters, brightness parameters, frame rate parameters, and the like.
It is understood that the user can set different shooting parameters to achieve the video stream of the effect desired by the user, so that the desired fusion material (i.e., the first video stream and the second video stream) can be obtained.
Optionally, in this embodiment of the application, the video processing apparatus may display video frames in the first video stream and the second video stream respectively in the form of thumbnails, so as to display the first video stream and the second video stream.
It should be noted that "displaying video frames in the first video stream and the second video stream respectively in the form of thumbnails" as described above can be understood as follows: the video processing device may display a thumbnail of each video frame in the first video stream and a thumbnail of each video frame in the second video stream.
For example, a video processing apparatus is taken as a mobile phone for description. As shown in fig. 2, the mobile phone may display video frames in the first video stream and the second video stream respectively in the form of thumbnails in a video processing interface (e.g., interface 10), that is, the video processing apparatus displays thumbnails (e.g., reference numeral 11 to reference numeral 16) of each video frame in the first video stream (e.g., video stream one) and thumbnails (e.g., reference numeral 17 to reference numeral 21) of each video frame in the second video stream (e.g., video stream two).
The following description will take the first video stream and the second video stream as examples, where the first video stream and the second video stream are both video streams captured by a camera of a video processing apparatus.
Optionally, in this embodiment of the present application, before the step 101, the video processing method provided in this embodiment of the present application may further include the following step 201 or step 202.
Step 201, the video processing apparatus records a first video stream and a second video stream in sequence through a camera.
Optionally, in this embodiment of the application, when the video processing apparatus starts the "dual exposure recording" mode, the video processing apparatus may start the rear main camera according to the click input of the user to the "single main camera recording" control in the shooting interface, and sequentially acquire the first video stream and the second video stream through the rear main camera according to the click input of the user to the "shooting" control in the shooting interface, so as to sequentially record the first video stream and the second video stream through one camera (i.e., the rear main camera).
For example, as shown in fig. 3A, the mobile phone displays a shooting interface (for example, the interface 22), and the mobile phone starts a "dual exposure recording" mode, where the interface 22 includes a "single main shooting recording" control, so that the mobile phone can start a rear main camera according to a click input of a user to the "single main shooting recording" control, so that the user can adjust shooting parameters such as a focus parameter, an AE parameter, a filter parameter, a beauty parameter, a brightness parameter, and a frame rate parameter, and align the rear main camera with a person, and then click input the "shooting" control in the interface 22, so that the mobile phone can acquire a first video stream of the person through the rear main camera according to the adjusted shooting parameters; as shown in fig. 3B, after the mobile phone collects the first video stream of the person, the mobile phone may display a "save data stream" control, so that the mobile phone may store the first data stream according to the click input of the user to the "save data stream" control, and open the rear main camera, so that the user may adjust the shooting parameters again, align the rear main camera to the landscape, and then click-input the "shooting" control in the interface 22; as shown in fig. 3C, after the user clicks on the "shooting" control, the mobile phone may capture a second video stream of the landscape through the rear main camera according to the adjusted shooting parameters.
Step 202, the video processing apparatus records the first video stream and the second video stream in parallel through the two cameras.
Optionally, in this embodiment of the application, when the video processing apparatus starts the "dual exposure recording" mode, the video processing apparatus may start the front camera and the rear camera (or start two rear cameras) according to a click input of a user to the "front-and-back dual shooting recording" control (or the "rear-side dual shooting recording" control) in the shooting interface, and simultaneously acquire the first video stream and the second video stream through the front camera and the rear camera (or start two rear cameras) according to a click input of the user to the "shooting" control in the shooting interface, so as to concurrently record the first video stream and the second video stream through the two cameras.
For example, as shown in fig. 4A, a mobile phone displays a shooting interface (for example, an interface 23), where the interface 23 includes a "front-back double-shooting and recording" control, so that the mobile phone can start a front camera and a rear camera according to a click input of a user to the "front-back double-shooting and recording" control, so that the user can adjust shooting parameters such as a focus parameter, an AE parameter, a filter parameter, a beauty parameter, a brightness parameter, a frame rate parameter, and the like, align the front camera and the rear main camera with a person 1 and a person 2, respectively, and then click and input the "shooting" control in the interface 23, so that the mobile phone can collect a first video stream of the person 1 through the front camera and collect a second video stream of the person 2 through the rear camera, respectively; as shown in fig. 4B, after the user clicks on the "shoot" control, the mobile phone may capture a first video stream of person 1 through the front camera and capture a second video stream of person 2 through the rear camera, and display in the interface 23 a capture first window (e.g., window 24) and a second window (e.g., window 25), where the window 24 is used for displaying video frames of the first video stream of person 1, and the window 25 is used for displaying video frames of the second video stream of person 2.
Therefore, the video processing device can record the first video stream and the second video stream respectively through different recording modes to obtain the first video stream and the second video stream with different effects, so that the diversity of special effects of the shot target video can be improved.
Of course, in the case where the video processing apparatus displays the first video stream and the second video stream, the user may also preview video frames in the first video stream (or the second video stream), as will be exemplified below.
Optionally, in this embodiment of the present application, the step 101 may be specifically implemented by a step 101a described below, and after the step 101, the video processing method provided in this embodiment of the present application may further include a step 301 and a step 302 described below.
In step 101a, the video processing device displays video frames in the first video stream and the second video stream in the form of thumbnails in the video processing interface.
Further optionally, in this embodiment of the application, the video processing apparatus may display a thumbnail of each video frame of the first video stream in a first display area in the video processing interface, and display a thumbnail of each video frame of the second video stream in a second display area in the video processing interface. Wherein the first display area is adjacent to the second display area.
The above "display region adjacent to the first display region" may be understood as: and the display area is positioned in the preset range of the first display area.
For example, referring to fig. 3C, as shown in fig. 5, the mobile phone may update the interface 23 to a video processing interface (e.g., the interface 26), where a first display area (e.g., the display area 27) in the interface 26 displays a thumbnail of a video frame of a first video stream (e.g., the video stream one), a second display area (e.g., the display area 28) in the interface 26 displays a thumbnail of a video frame of a second video stream (e.g., the video stream two), and the display area 28 is: a display area adjacent to the display area 27.
In step 301, the video processing apparatus receives a third input of the first thumbnail from the user.
In this embodiment of the application, the first thumbnail is a thumbnail of any video frame in a target video stream, where the target video stream is: a first video stream or a second video stream.
It is understood that the target video frame is a video stream operated by a user.
In this embodiment, the third input is used to select a video frame for preview.
Further optionally, in this embodiment of the application, the third input may specifically be: the touch input of the user to the display screen, or the voice instruction input by the user, or the specific gesture input by the user, or other feasible input may be specifically determined according to the actual use requirement, and the embodiment of the present application is not limited.
The specific gesture can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input may be a single click input, a double click input, or any number of click inputs, and may also be a long-press input or a short-press input. For example, the third input may be: a user's single click input to a first thumbnail displayed on a display screen.
Step 302, the video processing apparatus displays, in response to the third input, a first video frame corresponding to the first thumbnail in the video processing interface.
Further optionally, in this embodiment of the application, the video processing apparatus may update the thumbnail in the video processing interface to be the first video frame corresponding to the first thumbnail, so as to display the first video frame corresponding to the first thumbnail.
For example, referring to fig. 5, after the mobile phone displays the thumbnail of the video frame of the first video stream and the thumbnail of the video frame of the second video stream in the interface 26 as shown in fig. 6, the mobile phone may update the thumbnail of the video frame of the first video stream and the thumbnail of the video frame of the second video stream to the first video frame (e.g., video frame 29) corresponding to the first thumbnail according to a third input (e.g., a click input) to the first thumbnail by the user.
Therefore, the video processing device can display the thumbnails of the video frames of the first video stream and the second video stream respectively, and display any video frame according to the input of the thumbnail of any video frame by the user, so that the user can quickly check whether any video frame is the video frame required by the user.
At step 102, the video processing apparatus receives a first input from a user.
In this embodiment of the application, the first input is used to select a video frame for video processing, and the video processing may include: fusing images; alternatively, image fusion and video frame insertion.
Optionally, in an embodiment of the present application, the first input specifically includes: the touch input of the user to the display screen, or the voice instruction input by the user, or the specific gesture input by the user, or other feasible input may be specifically determined according to the actual use requirement, and the embodiment of the present application is not limited.
The specific gesture can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input may be a single click input, a double click input, or any number of click inputs, and may also be a long-press input or a short-press input. For example, the first input may be: and clicking input of a control displayed on the display screen by a user.
Optionally, in this embodiment of the present application, the first input may include at least one input.
Wherein, under the condition that the video processing comprises image fusion and video frame insertion, one part of the at least one input is used for selecting the video frame for image fusion, and the other part of the at least one input is used for selecting the video frame for video frame insertion.
And 103, the video processing device responds to the first input, and carries out image fusion on the N first video frames and the M second video frames selected by the first input to obtain the target video.
In this embodiment of the application, the N first video frames are video frames in a first video stream, the M second video frames are video frames in a second video stream, and N, M are positive integers.
Optionally, in this embodiment of the present application, the number of N first video frames and the number of M second video frames may be the same or different, that is, N may be equal to M, or N may not be equal to M.
Optionally, in this embodiment of the application, when the video processing includes image fusion, the video processing apparatus may perform image fusion on the N first video frames and the M second video frames to obtain the target video.
Optionally, in this embodiment of the application, when N is equal to M, the video processing apparatus may perform image fusion on a first video frame and a first second video frame to obtain a first video frame, perform image fusion on a second first video frame and a second video frame to obtain a second video frame, and so on until performing image fusion on a last first video frame and a last second video frame to obtain a last video frame to obtain N video frames, so that the video processing apparatus may encapsulate the N video frames into a video file to obtain the target video.
Optionally, in this embodiment of the application, in the case that N is less than M, the video processing apparatus may perform image fusion on the first video frame and the first second video frame to obtain a first video frame, perform image fusion on the second first video frame and the second video frame to obtain a second video frame, and so on until the nth first video frame and the nth second video frame are image fused to obtain an nth video frame, so that the video processing apparatus may encapsulate the N video frames and the M-N second video frames into a video file to obtain the target video. Wherein the M-N second video frames are: all but the N second video frames of the M second video frames.
It should be noted that, in the case that N is greater than M, the video processing apparatus performs image fusion on the N first video frames and the M second video frames, which may refer to specific description of the case that N is less than M, and this embodiment of the present application is not described herein again.
Optionally, in a possible implementation manner of the embodiment of the present application, the video processing apparatus may adjust transparency values of the N first video frames and the M second video frames, and perform image fusion on the N first video frames and the M second video frames to obtain the target video. Wherein the video frames of the target video have an overlapping (ghosting) special effect.
Further optionally, in this embodiment of the application, the video processing apparatus may turn down (or turn up) transparency values of the N first video frames, and turn down (or turn up) transparency values of the M second video frames; alternatively, the video processing apparatus may increase (or decrease) the transparency values of the N first video frames and decrease (or increase) the transparency values of the M second video frames to adjust the transparency values of the N first video frames and the M second video frames to image-fuse the N first video frames and the M second video frames.
Exemplarily, assuming that the image contents of the first video stream and the second video stream each include a person, person 1 in the first video stream walks to the left, and person 2 in the second video stream walks to the right, the video processing apparatus may adjust the transparency values of N first video frames down, and adjust the transparency values of M second video frames down, and perform image fusion on the adjusted N first video frames and the adjusted M second video frames, so that a special effect through which two persons (i.e., person 1 and person 2) penetrate, that is, a special effect through which person 1 penetrates person 2, may be formed in the obtained target video.
Optionally, in another possible implementation manner of the embodiment of the present application, the video processing apparatus may determine a certain region of the N first video frames (or the M second video frames) as a foreground region, lower the image transparency of the certain region, and raise the image transparency of all sub-regions except the certain region, so that the video processing apparatus may perform image fusion on the N first video frames and the M second video frames to obtain the target video. Wherein all sub-regions of the video frames of the target video except the certain region have an overlapping (ghosting) special effect.
For example, assuming that the first video stream includes a person and the second video stream includes a landscape, the video processing apparatus may decrease the transparency of the image of the region where the person is located and increase the transparency of the image of all sub-regions except the region where the person is located (i.e., the region where the landscape is located), so that the video processing apparatus may perform image fusion on the N adjusted first video frames and the M second video frames, and thus the obtained target video may form a landscape with an overlapping (ghost) special effect except the person.
Optionally, in yet another possible implementation manner of the embodiment of the present application, the video processing apparatus may extract N sub-images from N first video frames and extract M sub-images from M second video frames, so that the video processing apparatus may perform image fusion on the N sub-images and the M sub-images to obtain the target video. The N sub-images are matched with the M sub-images in image content, and video frames of the target video have special effects of different image fusion.
The above "special effect of different screen fusion" can be understood as follows: the special effects of the objects in different time (or different places) of the same scenery appear in the pictures of the video. Wherein the object may be any one of: characters, objects, animals, plants, etc.
Optionally, in this embodiment of the application, in a case that the video processing includes image fusion and video frame insertion, the video processing apparatus may insert an inserted video frame (for example, an intermediate video frame in the following embodiment) into a video frame obtained by performing image fusion on the N first video frames and the M second video frames, so as to obtain the target video.
Optionally, in this embodiment of the application, after the video processing apparatus obtains the target video, the video processing apparatus may update the video processing interface, and display a third window in the updated interface, where the third window is used to display a video frame of the target video, so that a user may view the target video in the third window, and may input a "save" control in the updated interface, so that the video processing apparatus may store the target video.
For example, the mobile phone may perform image fusion on the N video frames and the M video frames to obtain the target video. Referring to fig. 6, as shown in fig. 7, after the target video is obtained by the mobile phone, the mobile phone may update the interface 26 to the interface 30, and display a third window (for example, a window 31) in the interface 30, where the window 31 is used to display a video frame of the target video, so that the user may view the target video in the window 31.
In the embodiment of the application, if a user wants to add an overlapping (ghost) special effect to a video file, the user can directly trigger the video processing device to acquire the first video stream and the second video stream and perform one-time input, so that the video processing device can perform image fusion on a part of video frames in the first video stream and a part of video frames in the second video stream selected by the one-time input to obtain a target video with the overlapping (ghost) special effect.
The user can also trigger the video processing device to acquire the first video stream and the second video stream with different shooting parameters, so that the target video obtained by the video processing device has an overlapping (ghost) special effect (namely, a double-exposure special effect).
The user can also trigger the video processing device to acquire a first video stream and a second video stream with different shooting parameters and the same image content of the video frames, so that the target video obtained by the video processing device has a high-dynamic range (HDR) effect and a higher frame rate than the first video stream (or the second video stream).
According to the video processing method provided by the embodiment of the application, the video processing device can display the first video stream and the second video stream, and perform image fusion on N first video frames in the first video stream and M second video frames in the second video stream selected by first input according to the first input of a user to obtain the target video. The video processing device can display the first video frame of the first video stream and the second video frame of the second video stream firstly, so that a user can input the video frames once according to the requirement of the user, the video processing device can perform image fusion on the N first video frames and the M second video frames required by the user to obtain the target video with the special effect required by the user, namely the user can customize the video frames required to perform image fusion, so that the video processing device can obtain the video with the double-exposure special effect required by the user without performing multiple operations by the user, the operation of the user in the process of shooting the video file satisfied by the user can be simplified, the consumed time is reduced, and the efficiency of shooting the video file by the video processing device can be improved.
Moreover, the video processing device can always store the video streams in the video recording process, respectively store the two video streams in the two video queues for double exposure post-processing, is suitable for multiple scenes and multiple space-time spaces, and has more creativity and playability.
How the video processing apparatus performs image fusion will be illustrated below with different examples.
Optionally, in a possible implementation manner of the embodiment of the present application, before the step 103, the video processing method provided in the embodiment of the present application may further include the following steps 401 to 403.
Step 401, the video processing apparatus divides the third video frame into at least two sub-areas.
In this embodiment of the application, the third video frame is any one of N first video frames or M second video frames.
Further optionally, in this embodiment of the application, in a case where the video processing apparatus displays the third video frame, the video processing apparatus may divide the third video frame into at least two sub-areas.
It should be noted that, for the description of the video processing apparatus displaying the third video frame, reference may be made to specific description of the video processing apparatus displaying the first video frame, and details of the embodiment of the present application are not repeated herein.
Further optionally, in this embodiment of the application, the video processing apparatus may perform image recognition on a third video frame, and divide the third video frame into at least two sub-regions according to a video object in the third video frame obtained by the image recognition; alternatively, the third video frame may be divided into at least two sub-areas according to a user input.
Optionally, in this embodiment of the present application, the step 401 may be specifically implemented by the following step 401a or step 401 b.
Step 401a, the video processing apparatus divides the third video frame into at least two sub-regions according to the object type of the video object in the third video frame.
Further optionally, in this embodiment of the application, the object type may include any one of the following: static objects or dynamic objects.
It should be noted that the above "static object" can be understood as: objects that are stationary in the video stream, such as buildings, roads, etc. The above-mentioned "dynamic object" can be understood as: objects moving in the video stream, such as pedestrians, vehicles in motion, etc.
Further optionally, in this embodiment of the application, the video processing apparatus may divide a region in which video objects with the same object type in the third video frame are located into one sub-region, so as to obtain at least two sub-regions.
It is understood that the at least two sub-regions comprise: the area of the static object and the area of the dynamic object.
In the embodiment of the present application, since the image fusion modes of the objects (i.e., the static object, the dynamic object, and the like) that may be required by the user are different, the video processing apparatus may divide the third video frame into the region where the static object is located and the region where the dynamic object is located, so that the user may set the image fusion modes of different sub-regions.
Therefore, the video processing device can directly divide the third video frame into the area where the static object is located and the area where the dynamic object is located, so that a user can set the image fusion modes of different sub-areas according to requirements, and therefore, the flexibility of setting the image fusion mode of the video frame can be improved.
Step 401b, in the case that the video processing apparatus receives a sliding input of the third video frame from the user, the video processing apparatus divides the third video frame into at least two sub-areas based on the sliding track of the sliding input.
Further optionally, in this embodiment of the application, the sliding track may be any one of the following: closed tracks, unclosed tracks.
Further optionally, in this embodiment of the application, in a case that the sliding track is a closed track, the video processing apparatus may determine the areas enclosed by the closed track as at least two sub-areas respectively.
Specifically, the video processing apparatus may first determine the area enclosed by the closed track as a first sub-area to determine at least two first sub-areas, and then identify the edge lines of the video object in the at least two first sub-areas by an Artificial Intelligence (AI) algorithm to determine the at least two sub-areas according to the edge lines of the video object.
For example, referring to fig. 6, as shown in fig. 8, the user may perform a sliding input on the video frame 29, where the sliding track of the sliding input is the track 32, so that the mobile phone may determine an area (e.g., the area 33) surrounded by the input track 32 as one sub-area to obtain at least two sub-areas, namely the area 33 and the area 34, where the area 34 is all the areas except the area 33.
Further optionally, in this embodiment of the application, in a case that the sliding track is an unclosed track, the video processing apparatus may determine, as one sub-area, an area within a preset range of the unclosed track, so as to determine at least two sub-areas.
Further optionally, in this embodiment of the application, after the video processing apparatus determines the at least two sub-areas, the video processing apparatus may mark the at least two sub-areas in a first marking manner, so that the user may view the at least two sub-areas.
Specifically, the first marking manner may include at least one of: a dotted frame marking mode, a highlight marking mode, a color marking mode, a gray marking mode, a preset transparency marking mode, a flashing marking mode and the like.
In the embodiment of the application, since a situation that a user wants to set an image fusion mode of a sub-region in a third video frame may occur, the user may perform a slide input on the third video frame, so that the video processing apparatus may divide the third video frame into at least two sub-regions based on a slide track of the slide input, and thus the user may set an image fusion mode of different sub-regions.
Therefore, the video processing device can divide the third video frame into at least two sub-areas according to the sliding input of the user on the third video frame, so that the user can set the image fusion modes of different sub-areas according to the requirements, and the flexibility of setting the image fusion mode of the video frame can be improved.
Step 402, the video processing apparatus receives a second input from the user to a first sub-area of the at least two sub-areas.
In an embodiment of the present application, the second input is used to set the image transparency of the sub-region.
Further optionally, in this embodiment of the application, the second input may specifically be: the touch input of the user to the display screen, or the voice instruction input by the user, or the specific gesture input by the user, or other feasible input may be specifically determined according to the actual use requirement, and the embodiment of the present application is not limited.
The specific gesture can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure recognition gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input may be a single click input, a double click input, or any number of click inputs, and may also be a long-press input or a short-press input. For example, the second input may be: and clicking input of the first sub-area displayed on the display screen by the user.
In step 403, the video processing apparatus responds to the second input, and reduces the image transparency of the first sub-area and increases the image transparency of the second sub-area.
In an embodiment of the present application, the second sub-region includes all sub-regions except the first sub-region in the at least two sub-regions.
In this embodiment, the video processing apparatus may divide the third video frame into at least two sub-regions, so that a user may input a first sub-region of the at least two sub-regions, so that the video processing apparatus may determine the first sub-region as a foreground region, and the video processing apparatus may perform image fusion on the N first video frames and the M second video frames to obtain the target video. Wherein all sub-regions of the video frame of the target video except the first sub-region have overlapping (ghosting) special effects.
Therefore, the video processing device can divide any one of the N first video frames or the M second video frames into at least two sub-regions, so that a user can perform one-time input, so that the video processing device can reduce the image transparency of the first sub-region required by the user and increase the image transparency of the second sub-region to obtain the target video with the special effect required by the user, thereby simplifying the user operation in the process of shooting the video file satisfactory to the user and reducing the time consumption.
Optionally, in another possible implementation manner of the embodiment of the present application, before the step 103, the video processing method provided in the embodiment of the present application may further include the following step 501 and step 502, and the step 103 may be specifically implemented by the following step 103 a.
Step 501, the video processing apparatus scratches an ith first sub-image from an ith first video frame of the N first video frames.
In the embodiment of the application, i is a positive integer.
Step 502, the video processing apparatus scratches an ith second sub-image from an ith second video frame of the M second video frames.
In the embodiment of the present application, the i second sub-images are matched with the image content of the i first sub-image.
Further optionally, in this embodiment of the application, the video processing apparatus may first perform image recognition on the ith first video frame to obtain first image content, and perform image recognition on the ith second video frame to obtain second image content, so that the video processing apparatus may respectively extract the ith first sub-image and the ith second sub-image according to the first image content and the second image content.
It should be noted that the above "matching" can be understood as: the same, or the difference between the two is less than or equal to the preset threshold.
And 103a, the video processing device performs image fusion on the ith first sub-image and the ith second sub-image to obtain a target video.
It is understood that, for each of the N first video frames and each of the M second video frames, the video processing apparatus may perform the above steps 501, 502, and 103a, respectively, so as to obtain the target video.
In the embodiment of the application, the video processing device can identify the picture scenes of the same video object (such as an object, a building, a person and the like) of two video frames, cut out the regions where the two video frames do not overlap, and reserve the regions where the two video frames overlap (i.e. the first sub-image and the second sub-image), and then fuse the regions where the two video frames overlap to obtain the target video, wherein in the video frames of the target video, the pictures are unchanged after the scenes with the same pictures are fused, and the scenes with different pictures have special overlapping effects.
Therefore, the video processing device can extract the ith first sub-image and the ith second sub-image which are matched with the image content in the ith first video frame and the ith second video frame, and perform image fusion on the ith first sub-image and the ith second sub-image, so that the diversity of special effects of the shot target video can be improved.
The following description will be given by taking video processing including image fusion and video frame insertion as an example.
It will be appreciated that a portion of the input of the first input is used to select video frames for image fusion and another portion is used to select video frames for video frame insertion.
Optionally, in this embodiment of the application, the first input further includes an input of an intermediate video frame in the first video stream or in the second video stream by a user. Specifically, the step 103 can be realized by the step 103b described below.
And 103b, responding to the first input by the video processing device, carrying out image fusion on all video frames except the intermediate video frame in the N first video frames and the M second video frames, and inserting the intermediate video frame into the video frame obtained by image fusion to obtain the target video.
It will be appreciated that the intermediate video frame may be a video frame selected for video frame insertion for another portion of the first input.
Further optionally, in this embodiment of the application, after the video processing apparatus inserts the intermediate video frame into the video frame obtained by image fusion, the video processing apparatus may further add a special effect material to the intermediate video frame according to an input of a user.
In particular, the special effects material may include at least one of: optical special effect materials, sticker materials and the like.
Further optionally, in this embodiment of the application, the video processing apparatus may insert the intermediate video frame in front of (or behind) any one of the video frames obtained by image fusion, so that the video processing apparatus may encapsulate the inserted video frame into a video file to obtain the target video. It is understood that the intermediate video frame is a video frame (or a subsequent video frame) that precedes the any video frame.
Exemplarily, assuming that the N first video frames include 30 first video frames, the M second video frames include 31 second video frames, the intermediate video frame is a first one of the M second video frames, the video frame processing device may perform image fusion on the first one of the 30 first video frames and a second one of the 31 second video frames to obtain a first fused video frame, perform image fusion on the second one of the 30 first video frames and a third one of the 31 second video frames to obtain a second fused video frame, and so on to obtain 30 fused video frames, so that the video processing device may insert the intermediate video frame into the 30 fused video frames, for example, into the front of a fifteenth one of the 30 fused video frames to obtain the target video, i.e. the intermediate video frame is the fifteenth video frame of the target video.
In the embodiment of the present application, since a situation may occur in which a user wants to display some video frames in the first video stream (or the second video stream) separately (i.e., without performing image fusion), the video processing apparatus may perform image fusion on all video frames except the intermediate video frame in the N first video frames and the M second video frames, and insert the intermediate video frame into the video frame obtained by image fusion to obtain the target video.
Therefore, the video processing device can perform image fusion on a part of the N first video frames and a part of the M second video frames, and insert the other part of the N first video frames and the M second video frames into the video frames obtained by image fusion, that is, the video processing device can process the video frames by adopting different video processing modes, so that the diversity of special effects of the shot target video can be improved.
In this embodiment, the video processing interface further includes at least one control, and each control is respectively used to select a video frame in one video stream, so that a user can input the at least one control, so that the video processing apparatus can select the N first video frames and the M second video frames.
Optionally, in this embodiment of the application, the video processing interface further includes a first control and a second control; the first control is used to select a video frame in the first video stream and the second control is used to select a video frame in the second video stream. Specifically, with reference to fig. 1, as shown in fig. 9, the step 102 may be specifically implemented by a step 102a described below, and before the step 103 of "performing image fusion on the N first video frames and the M second video frames selected by the first input", the video processing method provided in the embodiment of the present application may further include steps 601 to 603 described below, and the step 103 may be specifically implemented by a step 103c described below.
Step 102a, the video processing apparatus receives a first input of a user to the first control and the second control.
Further optionally, in this embodiment of the application, the first input may include a first sub-input and a second sub-input. The first sub-input is input of a first control by a user, and the first sub-input is used for selecting a video frame for image fusion in a first video stream; the second sub-input is input to the second control by the user, and the second sub-input is used for selecting a video frame for image fusion in the second video stream.
Specifically, the first sub-input may be specifically a drag input of the user dragging the first control to the first position. The second sub-input may specifically be a drag input in which the user drags the second control to the second position.
Step 601, the video processing apparatus determines a first start video frame and a first end video frame in the first video stream and a second start video frame and a second end video frame in the second video stream according to a first input in response to the first input.
Further alternatively, in this embodiment of the application, the video processing apparatus may determine the first start video frame and the first end video frame respectively according to the first position, and determine the second start video frame and the second end video frame respectively according to the second position.
Specifically, the video processing apparatus may determine the video frames corresponding to the thumbnails located at the first positions as the first start video frame and the first end video frame, respectively, and determine the video frames corresponding to the thumbnails located at the second positions as the second start video frame and the second end video frame, respectively.
For example, referring to fig. 5, as shown in fig. 10, the interface 26 of the mobile phone further includes a first control 35 and a second control 36, so that the user can make a first sub-input to the first control 35, so that the mobile phone can determine the first start video frame 37 and the first end video frame 38 according to the first position, respectively, and make a second sub-input to the second control 36, so that the mobile phone can determine the second start video frame 39 and the second end video frame 40 according to the second position, respectively.
In step 602, the video processing apparatus determines N first video frames between a first start video frame and a first end video frame from the first video stream.
It is understood that the N first video frames include a first start video frame, a first end video frame, and all first video frames between the first start video frame and the first end video frame.
Step 603, the video processing apparatus determines M second video frames between the second start video frame and the second end video frame from the second video stream.
It is understood that the M second video frames include a second start video frame, a second end video frame, and all second video frames between the second start video frame and the second end video frame.
Note that, the execution order of the step 602 and the step 603 is not limited in the embodiment of the present application. In a possible implementation manner, the video processing apparatus may perform step 602 first, and then perform step 603; in another possible implementation manner, the video processing apparatus may perform step 603 first, and then perform step 603; in yet another possible implementation, the video processing apparatus may perform step 603 while performing step 602.
And 103c, the video processing device performs image fusion on the N first video frames and the M second video frames to obtain the target video.
Therefore, the user can input the first control and the second control, so that the video processing device can directly determine the N first video frames and the M second video frames without inputting the video processing device for multiple times, the operation of the user in the process of shooting the video file satisfactory to the user can be simplified, and the time consumption is reduced.
It should be noted that, for the description of selecting the intermediate video frame by the video processing apparatus, reference may be made to specific description of selecting N first video frames and M second video frames by the video processing apparatus, and details of this embodiment are not repeated herein.
It should be noted that, in the video processing method provided in the embodiment of the present application, the execution subject may be a video processing apparatus, or a control module in the video processing apparatus for executing the video processing method. In the embodiment of the present application, a video processing apparatus executing a video processing method is taken as an example to describe the video processing apparatus provided in the embodiment of the present application.
Fig. 11 shows a schematic diagram of a possible structure of a video processing apparatus according to an embodiment of the present application. As shown in fig. 11, the video processing apparatus 60 may include: and a display module 61, configured to display the first video stream and the second video stream. The receiving module 62 is configured to receive a first input from a user. The processing module 63 is configured to synthesize, in response to the first input received by the receiving module 62, the N first video frames and the M second video frames selected by the first input to obtain a target video; the N first video frames are video frames in a first video stream, the M second video frames are video frames in a second video stream, and N, M are positive integers.
In a possible implementation manner, the processing module 63 is further configured to divide a third video frame into at least two sub-areas, where the third video frame is any one of N first video frames or M second video frames. The receiving module 62 is further configured to receive a second input from the user to the first sub-area of the at least two sub-areas divided by the processing module 63. The processing module 63 is further configured to, in response to the second input received by the receiving module 62, lower the image transparency of the first sub-area and raise the image transparency of the second sub-area; wherein the second sub-area includes all sub-areas except the first sub-area of the at least two sub-areas.
In a possible implementation manner, the processing module 63 is specifically configured to divide the third video frame into at least two sub-regions according to an object type of a video object in the third video frame; or, in the case that a sliding input of the user to the third video frame is received, the third video frame is divided into at least two sub-areas based on a sliding track of the sliding input.
In a possible implementation manner, the video processing interface further includes a first control and a second control; the first control is used to select a video frame in the first video stream and the second control is used to select a video frame in the second video stream. The receiving module 62 is specifically configured to receive a first input of the first control and the second control from the user. The processing module 63 is further configured to determine, according to the first input received by the receiving module 62, a first start video frame and a first end video frame in the first video stream, and a second start video frame and a second end video frame in the second video stream; determining N first video frames between a first starting video frame and a first ending video frame from the first video stream; and determining M second video frames between the second start video frame and the second end video frame from the second video stream.
In one possible implementation, the first input includes a user input of an intermediate video frame in the first video stream or in the second video stream. The processing module 63 is specifically configured to perform image fusion on all video frames, except for the intermediate video frame, in the N first video frames and the M second video frames, and insert the intermediate video frame into the video frame obtained by image fusion.
In a possible implementation manner, the processing module 63 is further configured to scratch an ith first sub-image from an ith first video frame of the N first video frames, where i is a positive integer; and keying an ith second sub-image from an ith second video frame of the M second video frames, wherein the ith second sub-image is matched with the image content of the ith first sub-image. The processing module 63 is specifically configured to perform image fusion on the ith first sub-image and the ith second sub-image to obtain the target video.
In a possible implementation manner, the processing module 63 is further configured to record a first video stream and a second video stream sequentially through a camera; or, the first video stream and the second video stream are recorded in parallel through two cameras.
In a possible implementation manner, the display module 61 is specifically configured to display, in a video processing interface, video frames in the first video stream and the second video stream respectively in a thumbnail mode. The receiving module 62 is further configured to receive a third input of the first thumbnail from the user, where the first thumbnail is a thumbnail of any video frame in the target video stream, and the target video stream is: a first video stream or a second video stream. The display module 63 is further configured to display, in response to the third input received by the receiving module 62, a first video frame corresponding to the first thumbnail in the video processing interface.
According to the video processing device provided by the embodiment of the application, the video processing device can display the first video frame of the first video stream and the second video frame of the second video stream firstly, so that a user can input once according to the requirement of the user, the video processing device can perform image fusion on the N first video frames and the M second video frames required by the user to obtain the target video with the special effect required by the user, namely, the user can customize the video frames required to perform image fusion, the video processing device can obtain the video with the double exposure special effect required by the user, and the user does not need to perform multiple operations, so that the operation of the user in the process of shooting the video file satisfied by the user can be simplified, the time consumption is reduced, and therefore, the efficiency of shooting the video file by the video processing device can be improved.
The video processing apparatus in the embodiment of the present application may be an apparatus, or may be a component, an integrated circuit, or a chip in a terminal. The device can be mobile electronic equipment or non-mobile electronic equipment. By way of example, the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palm top computer, a vehicle-mounted electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and the non-mobile electronic device may be a server, a Network Attached Storage (NAS), a Personal Computer (PC), a Television (TV), a teller machine or a self-service machine, and the like, and the embodiment of the present application is not particularly limited.
The video processing apparatus in the embodiment of the present application may be an apparatus having an operating system. The operating system may be an Android operating system (Android), an iOS operating system, or other possible operating systems, which is not specifically limited in the embodiments of the present application.
The video processing apparatus provided in the embodiment of the present application can implement each process implemented by the method embodiments in fig. 1 to fig. 10, and is not described herein again to avoid repetition.
Optionally, in this embodiment, as shown in fig. 12, an electronic device 70 is further provided in this embodiment of the present application, and includes a processor 71, a memory 72, and a program or an instruction stored in the memory 72 and capable of being executed on the processor 71, where the program or the instruction is executed by the processor 71 to implement each process of the above-mentioned video processing method embodiment, and can achieve the same technical effect, and no further description is provided here to avoid repetition.
It should be noted that the electronic device in the embodiment of the present application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 13 is a schematic hardware structure diagram of an electronic device implementing an embodiment of the present application.
The electronic device 100 includes, but is not limited to: a radio frequency unit 101, a network module 102, an audio output unit 103, an input unit 104, a sensor 105, a display unit 106, a user input unit 107, an interface unit 108, a memory 109, and a processor 110.
Those skilled in the art will appreciate that the electronic device 100 may further comprise a power source (e.g., a battery) for supplying power to various components, and the power source may be logically connected to the processor 110 through a power management system, so as to implement functions of managing charging, discharging, and power consumption through the power management system. The electronic device structure shown in fig. 13 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description is omitted here.
The display unit 106 is configured to display the first video stream and the second video stream.
A user input unit 107 for receiving a first input of a user.
And the processor 110 is configured to respond to the first input, and synthesize the N first video frames and the M second video frames selected by the first input to obtain the target video.
The N first video frames are video frames in a first video stream, the M second video frames are video frames in a second video stream, and N, M are positive integers.
According to the electronic equipment provided by the embodiment of the application, the first video frame of the first video stream and the second video frame of the second video stream can be displayed by the electronic equipment firstly, so that a user can input the first video frame and the second video frame according to the requirement of the user once, the electronic equipment can perform image fusion on the N first video frames and the M second video frames required by the user to obtain the target video with the special effect required by the user, namely the user can customize the video frames required to perform image fusion, the electronic equipment can obtain the video with the double-exposure special effect required by the user, and the user does not need to perform multiple operations, so that the operation of the user in the process of shooting the video file satisfied by the user can be simplified, the time consumption is reduced, and therefore, the efficiency of shooting the video file by the electronic equipment can be improved.
Optionally, in this embodiment of the application, the processor 110 is further configured to divide a third video frame into at least two sub-areas, where the third video frame is any one of the N first video frames or the M second video frames.
The user input unit 107 is further configured to receive a second input from the user to the first sub-area of the at least two sub-areas.
The processor 110 is further configured to adjust the transparency of the image of the first sub-area down and adjust the transparency of the image of the second sub-area up in response to the second input.
Wherein the second sub-area includes all sub-areas except the first sub-area of the at least two sub-areas.
Therefore, the electronic equipment can divide any one of the N first video frames or the M second video frames into at least two sub-regions, so that a user can perform one-time input, the image transparency of the first sub-region required by the user can be reduced, the image transparency of the second sub-region can be increased, and the target video with the special effect required by the user can be obtained.
Optionally, in this embodiment of the application, the processor 110 is specifically configured to divide the third video frame into at least two sub-regions according to an object type of a video object in the third video frame; or, in the case that a sliding input of the user to the third video frame is received, the third video frame is divided into at least two sub-areas based on a sliding track of the sliding input.
Therefore, the electronic device can directly divide the third video frame into the region where the static object is located and the region where the dynamic object is located, so that the user can set the image fusion modes of different sub-regions according to the requirements, and therefore, the flexibility of setting the image fusion mode of the video frame can be improved.
Therefore, the electronic equipment can divide the third video frame into at least two sub-regions according to the sliding input of the user on the third video frame, so that the user can set the image fusion modes of different sub-regions according to the requirements, and the flexibility of setting the image fusion mode of the video frame can be improved.
Optionally, in this embodiment of the application, the video processing interface further includes a first control and a second control; the first control is used to select a video frame in the first video stream and the second control is used to select a video frame in the second video stream.
The user input unit 107 is specifically configured to receive a first input of the first control and the second control from a user.
A processor 110, further configured to determine, according to a first input, a first starting video frame and a first ending video frame in the first video stream and a second starting video frame and a second ending video frame in the second video stream; determining N first video frames between a first starting video frame and a first ending video frame from the first video stream; and determining M second video frames between the second start video frame and the second end video frame from the second video stream.
Therefore, the user can input the first control and the second control, so that the electronic equipment can directly determine the N first video frames and the M second video frames without inputting the video files for multiple times by the user, the operation of the user in the process of shooting the video files satisfactory to the user can be simplified, and the time consumption is reduced.
Optionally, in this embodiment of the present application, the first input includes an input of an intermediate video frame in the first video stream or in the second video stream by a user.
The processor 110 is specifically configured to perform image fusion on all video frames except the intermediate video frame in the N first video frames and the M second video frames, and insert the intermediate video frame into the video frame obtained by image fusion.
Therefore, the electronic device can perform image fusion on a part of the N first video frames and a part of the M second video frames, and insert the other part of the N first video frames and the M second video frames into the video frames obtained by image fusion, that is, the video frames can be processed in different video processing manners, so that the diversity of special effects of the shot target video can be improved.
Optionally, in this embodiment of the present application, the processor 110 is further configured to scratch an ith first sub-image from an ith first video frame of the N first video frames, where i is a positive integer; and keying an ith second sub-image from an ith second video frame of the M second video frames, wherein the ith second sub-image is matched with the image content of the ith first sub-image.
The processor 110 is specifically configured to perform image fusion on the ith first sub-image and the ith second sub-image to obtain the target video.
Therefore, the electronic equipment can scratch the ith first sub-image and the ith second sub-image which are matched with each other in the ith first video frame and the ith second video frame and perform image fusion on the ith first sub-image and the ith second sub-image, so that the diversity of special effects of the shot target video can be improved.
Optionally, in this embodiment of the application, the processor 110 is further configured to record the first video stream and the second video stream sequentially through one camera; or, the first video stream and the second video stream are recorded in parallel through two cameras.
Therefore, the electronic equipment can record the first video stream and the second video stream respectively through different recording modes to obtain the first video stream and the second video stream with different effects, and therefore the diversity of special effects of the shot target video can be improved.
Optionally, in this embodiment of the present application, the display unit 106 is specifically configured to display, in a video processing interface, video frames in the first video stream and the second video stream in a thumbnail mode, respectively.
The user input unit 107 is further configured to receive a third input of a first thumbnail from a user, where the first thumbnail is a thumbnail of any video frame in a target video stream, and the target video stream is: a first video stream or a second video stream.
And the display unit 106 is further configured to display, in response to a third input, a first video frame corresponding to the first thumbnail in the video processing interface.
Therefore, the electronic equipment can display the thumbnails of the video frames of the first video stream and the second video stream respectively, and display any video frame according to the input of the user to the thumbnail of the video frame, so that the user can quickly check whether the video frame is the video frame required by the user.
It should be understood that, in the embodiment of the present application, the input unit 104 may include a Graphics Processing Unit (GPU) 1041 and a microphone 1042, and the graphics processing unit 1041 processes image data of a still picture or a video obtained by an image capturing device (e.g., a camera) in a video capturing mode or an image capturing mode. The display unit 106 may include a display panel 1061, and the display panel 1061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 107 includes a touch panel 1071 and other input devices 1072. The touch panel 1071 is also referred to as a touch screen. The touch panel 1071 may include two parts of a touch detection device and a touch controller. Other input devices 1072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein. The memory 109 may be used to store software programs as well as various data including, but not limited to, application programs and an operating system. The processor 110 may integrate an application processor, which primarily handles operating systems, user interfaces, applications, etc., and a modem processor, which primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 110.
The embodiments of the present application further provide a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the video processing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a read-only memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and the like.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the above video processing method embodiment, and can achieve the same technical effect, and the details are not repeated here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element identified by the phrase "comprising an … …" does not exclude the presence of other identical elements in the process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the description of the foregoing embodiments, it is clear to those skilled in the art that the method of the foregoing embodiments may be implemented by software plus a necessary general hardware platform, and certainly may also be implemented by hardware, but in many cases, the former is a better implementation. Based on such understanding, the technical solutions of the present application or portions thereof that contribute to the prior art may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (which may be a mobile phone, a computer, a server, or a network device, etc.) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A method of video processing, the method comprising:
displaying the first video stream and the second video stream;
receiving a first input of a user;
responding to the first input, and carrying out image fusion on N first video frames and M second video frames selected by the first input to obtain a target video;
wherein the N first video frames are video frames in the first video stream, the M second video frames are video frames in the second video stream, and N, M are positive integers.
2. The method according to claim 1, wherein before the image fusion of the N first video frames and the M second video frames selected by the first input to obtain the target video, the method further comprises:
dividing a third video frame into at least two sub-regions, wherein the third video frame is any one of the N first video frames or the M second video frames;
receiving a second input of a user to a first sub-area of the at least two sub-areas;
in response to the second input, reducing the image transparency of the first sub-area and increasing the image transparency of the second sub-area;
wherein the second sub-area includes all of the at least two sub-areas except the first sub-area.
3. The method of claim 2, wherein the dividing the third video frame into at least two sub-regions comprises:
dividing the third video frame into the at least two sub-regions according to the object type of the video object in the third video frame;
or, in the case that a sliding input of the third video frame by a user is received, dividing the third video frame into the at least two sub-areas based on a sliding track of the sliding input.
4. The method of claim 1, wherein the video processing interface further comprises a first control and a second control; the first control is used for selecting a video frame in the first video stream, and the second control is used for selecting a video frame in the second video stream;
the receiving a first input of a user comprises:
receiving a first input of a user to the first control and the second control;
before the image fusion is performed on the N first video frames and the M second video frames selected by the first input, the method further includes:
determining, from the first input, a first starting video frame and a first ending video frame in the first video stream and a second starting video frame and a second ending video frame in the second video stream;
determining the N first video frames between the first start video frame and the first end video frame from the first video stream;
determining, from the second video stream, the M second video frames between the second start video frame and the second end video frame.
5. The method of claim 1, wherein the first input comprises a user input of an intermediate video frame in the first video stream or in a second video stream;
the image fusion of the N first video frames and the M second video frames selected by the first input to obtain the target video includes:
and performing image fusion on all the video frames except the intermediate video frame in the N first video frames and the M second video frames, and inserting the intermediate video frame into the video frame obtained by image fusion.
6. The method according to claim 1, wherein prior to said image fusing the N first video frames and the M second video frames selected by the first input, the method further comprises:
scratching an ith first sub-image from an ith first video frame in the N first video frames, wherein i is a positive integer;
matting an ith second sub-image from an ith second video frame of the M second video frames, the ith second sub-image matching the image content of the ith first sub-image;
the image fusion of the N first video frames and the M second video frames selected by the first input to obtain the target video includes:
and carrying out image fusion on the ith first sub-image and the ith second sub-image to obtain the target video.
7. The method of claim 1, wherein prior to said displaying the first video stream and the second video stream, the method further comprises:
sequentially recording the first video stream and the second video stream through a camera;
or the first video stream and the second video stream are recorded in parallel through two cameras.
8. The method of claim 1, wherein displaying the first video stream and the second video stream comprises:
displaying video frames in the first video stream and the second video stream respectively in a video processing interface in a thumbnail mode;
after the displaying the first video stream and the second video stream, the method further comprises:
receiving a third input of a user to a first thumbnail, wherein the first thumbnail is a thumbnail of any video frame in a target video stream, and the target video stream is: the first video stream or the second video stream;
and responding to the third input, and displaying a first video frame corresponding to the first thumbnail in the video processing interface.
9. A video processing apparatus, characterized in that the video processing apparatus comprises:
a display module for displaying the first video stream and the second video stream;
the receiving module is used for receiving a first input of a user;
the processing module is used for responding to the first input received by the receiving module, and carrying out image fusion on N first video frames and M second video frames selected by the first input to obtain a target video;
wherein the N first video frames are video frames in the first video stream, the M second video frames are video frames in the second video stream, and N, M are positive integers.
10. An electronic device comprising a processor, a memory, and a program or instructions stored on the memory and executable on the processor, the program or instructions when executed by the processor implementing the steps of the video processing method according to any one of claims 1 to 8.
CN202210109250.3A 2022-01-28 2022-01-28 Video processing method and device and electronic equipment Active CN114520875B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210109250.3A CN114520875B (en) 2022-01-28 2022-01-28 Video processing method and device and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210109250.3A CN114520875B (en) 2022-01-28 2022-01-28 Video processing method and device and electronic equipment

Publications (2)

Publication Number Publication Date
CN114520875A true CN114520875A (en) 2022-05-20
CN114520875B CN114520875B (en) 2024-04-02

Family

ID=81596623

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210109250.3A Active CN114520875B (en) 2022-01-28 2022-01-28 Video processing method and device and electronic equipment

Country Status (1)

Country Link
CN (1) CN114520875B (en)

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005244536A (en) * 2004-02-26 2005-09-08 Seiko Epson Corp Image composition for generating composite image by overlapping image
US20110242130A1 (en) * 2010-03-30 2011-10-06 Casio Computer Co., Ltd. Image processing apparatus, image processing method, and computer-readable medium
US20160035116A1 (en) * 2013-03-29 2016-02-04 Rakuten, Inc. Image processing device, image processing method, information storage medium, and program
US20180070045A1 (en) * 2016-09-02 2018-03-08 Russell Holmes Systems And Methods For Providing Real-Time Composite Video From Multiple Source Devices
JP2018198463A (en) * 2018-09-11 2018-12-13 日本電信電話株式会社 Image processing device, image processing method, and computer program
CN109104588A (en) * 2018-07-24 2018-12-28 房梦琦 A kind of video monitoring method, equipment, terminal and computer storage medium
WO2020135055A1 (en) * 2018-12-28 2020-07-02 广州市百果园信息技术有限公司 Method, device and apparatus for adding video special effects and storage mediem
US20200234497A1 (en) * 2017-10-02 2020-07-23 Candera Japan Inc. Moving body image generation recording display device and program product
JP2020154428A (en) * 2019-03-18 2020-09-24 株式会社リコー Image processing apparatus, image processing method, image processing program, electronic device and photographing apparatus
CN112135049A (en) * 2020-09-24 2020-12-25 维沃移动通信有限公司 Image processing method and device and electronic equipment
CN112995500A (en) * 2020-12-30 2021-06-18 维沃移动通信(杭州)有限公司 Shooting method, shooting device, electronic equipment and medium
CN113810624A (en) * 2021-09-18 2021-12-17 维沃移动通信有限公司 Video generation method and device and electronic equipment
CN113905175A (en) * 2021-09-27 2022-01-07 维沃移动通信有限公司 Video generation method and device, electronic equipment and readable storage medium

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2005244536A (en) * 2004-02-26 2005-09-08 Seiko Epson Corp Image composition for generating composite image by overlapping image
US20110242130A1 (en) * 2010-03-30 2011-10-06 Casio Computer Co., Ltd. Image processing apparatus, image processing method, and computer-readable medium
US20160035116A1 (en) * 2013-03-29 2016-02-04 Rakuten, Inc. Image processing device, image processing method, information storage medium, and program
US20180070045A1 (en) * 2016-09-02 2018-03-08 Russell Holmes Systems And Methods For Providing Real-Time Composite Video From Multiple Source Devices
US20200234497A1 (en) * 2017-10-02 2020-07-23 Candera Japan Inc. Moving body image generation recording display device and program product
CN109104588A (en) * 2018-07-24 2018-12-28 房梦琦 A kind of video monitoring method, equipment, terminal and computer storage medium
JP2018198463A (en) * 2018-09-11 2018-12-13 日本電信電話株式会社 Image processing device, image processing method, and computer program
WO2020135055A1 (en) * 2018-12-28 2020-07-02 广州市百果园信息技术有限公司 Method, device and apparatus for adding video special effects and storage mediem
JP2020154428A (en) * 2019-03-18 2020-09-24 株式会社リコー Image processing apparatus, image processing method, image processing program, electronic device and photographing apparatus
CN112135049A (en) * 2020-09-24 2020-12-25 维沃移动通信有限公司 Image processing method and device and electronic equipment
CN112995500A (en) * 2020-12-30 2021-06-18 维沃移动通信(杭州)有限公司 Shooting method, shooting device, electronic equipment and medium
CN113810624A (en) * 2021-09-18 2021-12-17 维沃移动通信有限公司 Video generation method and device and electronic equipment
CN113905175A (en) * 2021-09-27 2022-01-07 维沃移动通信有限公司 Video generation method and device, electronic equipment and readable storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
S. R. KHARABE 等: "Matching of video objects taken from different camera views by using multi-feature fusion and evolutionary learning methods", 《IEEE》, 31 October 2016 (2016-10-31) *
张勇;陈大建;: "区域图像融合算法在红外图像分析中的应用", 光电技术应用, no. 03, 15 June 2011 (2011-06-15) *

Also Published As

Publication number Publication date
CN114520875B (en) 2024-04-02

Similar Documents

Publication Publication Date Title
CN113093968B (en) Shooting interface display method and device, electronic equipment and medium
CN112135046B (en) Video shooting method, video shooting device and electronic equipment
CN112492209B (en) Shooting method, shooting device and electronic equipment
CN111654635A (en) Shooting parameter adjusting method and device and electronic equipment
CN112954210B (en) Photographing method and device, electronic equipment and medium
CN111857512A (en) Image editing method and device and electronic equipment
CN112738402B (en) Shooting method, shooting device, electronic equipment and medium
CN113794829B (en) Shooting method and device and electronic equipment
CN112911147B (en) Display control method, display control device and electronic equipment
CN112995500A (en) Shooting method, shooting device, electronic equipment and medium
CN112565613B (en) Preview image display control method, device, electronic equipment and medium
CN111669495B (en) Photographing method, photographing device and electronic equipment
CN112437232A (en) Shooting method, shooting device, electronic equipment and readable storage medium
CN113794834A (en) Image processing method and device and electronic equipment
CN113194256B (en) Shooting method, shooting device, electronic equipment and storage medium
CN113207038B (en) Video processing method, video processing device and electronic equipment
CN113794831B (en) Video shooting method, device, electronic equipment and medium
CN112367467B (en) Display control method, display control device, electronic apparatus, and medium
CN114520875B (en) Video processing method and device and electronic equipment
CN114025100A (en) Shooting method, shooting device, electronic equipment and readable storage medium
CN113923392A (en) Video recording method, video recording device and electronic equipment
CN113873168A (en) Shooting method, shooting device, electronic equipment and medium
CN112165584A (en) Video recording method, video recording device, electronic equipment and readable storage medium
CN113014799A (en) Image display method and device and electronic equipment
CN114040099B (en) Image processing method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant