WO2022222835A1 - Procédé de traitement vidéo, appareil de traitement vidéo et dispositif électronique - Google Patents

Procédé de traitement vidéo, appareil de traitement vidéo et dispositif électronique Download PDF

Info

Publication number
WO2022222835A1
WO2022222835A1 PCT/CN2022/086751 CN2022086751W WO2022222835A1 WO 2022222835 A1 WO2022222835 A1 WO 2022222835A1 CN 2022086751 W CN2022086751 W CN 2022086751W WO 2022222835 A1 WO2022222835 A1 WO 2022222835A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
video frames
image
sub
images
Prior art date
Application number
PCT/CN2022/086751
Other languages
English (en)
Chinese (zh)
Inventor
韩桂敏
Original Assignee
维沃移动通信(杭州)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 维沃移动通信(杭州)有限公司 filed Critical 维沃移动通信(杭州)有限公司
Publication of WO2022222835A1 publication Critical patent/WO2022222835A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip

Definitions

  • the present application belongs to the technical field of electronic equipment, and specifically relates to a video processing method, a video processing apparatus and electronic equipment.
  • life can be recorded by shooting a video.
  • the video cannot be processed during the video shooting process.
  • editing processing such as segment division, interception or merging.
  • the video obtained after the editing process generally does not change the original video frame image. If the user wants to obtain a video different from the original video frame image, it cannot be achieved.
  • the purpose of the embodiments of the present application is to provide a video processing method, a video processing apparatus, and an electronic device, so as to solve the problem of poor video processing flexibility in the prior art.
  • an embodiment of the present application provides a video processing method, which includes:
  • M second images corresponding to M second video frames in the N video frames, wherein the M second images are images obtained by removing the third image from the M second video frames, and the M second images are obtained by removing the third image from the M second video frames.
  • the third image corresponds to the same object as the first image;
  • a first video is obtained.
  • an embodiment of the present application provides a video processing apparatus, including:
  • the first acquisition module is used to acquire N video frames
  • a receiving module configured to receive a first input for a first video frame in the N video frames
  • a second acquiring module configured to acquire the first image in the first video frame in response to the first input
  • a third acquiring module configured to acquire M second images corresponding to M second video frames in the N video frames, wherein the M second images are the M second video frames minus the third an image obtained from an image, the third image corresponds to the same object as the first image;
  • a first processing module configured to obtain M third video frames by merging the first image with the M second images respectively;
  • the second processing module is configured to obtain the first video according to the M third video frames.
  • embodiments of the present application provide an electronic device, the electronic device includes a processor, a memory, and a program or instruction stored on the memory and executable on the processor, the program or instruction being The processor implements the steps of the method according to the first aspect when executed.
  • an embodiment of the present application provides a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or instruction is executed by a processor, the steps of the method according to the first aspect are implemented .
  • an embodiment of the present application provides a chip, the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction, and implement the first aspect the method described.
  • the video processing apparatus can obtain the specific image in the specific video frame, and remove the other video frames with the same image. After the specific image is an image of the same object, the specific image is merged with other video frames to obtain a new video frame.
  • a new video frame that is different from the original video frame image can be obtained, and the user can also obtain a video different from the original video frame image. It can be seen that this can improve the flexibility of video processing.
  • FIG. 1 is a schematic flowchart of a video processing method provided by an embodiment of the present application.
  • FIGS. 2 to 5 are schematic diagrams of a freeze-frame shooting process provided by an embodiment of the present application.
  • 6 to 7 are schematic diagrams of performing background image restoration through image fusion technology provided by an embodiment of the present application.
  • FIG. 8 is a schematic diagram of a video frame provided by an embodiment of the present application.
  • 11 to 12 are schematic diagrams of adding graffiti material to a video frame provided by an embodiment of the present application.
  • FIG. 13 is a schematic structural diagram of a video processing apparatus provided by an embodiment of the present application.
  • FIG. 14 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
  • FIG. 15 is a schematic diagram of a hardware structure of an electronic device provided by an embodiment of the present application.
  • first, second and the like in the description and claims of the present application are used to distinguish similar objects, and are not used to describe a specific order or sequence. It is to be understood that the data so used are interchangeable under appropriate circumstances so that the embodiments of the present application can be practiced in sequences other than those illustrated or described herein, and distinguish between “first”, “second”, etc.
  • the objects are usually of one type, and the number of objects is not limited.
  • the first object may be one or more than one.
  • “and/or” in the description and claims indicates at least one of the connected objects, and the character “/" generally indicates that the associated objects are in an "or” relationship.
  • FIG. 1 shows a schematic flowchart of a video processing method provided by an embodiment of the present application.
  • the video processing method includes the following steps:
  • Step 101 Acquire N video frames
  • Step 102 receiving a first input for a first video frame in the N video frames
  • Step 103 Acquire a first image in the first video frame in response to the first input
  • Step 104 Acquire M second images corresponding to M second video frames in the N video frames, wherein the M second images are images obtained by removing the third image from the M second video frames , the third image corresponds to the same object as the first image;
  • Step 105 Merging the first image and the M second images respectively to obtain M third video frames
  • Step 106 Obtain a first video according to the M third video frames.
  • the video processing methods provided in the embodiments of the present application may be executed by a video processing apparatus.
  • the video processing apparatus may acquire N video frames during the video shooting process, or may acquire N video frames after the video shooting, which is not limited in this application.
  • the N video frames may be all or part of video frames of a certain video, or may be all or part of video frames of multiple videos.
  • the video processing apparatus receives a first input for the first video frame among the N video frames, where the first input can be understood as an input for selecting the first video frame, for example, in the process of playing the N video frames , and a click operation of the user is received at the moment of playing the first video frame, the click operation may be regarded as the first input for the first video frame.
  • the first video frame can be understood as the video frame that the user wants to freeze, and the moment when the user inputs the first input can be understood as the freeze moment.
  • the video processing apparatus may acquire the first image in the first video frame in response to the first input.
  • the first image may be understood as an image of the target object that the user wants to freeze in the first video frame, and the first image may be referred to as a freeze image.
  • the target object that the user wants to freeze can be either a moving object, such as people, animals, vehicles, etc., or a static object, such as buildings, objects, etc.
  • the target object the user wants to freeze can be either a foreground image or a background. image.
  • the first input may also be an input for the first image in the first video frame.
  • the video processing apparatus may extract the first image of the target object from the first video frame based on the semantic segmentation technology. Specifically, when the target object is a human being, the video processing apparatus may extract the human portrait from the first video frame based on the human portrait segmentation technology.
  • the term "portrait segmentation” refers to the separation of the portrait and the background in the picture, into different areas, and using different labels to distinguish them. It is suitable for application scenarios based on image content understanding, such as background replacement, rendering, blurring, etc.
  • the video processing apparatus may acquire M second images corresponding to the M second video frames in the N video frames, where the M second images are images obtained by removing the third image from the M second video frames, and the M second images are obtained by removing the third image from the M second video frames.
  • the three images correspond to the same object as the first image.
  • the M second video frames can be understood as video frames other than the first video frame among the N video frames, and can also be understood as the video frames located after the first video frame among the N video frames, and can also be understood as N video frames.
  • the video frame that is located before the first video frame in the video frame is not described in detail in this embodiment of the present application.
  • the first image is the image of person A in the first video frame
  • the third image is also the image of person A in the M second video frames
  • the M second images are the removed images in the M second video frames The image obtained after adding the image of Person A.
  • the video processing apparatus may segment the third image and the second image from the M second video frames based on the semantic segmentation technology, so as to obtain the M second images.
  • the video processing apparatus may combine the first image and the M second images respectively to obtain M third video frames.
  • the freeze-motion image is fused into other video frames, thereby obtaining a new video frame that is different from the original video frame image.
  • the position of the first image in the M third video frames may or may not correspond to the position of the third image in the M second video frames.
  • the video processing apparatus may obtain the first video according to the M third video frames.
  • the M third video frames may be combined to form the first video, or the M third video frames may be combined with other video frames to form the first video.
  • the video processing apparatus in addition to obtaining the first video, can also combine the first image and the second image to obtain the target photo.
  • the video processing apparatus can obtain the specific image in the specific video frame, and remove the other video frames with the same image. After the specific image is an image of the same object, the specific image is merged with other video frames to obtain a new video frame.
  • a new video frame that is different from the original video frame image can be obtained, which enables the user to obtain a video different from the original video frame image. It can be seen that this can improve the flexibility of video processing.
  • the first video frame includes a first sub-video frame and a second sub-video frame;
  • the acquiring the first image in the first video frame includes:
  • the first sub-image and the second sub-image are respectively combined with the M second images to obtain M third video frames.
  • the user can input the first input for multiple video frames as needed, or in other words, the user can input multiple first inputs, and the video processing device can freeze the target object once each time the user inputs the first input. , so as to obtain multiple images of the target object presented on the shooting interface.
  • the target object may be in a moving state, for example, the target object may change its position or posture.
  • the video processing device can obtain a clear and accurate image of the target object. A plurality of first images that change dynamically in sequence.
  • the video processing apparatus may also add a freeze mark to the first video frame, so that the user can clearly view the video frame freezed by the user by browsing the freeze mark.
  • the video processing apparatus may combine the multiple first images with the M second images to obtain M third video frames, so that the M third video frames can be presented clearly and smoothly.
  • the first video can present multiple clear and dynamic images of the target object, the first video has a streamer effect.
  • a portrait streamer video can be shot through the above process, and the portrait streamer video It can clearly retain each stop-motion portrait during the shooting process, expand the application scenarios of long-exposure photography, and improve the fun of shooting.
  • the video processing apparatus may photograph the target object in a long-exposure photographing mode, thereby acquiring N video frames.
  • the user can select the long-exposure shooting mode.
  • the user can click the "start shooting button" in the shooting preview interface.
  • the video processing device can shoot the target object. Take a shot, and acquire N video frames.
  • the user can freeze the target object by inputting a first input.
  • the first input can be the user clicking the "photograph button" in the shooting interface, or the user pressing the video processing
  • a certain physical button of the device may also be a user inputting a voice command, and so on.
  • the video processing apparatus may obtain the image (ie, the first image) of the target object presented on the shooting interface at the moment when the user inputs the first input.
  • the video processing apparatus may acquire the first image of the target object from the video frame captured when the first input is received.
  • the user can input a second input to end the shooting process.
  • a certain physical button it can also be a user input voice command, and so on.
  • the video processing device can obtain the first video, so that the first video can present a plurality of clear and dynamic images of the target object.
  • the camera device 20 freezes the target object 22 , and the camera device 20 captures the target object at the moment of the click. The image presented in the capture interface.
  • the camera 20 freezes the target object 22 again, and the camera 20 obtains the image of the target object displayed on the shooting interface again at the click moment. So far, the camera 20 can obtain two images of the target object 22 being frozen.
  • the camera 20 can generate a video including two images of the target object 22 being frozen.
  • the target object in the process of photographing the target object, is freezed multiple times through multiple first inputs, and an image of the target object presented on the photographing interface is acquired for each first input, that is, Generate a video frame containing multiple images of the target object.
  • an image of the target object presented on the photographing interface is acquired for each first input, that is, Generate a video frame containing multiple images of the target object.
  • the video processing device in addition to obtaining the first video, can also combine the first sub-image, the second sub-image and the second image to obtain a target photo, which can present a clear and dynamic image of the target object Multiple images that change in sequence can be called streamer photos.
  • the method further includes:
  • the K fourth images in the K fourth video frames are removed, and the removed areas of the K fourth video frames are repaired to obtain K fifth video frames, wherein the fourth images and the first image corresponds to the same object;
  • the obtaining of the M second images corresponding to the M second video frames in the N video frames includes:
  • the obtaining the first video according to the M third video frames includes:
  • a first video is obtained according to the K fifth video frames and the M third video frames.
  • the video processing apparatus may perform different processing on the first video frame, the K fourth video frames located before the first video frame, and the M second video frames located after the first video frame among the N video frames. processing to further expand the application scenarios of long exposure photography and enhance the fun of shooting.
  • the video processing device can analyze and process each video frame captured in real time during the shooting process, separate the target object and the background in each video frame, and mark them with different labels.
  • the image of the target object ie, the first image
  • the background image may be retained.
  • the image of the target object ie the third image
  • only the background image ie the second image
  • the first image and the second image are respectively combined to obtain the third video frame.
  • the image of the target object (that is, the fourth image) can be eliminated, and only the background image can be retained.
  • Image Fusion refers to the process of image processing and computer technology for the image data of the same target collected by multi-source channels, to maximize the extraction of favorable information in the respective channels, and finally to synthesize them into high-quality images. image.
  • FIGS. 6 to 7 show schematic diagrams of background image restoration through image fusion technology.
  • frame N is an unfrozen video frame
  • FIG. 7 extract the background image 25 of the corresponding area in frame N-2, cover the target object 22 in frame N, and fuse to generate a new video frame N in which the target object has been eliminated and only the background image 25 is retained.
  • frame N-2 the background image 25 of the corresponding area in frame N-2
  • the streamer video can not only clearly retain each frozen image of the target object during the shooting process, but also clearly retain the environment image where the target object is located, further expanding the long-exposure photography.
  • the application scene enhances the fun of shooting.
  • the method further includes:
  • the obtaining of the M second images corresponding to the M second video frames in the N video frames includes:
  • the obtaining the first video according to the M third video frames includes:
  • a first video is obtained according to the S seventh video frames and the M third video frames.
  • the entire process of the dynamic change of the target object can be displayed.
  • the first video when the first video is played to the node at the first freeze moment (ie, the first sub-video frame), the image of the target object presented on the shooting interface at the freeze moment can be displayed.
  • the first video When the first video is played to the node at the second freeze moment (ie, the second sub-video frame), the image of the target object presented on the shooting interface at the first freeze moment and the second freeze moment can be displayed, and so on.
  • the first video shows the process of the dynamic change of the target object.
  • the first video When the first video is played to a node at a non-freeze moment, an image of the target object presented on the shooting interface when the node including the last freeze moment can be displayed. That is to say, once the target object displayed by the node at any freeze moment is displayed, it will always exist in the video segment after the freeze moment.
  • frame m+3 is the video frame at the first freeze moment
  • frame m+11 is the second freeze frame
  • the video frame of the moment showing the images of the target object 22 presented on the shooting interface at the first freeze moment and the second freeze moment.
  • frame m+4 to frame m+10 can all display the image presented by the target object 22 on the shooting interface at the first freeze moment.
  • the method before the obtaining of the first video, the method further includes:
  • the obtaining the first video according to the M third video frames includes:
  • a first video is obtained according to the M third video frames and the dynamic graffiti.
  • the motion trajectory of the target object when the target object is a person, the motion trajectory of the target object may be a hand motion trajectory.
  • the motion trajectory of the target object may also be the footprint of the target object, which is not limited in this embodiment of the present application.
  • a virtual graffiti material can be added to the first video based on the human gesture recognition technology, and the user can select a pre-provided special effect material to describe the hand movement trajectory of the target object when shooting.
  • the user can click the material selection button 26 before clicking the start shooting button 21 to select the special effect for recording the hand movement track.
  • the user can click the material selection button 26 to switch between different virtual materials at any time during the shooting process.
  • Human gesture recognition technology can recognize gestures by processing real video images based on computer vision, including gesture segmentation, trajectory tracking and classification and recognition.
  • Tracking hand motion trajectory can use image tracking algorithm, such as optical flow method, continuously adaptive MeanShift algorithm (Continuously Adaptive Mean-SHIFT, camshift), Kernel Correlation Filter algorithm (Kernel Correlation Filter, KCF), deep learning and other algorithms, or , the hand position detected in each video frame may also be used directly to track the hand motion trajectory, which is not limited in this embodiment of the present application.
  • the video processing device can analyze the hand movement trajectory of the target object in real time through an algorithm, and follow the hand movement trajectory to add the selected graffiti material. For example, if the "love shape" material is selected, the heart-shaped graffiti will be displayed at the dynamic track traced by the hand movement track.
  • Figures 11 to 12 respectively show video frames to which graffiti materials 27 are added. These graffiti materials can present effects such as light painting, fireworks, etc., and present the dynamic trajectory of the target object.
  • the target can be presented. The dynamic trajectory of the object enhances the fun of shooting.
  • FIG. 13 shows a schematic structural diagram of a video processing apparatus provided by an embodiment of the present application.
  • the video processing apparatus 300 includes:
  • the first acquisition module 301 is used to acquire N video frames
  • a second acquiring module 303 configured to acquire the first image in the first video frame in response to the first input
  • a third obtaining module 304 configured to obtain M second images corresponding to M second video frames in the N video frames, wherein the M second images are the M second video frames except the first video frame. An image obtained from three images, the third image corresponds to the same object as the first image;
  • a first processing module 305 configured to combine the first image with the M second images respectively to obtain M third video frames
  • the second processing module 306 is configured to obtain a first video according to the M third video frames.
  • the video processing apparatus 300 further includes:
  • a fourth acquisition module configured to acquire K fourth video frames before the first video frame in the N video frames
  • the third processing module is configured to remove the K fourth images in the K fourth video frames, and repair the removed areas of the K fourth video frames to obtain K fifth video frames, wherein , the fourth image corresponds to the same object as the first image;
  • the third obtaining module 304 is specifically used for:
  • the second processing module 306 is specifically used for:
  • a first video is obtained according to the K fifth video frames and the M third video frames.
  • the first video frame includes a first sub-video frame and a second sub-video frame;
  • the second obtaining module 303 is specifically used for:
  • the first processing module 305 is specifically used for:
  • the first sub-image and the second sub-image are respectively combined with the M second images to obtain M third video frames.
  • the video processing apparatus 300 further includes:
  • a fifth acquisition module configured to acquire S sixth video frames located between the first sub-video frame and the second sub-video frame in the N video frames;
  • a sixth obtaining module configured to obtain S fifth images corresponding to the S sixth video frames, wherein the S fifth images are images obtained by removing the sixth image from the S sixth video frames, the sixth image corresponds to the same object as the first sub-image;
  • a fourth processing module configured to combine the first sub-image with the S fifth images to obtain S seventh video frames
  • the third obtaining module 304 is specifically used for:
  • the third processing module is specifically used for:
  • a first video is obtained according to the S seventh video frames and the M third video frames.
  • the video processing apparatus 300 further includes:
  • a collection module used for collecting the motion trajectory of the target object in the N video frames
  • a generating module for generating dynamic graffiti for depicting the motion trajectory according to the motion trajectory
  • the third processing module is specifically used for:
  • a first video is obtained according to the M third video frames and the dynamic graffiti.
  • the video processing apparatus in this embodiment of the present application may be an apparatus, or may be a component, an integrated circuit, or a chip in a terminal.
  • the apparatus may be a mobile electronic device or a non-mobile electronic device.
  • the mobile electronic device may be a mobile phone, a tablet computer, a notebook computer, a palmtop computer, an in-vehicle electronic device, a wearable device, an ultra-mobile personal computer (UMPC), a netbook, or a personal digital assistant (personal digital assistant).
  • UMPC ultra-mobile personal computer
  • PDA personal digital assistant
  • non-mobile electronic devices can be servers, network attached storage (Network Attached Storage, NAS), personal computer (personal computer, PC), television (television, TV), teller machine or self-service machine, etc., this application Examples are not specifically limited.
  • the video processing apparatus in this embodiment of the present application may be an apparatus having an operating system.
  • the operating system may be an Android (Android) operating system, an iOS operating system, or other possible operating systems, which are not specifically limited in the embodiments of the present application.
  • the video processing apparatus provided in the embodiments of the present application can implement the various processes implemented by the method embodiments in FIG. 1 to FIG. 12 , and can achieve the same beneficial effects. To avoid repetition, details are not repeated here.
  • the video processing apparatus can obtain the specific image in the specific video frame, and remove the other video frames with the same image. After the specific image is an image of the same object, the specific image is merged with other video frames to obtain a new video frame.
  • a new video frame that is different from the original video frame image can be obtained, which enables the user to obtain a video different from the original video frame image. It can be seen that this can improve the flexibility of video processing.
  • an embodiment of the present application further provides an electronic device 400, including a processor 401, a memory 402, a program or instruction stored in the memory 402 and executable on the processor 401, When the program or instruction is executed by the processor 401, each process of the above video processing method embodiment is implemented, and the same technical effect can be achieved. To avoid repetition, details are not described here.
  • the electronic devices in the embodiments of the present application include the aforementioned mobile electronic devices and non-mobile electronic devices.
  • FIG. 15 is a schematic diagram of a hardware structure of an electronic device implementing an embodiment of the present application.
  • the electronic device 500 includes but is not limited to: a radio frequency unit 501, a network module 502, an audio output unit 503, an input unit 504, a sensor 505, a display unit 506, a user input unit 507, an interface unit 508, a memory 509, and a processor 5010, etc. part.
  • the electronic device 500 may also include a power source (such as a battery) for supplying power to various components, and the power source may be logically connected to the processor 5010 through a power management system, so as to manage charging, discharging, and power management through the power management system. consumption management and other functions.
  • a power source such as a battery
  • the structure of the electronic device shown in FIG. 15 does not constitute a limitation to the electronic device.
  • the electronic device may include more or less components than the one shown, or combine some components, or arrange different components, which will not be repeated here. .
  • the processor 5010 is used for: acquiring N video frames;
  • the user input unit 507 is configured to: receive a first input for a first video frame in the N video frames;
  • the processor 5010 is further configured to: in response to the first input, obtain a first image in the first video frame; obtain M second images corresponding to M second video frames in the N video frames, Wherein, the M second images are images obtained by removing a third image from the M second video frames, and the third image corresponds to the same object as the first image; The M second images are combined to obtain M third video frames; according to the M third video frames, a first video is obtained.
  • the video processing apparatus can obtain the specific image in the specific video frame, and remove the other video frames with the same image. After the specific image is an image of the same object, the specific image is merged with other video frames to obtain a new video frame.
  • a new video frame that is different from the original video frame image can be obtained, which enables the user to obtain a video different from the original video frame image. It can be seen that this can improve the flexibility of video processing.
  • processor 5010 is further configured to:
  • the K fourth images in the K fourth video frames are removed, and the removed areas of the K fourth video frames are repaired to obtain K fifth video frames, wherein the fourth images and the first image corresponds to the same object;
  • a first video is obtained according to the K fifth video frames and the M third video frames.
  • the first video frame includes a first sub-video frame and a second sub-video frame;
  • Processor 5010 is also used to:
  • the first sub-image and the second sub-image are respectively combined with the M second images to obtain M third video frames.
  • processor 5010 is further configured to:
  • a first video is obtained according to the S seventh video frames and the M third video frames.
  • processor 5010 is further configured to:
  • a first video is obtained according to the M third video frames and the dynamic graffiti.
  • the input unit 504 may include a graphics processor (Graphics Processing Unit, GPU) 5041 and a microphone 5042. Such as camera) to obtain still pictures or video image data for processing.
  • the display unit 506 may include a display panel 5061, which may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like.
  • the user input unit 507 includes a touch panel 5071 and other input devices 5072 .
  • the touch panel 5071 is also called a touch screen.
  • the touch panel 5071 may include two parts, a touch detection device and a touch controller.
  • Other input devices 5072 may include, but are not limited to, physical keyboards, function keys (such as volume control keys, switch keys, etc.), trackballs, mice, and joysticks, which are not described herein again.
  • Memory 509 may be used to store software programs as well as various data, including but not limited to application programs and operating systems.
  • the processor 5010 may integrate an application processor and a modem processor, wherein the application processor mainly processes an operating system, a user interface, an application program, and the like, and the modem processor mainly processes wireless communication. It can be understood that, the above-mentioned modulation and demodulation processor may not be integrated into the processor 5010.
  • Embodiments of the present application further provide a readable storage medium, where a program or an instruction is stored on the readable storage medium.
  • a program or an instruction is stored on the readable storage medium.
  • the processor is the processor in the electronic device described in the foregoing embodiments.
  • the readable storage medium includes a computer-readable storage medium, and examples of the computer-readable storage medium include non-transitory machine-readable storage media, such as computer read-only memory (Read-Only Memory, ROM), random access memory ( Random Access Memory, RAM), disk or CD, etc.
  • An embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement the above video processing method embodiments.
  • the chip includes a processor and a communication interface
  • the communication interface is coupled to the processor
  • the processor is configured to run a program or an instruction to implement the above video processing method embodiments.
  • the chip mentioned in the embodiments of the present application may also be referred to as a system-on-chip, a system-on-chip, a system-on-a-chip, or a system-on-a-chip, or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)

Abstract

La présente demande, qui appartient au domaine technique des dispositifs électroniques, concerne un procédé de traitement vidéo, un appareil de traitement vidéo et un dispositif électronique. Le procédé de traitement vidéo comprend : l'acquisition de N trames vidéo ; la réception d'une première entrée pour une première trame vidéo dans les N trames vidéo ; en réponse à la première entrée, l'acquisition d'une première image dans la première trame vidéo ; l'acquisition de M deuxièmes images correspondant à M deuxièmes trames vidéo dans les N trames vidéo, les M deuxièmes images étant des images qui sont obtenues en supprimant une troisième image des M deuxièmes trames vidéo, et la troisième image correspondant au même objet que la première image ; la fusion respective de la première image avec les M deuxièmes images pour obtenir M troisièmes trames vidéo ; et selon les M troisièmes trames vidéo, l'obtention d'une première vidéo.
PCT/CN2022/086751 2021-04-21 2022-04-14 Procédé de traitement vidéo, appareil de traitement vidéo et dispositif électronique WO2022222835A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110432361.3A CN113207038B (zh) 2021-04-21 2021-04-21 视频处理方法、视频处理装置和电子设备
CN202110432361.3 2021-04-21

Publications (1)

Publication Number Publication Date
WO2022222835A1 true WO2022222835A1 (fr) 2022-10-27

Family

ID=77027721

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/086751 WO2022222835A1 (fr) 2021-04-21 2022-04-14 Procédé de traitement vidéo, appareil de traitement vidéo et dispositif électronique

Country Status (2)

Country Link
CN (1) CN113207038B (fr)
WO (1) WO2022222835A1 (fr)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113207038B (zh) * 2021-04-21 2023-04-28 维沃移动通信(杭州)有限公司 视频处理方法、视频处理装置和电子设备
CN113810624A (zh) * 2021-09-18 2021-12-17 维沃移动通信有限公司 视频生成的方法、装置和电子设备

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102164234A (zh) * 2010-02-09 2011-08-24 株式会社泛泰 具有拍摄功能的设备
CN106375670A (zh) * 2016-09-30 2017-02-01 努比亚技术有限公司 一种图片处理方法及终端
US20190026567A1 (en) * 2017-07-21 2019-01-24 Satori Worldwide, Llc System and method for counting objects on multiple roads using a pan, tilt, and zoom camera
CN110225241A (zh) * 2019-04-29 2019-09-10 努比亚技术有限公司 一种视频拍摄控制方法、终端及计算机可读存储介质
CN111601033A (zh) * 2020-04-27 2020-08-28 北京小米松果电子有限公司 视频处理方法、装置及存储介质
CN111614902A (zh) * 2020-05-27 2020-09-01 努比亚技术有限公司 一种视频拍摄方法、设备及计算机可读存储介质
CN113207038A (zh) * 2021-04-21 2021-08-03 维沃移动通信(杭州)有限公司 视频处理方法、视频处理装置和电子设备

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103905716B (zh) * 2012-12-27 2017-08-18 三星电子(中国)研发中心 拍摄照片时动态处理取景画面的照相设备及方法
CN105141974B (zh) * 2015-07-29 2018-11-09 百度在线网络技术(北京)有限公司 一种视频剪辑方法和装置
CN106385591B (zh) * 2016-10-17 2020-05-15 腾讯科技(上海)有限公司 视频处理方法及视频处理装置
CN107105310B (zh) * 2017-05-05 2020-07-10 广州盈可视电子科技有限公司 一种视频直播中人物形象替换方法、装置和一种录播系统
CN108537867B (zh) * 2018-04-12 2020-01-10 北京微播视界科技有限公司 根据用户肢体运动的视频渲染方法和装置
CN109348277B (zh) * 2018-11-29 2020-02-07 北京字节跳动网络技术有限公司 运动像素视频特效添加方法、装置、终端设备及存储介质
CN110047124A (zh) * 2019-04-23 2019-07-23 北京字节跳动网络技术有限公司 渲染视频的方法、装置、电子设备和计算机可读存储介质
CN110266973B (zh) * 2019-07-19 2020-08-25 腾讯科技(深圳)有限公司 视频处理方法、装置、计算机可读存储介质和计算机设备
CN111698553B (zh) * 2020-05-29 2022-09-27 维沃移动通信有限公司 视频处理方法、装置、电子设备及可读存储介质
CN112492209B (zh) * 2020-11-30 2022-04-29 维沃移动通信有限公司 拍摄方法、拍摄装置和电子设备
CN112565868B (zh) * 2020-12-04 2022-12-06 维沃移动通信有限公司 视频播放方法、装置及电子设备

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102164234A (zh) * 2010-02-09 2011-08-24 株式会社泛泰 具有拍摄功能的设备
CN106375670A (zh) * 2016-09-30 2017-02-01 努比亚技术有限公司 一种图片处理方法及终端
US20190026567A1 (en) * 2017-07-21 2019-01-24 Satori Worldwide, Llc System and method for counting objects on multiple roads using a pan, tilt, and zoom camera
CN110225241A (zh) * 2019-04-29 2019-09-10 努比亚技术有限公司 一种视频拍摄控制方法、终端及计算机可读存储介质
CN111601033A (zh) * 2020-04-27 2020-08-28 北京小米松果电子有限公司 视频处理方法、装置及存储介质
CN111614902A (zh) * 2020-05-27 2020-09-01 努比亚技术有限公司 一种视频拍摄方法、设备及计算机可读存储介质
CN113207038A (zh) * 2021-04-21 2021-08-03 维沃移动通信(杭州)有限公司 视频处理方法、视频处理装置和电子设备

Also Published As

Publication number Publication date
CN113207038B (zh) 2023-04-28
CN113207038A (zh) 2021-08-03

Similar Documents

Publication Publication Date Title
EP2750372B1 (fr) Dispositif photographique de production d'images composites et son procédé d'utilisation
WO2022222835A1 (fr) Procédé de traitement vidéo, appareil de traitement vidéo et dispositif électronique
US9137461B2 (en) Real-time camera view through drawn region for image capture
CN112714253B (zh) 视频录制方法、装置、电子设备和可读存储介质
WO2022078241A1 (fr) Procédé et appareil de photographie, et dispositif électronique
US20240144971A1 (en) System and Method for Performing a Rewind Operation with a Mobile Image Capture Device
CN111669495B (zh) 拍照方法、拍照装置和电子设备
WO2023134583A1 (fr) Procédé et appareil d'enregistrement vidéo et dispositif électronique
CN117152660A (zh) 图像显示方法及其装置
WO2023151527A1 (fr) Procédé et appareil de photographie d'image
CN114125297B (zh) 视频拍摄方法、装置、电子设备及存储介质
CN114025237B (zh) 视频生成方法、装置和电子设备
WO2022161261A1 (fr) Procédé et appareil d'affichage d'image, et dispositif électronique
CN112367467B (zh) 显示控制方法、装置、电子设备和介质
CN115801977A (zh) 分割视频的多模态系统、分割多媒体的多模态系统及方法
CN112887623B (zh) 图像生成方法、装置及电子设备
CN112887611A (zh) 图像处理方法、装置、设备和存储介质
CN112672059B (zh) 一种拍摄方法及拍摄装置
CN114500852B (zh) 拍摄方法、拍摄装置、电子设备和可读存储介质
CN113873160B (zh) 图像处理方法、装置、电子设备和计算机存储介质
CN117294932A (zh) 拍摄方法、拍摄装置和电子设备
CN117278842A (zh) 拍摄控制方法、装置、电子设备和可读存储介质
CN116156076A (zh) 视频录制方法、装置、电子设备及存储介质
CN114143455A (zh) 拍摄方法、装置及电子设备
CN115242976A (zh) 拍摄方法、拍摄装置和电子设备

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22790943

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22790943

Country of ref document: EP

Kind code of ref document: A1