WO2022100735A1 - 视频处理方法、装置、电子设备及存储介质 - Google Patents
视频处理方法、装置、电子设备及存储介质 Download PDFInfo
- Publication number
- WO2022100735A1 WO2022100735A1 PCT/CN2021/130708 CN2021130708W WO2022100735A1 WO 2022100735 A1 WO2022100735 A1 WO 2022100735A1 CN 2021130708 W CN2021130708 W CN 2021130708W WO 2022100735 A1 WO2022100735 A1 WO 2022100735A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- target
- image
- frame image
- video
- processing method
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims abstract description 41
- 230000000694 effects Effects 0.000 claims abstract description 82
- 230000000007 visual effect Effects 0.000 claims abstract description 24
- 230000004044 response Effects 0.000 claims abstract description 18
- 230000000295 complement effect Effects 0.000 claims description 63
- 238000000034 method Methods 0.000 claims description 50
- 238000012545 processing Methods 0.000 claims description 43
- 230000008569 process Effects 0.000 claims description 28
- 238000004590 computer program Methods 0.000 claims description 22
- 239000002245 particle Substances 0.000 claims description 7
- 230000008921 facial expression Effects 0.000 claims description 6
- 230000011218 segmentation Effects 0.000 claims description 5
- 230000008034 disappearance Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 20
- 230000006870 function Effects 0.000 description 9
- 238000004891 communication Methods 0.000 description 8
- 230000003287 optical effect Effects 0.000 description 6
- 230000003993 interaction Effects 0.000 description 4
- 241001465754 Metazoa Species 0.000 description 3
- 238000009432 framing Methods 0.000 description 3
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 101000822695 Clostridium perfringens (strain 13 / Type A) Small, acid-soluble spore protein C1 Proteins 0.000 description 1
- 101000655262 Clostridium perfringens (strain 13 / Type A) Small, acid-soluble spore protein C2 Proteins 0.000 description 1
- 101000655256 Paraclostridium bifermentans Small, acid-soluble spore protein alpha Proteins 0.000 description 1
- 101000655264 Paraclostridium bifermentans Small, acid-soluble spore protein beta Proteins 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000002311 subsequent effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/77—Retouching; Inpainting; Scratch removal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/2621—Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/222—Studio circuitry; Studio devices; Studio equipment
- H04N5/262—Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
- H04N5/272—Means for inserting a foreground image in a background image, i.e. inlay, outlay
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
- G06T2207/20104—Interactive definition of region of interest [ROI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Definitions
- the present disclosure relates to the technical field of video processing, and in particular, to a video processing method, apparatus, electronic device, and storage medium.
- the present disclosure provides a video processing method, device, electronic device, storage medium computer program product and computer program, which are used to solve the problem that in current video interactive applications, only the image recording of the object to be photographed cannot satisfy the user's need for video recording or playback. Technical questions about the need for diversity of interaction modes.
- an embodiment of the present disclosure provides a video processing method, including:
- the target area is to include the target object from The removed vacancy area in the target frame image.
- an embodiment of the present disclosure provides a video processing apparatus, including:
- an image acquisition module for fixing the target frame image of the target video in response to a trigger instruction
- the image processing module is used to remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image, so as to generate and display the completion frame image, the target area Including the vacant area after the target object is removed from the target frame image.
- an electronic device including:
- a memory for storing a computer program for the processor
- a display for displaying the video processed by the processor
- the processor is configured to implement the video processing method described above in the first aspect and various possible designs of the first aspect by executing the computer program.
- embodiments of the present disclosure provide a computer-readable storage medium, where computer-executable instructions are stored in the computer-readable storage medium, and when a processor executes the computer-executable instructions, the first aspect and the first Aspects of the video processing methods described in various possible designs.
- an embodiment of the present disclosure provides a computer program product, where the program product includes: a computer program, which implements the steps of the method described in the first aspect of the present disclosure when the computer program is executed by a processing device.
- a target frame image of a target video in response to a trigger instruction, a target frame image of a target video is fixed, and a target object in the target frame image is removed, and then the target frame image is removed.
- Complement the target area in the target frame image to generate a complement frame image, so as to realize the special effect that the target object disappears in the target video during the process of video recording or playback, and then serve the user in the video application.
- FIG. 1 is an application scenario diagram of a video processing method according to an exemplary embodiment of the present disclosure
- FIG. 2 is a schematic flowchart of a video processing method according to an exemplary embodiment of the present disclosure
- FIG. 3 is a schematic interface diagram of a video processing process shown in the present disclosure according to an exemplary embodiment
- FIG. 4 is another application scenario diagram of the video processing method shown in the present disclosure according to an exemplary embodiment
- FIG. 5 is a schematic diagram of an intermediate process of a video processing process shown in the present disclosure according to an exemplary embodiment
- FIG. 6 is a schematic diagram of another interface of a video processing process shown in the present disclosure according to an exemplary embodiment
- FIG. 7 is a schematic flowchart of a video processing method according to another exemplary embodiment of the present disclosure.
- FIG. 8 is a schematic structural diagram of a video processing apparatus according to an exemplary embodiment of the present disclosure.
- FIG. 9 is a schematic structural diagram of an electronic device according to an exemplary embodiment of the present disclosure.
- the term “including” and variations thereof are open-ended inclusions, ie, "including but not limited to”.
- the term “based on” is “based at least in part on.”
- the term “one embodiment” means “at least one embodiment”; the term “another embodiment” means “at least one additional embodiment”; the term “some embodiments” means “at least some embodiments”. Relevant definitions of other terms will be given in the description below.
- a target frame image of a target video is fixed, and the target object in the target frame image is removed, and then the target frame image is removed.
- Completion operation is performed on the target area to generate and display the completion frame image, so that during the process of video recording or playback, the user can input specific instructions to achieve the special effect of the target object disappearing in the target video, and then in the video.
- Visual effects that provide users with more personalized and better visual experience in interactive applications.
- FIG. 1 is an application scenario diagram of a video processing method according to an exemplary embodiment of the present disclosure.
- the video processing method provided in this embodiment can be executed by a terminal device with a camera and a display screen.
- the video recording of the target object (such as a person, an animal, a building, etc.) can be performed through the camera on the terminal device (for example, a front camera, a rear camera, an external camera, etc.), and the display of the terminal device can also be used. screen to play the target video.
- the target object can be used as an example to illustrate.
- the camera on the terminal device is usually aimed at the target person to shoot. It is understandable that in During the framing process, the camera will not only obtain the target person, but also obtain the target background.
- the user can input trigger commands (for example: target gesture command, target voice command, target expression command, target text instructions and target body instructions, etc.) to trigger the target effect, thereby fixing the target frame image in the currently captured target video.
- trigger commands for example: target gesture command, target voice command, target expression command, target text instructions and target body instructions, etc.
- the target video when the target video is played by the terminal device, the target video also includes the target background in addition to the target person.
- the user can input a trigger instruction (such as : target gesture command, target voice command, target expression command, target text command, target body command, etc.) to trigger the target effect, thereby fixing the target frame image in the currently captured target video.
- a trigger instruction such as : target gesture command, target voice command, target expression command, target text command, target body command, etc.
- the trigger instruction may be a trigger instruction input by the user, or a trigger instruction issued by the target object in the video, and may also include the terminal under preset conditions (for example, the video is played to a preset time, etc.) Trigger command issued.
- the target frame image may be the first frame image after the trigger instruction is acquired, or may be the frame image after the trigger instruction is acquired with a preset duration.
- After fixing the target frame image remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image. For example, a background completion operation may be performed to generate and display the completion frame image, wherein , the target area includes the vacant area after the target object is removed from the target frame image.
- the target special effect can be triggered by triggering instructions, so that the target object is removed from the target frame image and the completion operation is performed, so as to realize the visual disappearance of the target object from the video. Effect.
- FIG. 2 is a schematic flowchart of a video processing method according to an exemplary embodiment of the present disclosure. As shown in Figure 2, the video processing method provided by this embodiment includes:
- Step 101 in response to the trigger instruction, fix the target frame image of the target video.
- the camera on the terminal device is usually aimed at the target object to shoot. It is understandable that in the process of shooting the target object, the camera In addition to acquiring the target object, the target background is also acquired.
- the target video when the target video is played by a terminal device, the target video usually includes the target object and the target background. It is understandable that the target video is generated based on the target background when shooting. The target object is photographed. Among them, for the convenience of understanding, the target person may be selected as the target object for illustration. 1, when a terminal device is used to shoot a video of a target person, the camera on the terminal device is usually aimed at the target person to shoot. It is understandable that during the framing process, the camera will not only acquire the target person In addition, the target background is also obtained at the same time.
- the target video when the target video is played by the terminal device, since the target video will include the target object and the target background, during playback, the relevant picture based on the target object under the target background will be played.
- FIG. 3 is a schematic diagram of an interface of a video processing process according to an exemplary embodiment of the present disclosure. As shown in FIG. 3 , when a terminal device is used to shoot a video of a target person or play a target video, the target background and the target person will be displayed on the display screen of the terminal device at the same time.
- the terminal device fixes the target frame image of the target video in response to the trigger instruction.
- 4 is a diagram of another application scenario of the video processing method according to an exemplary embodiment of the present disclosure.
- the scene in FIG. 4 may be that in the process of recording the target video of the target person, the user triggers the effect of the person disappearing from the image by inputting the target gesture instruction (for example, the instruction to reach out).
- the target gesture instruction for example, the instruction to reach out
- the target frame image of the currently captured target video is fixed, wherein the current frame of the target video can be fixed.
- the terminal device when the terminal device is playing the target video, the user can trigger the effect of the character disappearing from the image through the target gesture command (for example, a hand reaching command).
- the terminal device recognizes the target gesture command, the The target frame image of the currently captured target video, which may be the current frame of the fixed target video.
- Step 102 remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image, so as to generate and display the completion frame image.
- the target frame image can be processed to achieve the effect that the target object disappears in the target frame image. Specifically, it may be to remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image to generate and display the completion frame image, and the target area is to remove the target object from the target frame image. behind the vacant area.
- FIG. 5 is a schematic diagram of an intermediate process of a video processing process according to an exemplary embodiment of the present disclosure. As shown in FIG. 5 , after fixing the target frame image, the target person in the target frame image may be removed. 6 is another interface schematic diagram of the video processing process shown in the present disclosure according to an exemplary embodiment. As shown in FIG. 6, after removing the target person in the target frame image, the target area, that is, the formed vacant area, Perform background completion operations.
- the user triggers by inputting a hand-out instruction, so as to realize that in the video recording process, in response to the user gesture, the target person is in the target video. disappearing effects.
- the user can also realize the special effect that the target character disappears in the target video by triggering an instruction.
- the target object may be in the form of a target person, a target animal, a target building, etc., which is not specifically limited here.
- the above trigger instruction may also be in the form of a target gesture command, a target voice command, a target facial expression command, a target text command, a target body command, etc., which are also not specifically limited herein.
- the target frame image of the target video in response to the trigger instruction, is fixed, and the target object in the target frame image is removed, and then the completion operation is performed on the target area in the target frame image to Generate and display the complementary frame image, so that the user can realize the special effect of the target object disappearing in the target video by inputting specific instructions, and then provide the user with a more personalized and better visual effect in the video application.
- FIG. 7 is a schematic flowchart of a video processing method according to another exemplary embodiment of the present disclosure. As shown in FIG. 7 , the video processing method provided by this embodiment includes:
- Step 201 in response to the trigger instruction, fix the target frame image of the target video.
- the camera on the terminal device is usually aimed at the target object to shoot. It is understandable that in the process of shooting the target object, the camera will not only acquire the target object , and also get the target background.
- the target video when the target video is played by a terminal device, the target video usually includes the target object and the target background. It is understandable that the target video is generated based on the target background when shooting. The target object is photographed.
- the terminal device can respond to the trigger instruction to fix the target frame image in the target video, wherein the target frame image can be the first frame image after acquiring the trigger instruction, or it can be the corresponding image after a preset time period after acquiring the trigger instruction. frame image.
- the target person may be selected as the target object for illustration.
- the camera on the terminal device is usually aimed at the target person to shoot. It is understandable that during the framing process, the camera will not only acquire the target person In addition, the target background is also obtained at the same time.
- the target video when the target video is played by the terminal device, since the target video will include the target object and the target background, during playback, the relevant picture based on the target object under the target background will be played.
- the target background and the target person will be displayed on the display screen of the terminal device at the same time.
- the terminal device fixes the target frame image of the target video in response to the trigger instruction.
- 4 is a diagram of another application scenario of the video processing method according to an exemplary embodiment of the present disclosure.
- the scene in FIG. 4 is that during the process of recording the target video of the target person, the target person triggers the effect of the person disappearing from the image by inputting a target gesture instruction (for example, a hand reaching instruction).
- a target gesture instruction for example, a hand reaching instruction
- the terminal device recognizes the target gesture instruction, it fixes the target frame image of the currently captured target video, which may be the current frame of the fixed target video.
- the terminal device when the terminal device is playing the target video, the user can trigger the effect of the character disappearing from the image through the target gesture command (for example, a hand reaching command).
- the terminal device recognizes the target gesture command, the The target frame image of the currently captured target video, which may be the current frame of the fixed target video.
- the special effect it is also possible to first determine whether the currently shot video contains a target type object. For example, it may be determined that the currently shot video contains a human object before responding to the corresponding trigger instruction to carry out Subsequent effects.
- Step 202 displaying a preset second special effect, where the preset second special effect is used to make the target frame image display a visual blur effect.
- a preset second special effect may be displayed, and the preset second special effect is used to make the target frame image display a visual blur effect.
- the preset second special effect is used to make the target frame image display a visual blur effect.
- Step 203 displaying a preset third special effect, which is used to make the target frame image display a visual shaking effect.
- a preset third special effect may also be displayed, wherein the preset third special effect is used to make the target frame image display a visual shaking effect.
- the preset third special effect is used to make the target frame image display a visual shaking effect.
- step 202 may be executed before step 203, or may be executed after step 203, or step 202 and step 203 may be performed independently. The execution may also be to execute only step 202 or step 203 .
- Step 204 remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image, so as to generate and display the completion frame image.
- the target frame image can be processed to achieve the effect that the target object disappears in the target frame image.
- the target object in the target frame image can be removed, and the background completion operation can be performed on the target area in the target frame image to generate and display the completion frame image, and the target area is to remove the target object from the target frame image. behind the vacant area.
- the target person can be used as the target object for illustration.
- the target person in the target frame image may be removed.
- the target area that is, the formed vacant area, may be subjected to a completion operation.
- the target binary image and the target frame image may be fused to obtain a model input image, wherein the target area in the target binary image includes each pixel whose pixel value is the target value.
- area, 0 means this pixel belongs to the target human body area.
- the target frame image and the target binary image are scaled according to a specific ratio (for example: 256/image long side), and then fused in the image processing channel, and the pixel value is normalized to (-1, 1 ) to obtain the model input image.
- a specific ratio for example: 256/image long side
- the image inpainting model it can be one model or a combination of multiple models, and the image inpainting model can be set on a cloud server or on a terminal device.
- the terminal device uploads the target frame image to the cloud server, and then uses the deep learning method to complete, and then returns the completed result image to the terminal device.
- uploading pictures to the cloud server will increase the time-consuming of interaction, and the network delay will also be uncertain, so it cannot meet the high real-time scene application.
- an image inpainting model eg, an image inpainting model including a first image inpainting model and a second image inpainting model
- the model input image is input to the first image inpainting model to generate a first inpainting model.
- Full image The model input image can be input into the lower-precision first image inpainting model to obtain a rough completion result map, that is, the first completed image.
- both the first image inpainting model and the second image inpainting model may be based on the UNet network structure.
- truncation may be performed pixel by pixel, that is, the pixel points with a pixel value less than -1 are set to -1, the pixel points with a pixel value greater than 1 are set to 1, and the pixel value is set to -1
- the pixel values of the pixels between and 1 remain the original value, and then the truncated result image, that is, the second complementary image, is input into the higher-precision second image inpainting model to obtain a revised complementary image.
- the full result image that is, the third complementary image.
- the user triggers by inputting a trigger instruction, so as to realize the target in response to the user's trigger instruction during the video recording process.
- the target video when using the terminal device to play the target video, in addition to the target person, the target video also includes the target background. command, target expression command, target text command, target body command, etc.) to trigger the target effect, thereby fixing the target frame image in the currently captured target video.
- the trigger instruction may be a trigger instruction input by a user, or a trigger instruction issued by a target object in the video, and may also include the terminal under preset conditions (for example, the video is played to a certain preset time) etc.) trigger commands.
- the above-mentioned description with the target person as the target object is only for the convenience of understanding.
- the target object may be in the form of a target person, a target animal, a target building, etc., which is not specifically limited here.
- the above trigger instruction may also be in the form of a target gesture command, a target voice command, a target facial expression command, a target body command, etc., which are also not specifically limited herein.
- Step 205 playing the first special effect sequence frame with the complementary frame image as the background, and the first special effect sequence frame is used to dynamically display special effect particles according to a preset path.
- the first special effect sequence frame may be played against the background of the complementary frame image, and the first special effect sequence frame is used to dynamically display special effect particles according to a preset path.
- the target frame image display the special effect particle effect
- the atmosphere can be enriched, and the effect of the target object can be displayed in a more natural and dynamic manner.
- the background completion operation can be performed on the target area continuously in the subsequent frame image of the target video, and the subsequent frame image is located in the completion frame image in the target video. after.
- the effect of the target object disappearing continuously in the target video is realized.
- FIG. 8 is a schematic structural diagram of a video processing apparatus according to an exemplary embodiment of the present disclosure. As shown in FIG. 8 , the video processing apparatus 300 provided in this embodiment includes:
- the image acquisition module 301 is used to fix the target frame image in the target video in response to the trigger instruction;
- the image processing module 302 is used to remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image, so as to generate and display the completion frame image, and the target area includes The vacant area after the target object is removed from the target frame image.
- the image processing module 302 is further configured to play a first special effect sequence frame with the complementary frame image as a background, where the first special effect sequence frame is used to follow a preset path Dynamically display effect particles.
- the image processing module 302 is further configured to display a preset second special effect, and the preset second special effect is used to make the target frame image display a visual blur effect.
- the image processing module 302 is further configured to display a preset third special effect, where the preset third special effect is used to make the target frame image display a visual shaking effect.
- the image processing module 302 is further configured to continuously perform a completion operation on the target area in a subsequent frame image of the target video, and the subsequent frame image is in the The target video is located after the complementary frame image.
- the image processing module 302 is further configured to determine that the target object is a target type object.
- the trigger instruction includes at least one of a target gesture instruction, a target voice instruction, a target facial expression instruction, a target body instruction, and a target text instruction.
- the image processing module 302 is configured to:
- the target area of the target object in the target frame image is determined according to the target binary image, wherein the target area in the target binary image includes each pixel whose pixel value is a target value.
- the image processing module 302 is configured to:
- the target binary image is fused with the target frame image, and the model input image is obtained;
- the model input image is input to an image inpainting model to generate the complement frame image.
- the image inpainting model is provided in a terminal device, and the terminal device is configured to record and/or play the target video for the target object.
- the image processing module 302 is specifically configured to:
- Pixel truncation is performed on the third complementary image by using the preset pixel threshold to generate a fourth complementary image, wherein the processed frame image includes the fourth complementary image.
- video processing apparatus provided by the embodiment shown in FIG. 8 can be used to execute the method steps provided by any of the above method embodiments, and the specific implementation manner and technical effect are similar, and details are not repeated here.
- FIG. 9 is a schematic structural diagram of an electronic device according to an exemplary embodiment of the present disclosure. As shown in FIG. 9 , it shows a schematic structural diagram of an electronic device 400 suitable for implementing an embodiment of the present disclosure.
- Terminal devices in the embodiments of the present disclosure may include, but are not limited to, such as mobile phones, notebook computers, digital broadcast receivers, PDAs (Personal Digital Assistant, personal digital assistants), PADs (Portable Android Device, tablet computers), PMPs (Portable Media Player, portable multimedia player), in-vehicle terminals (such as in-vehicle navigation terminals) and other mobile terminals with image acquisition functions, and fixed terminals with image acquisition devices such as digital TVs (Television, television sets), desktop computers, and the like.
- the electronic device shown in FIG. 9 is only an example, and should not impose any limitation on the function and scope of use of the embodiments of the present disclosure.
- the electronic device 400 may include a processor (eg, a central processing unit, a graphics processing unit, etc.) 401 , which may be based on a program stored in a read-only memory (Read-Only Memory, ROM) 402 or from a memory 408
- ROM read-only memory
- RAM random access memory
- various programs and data required for the operation of the electronic device 400 are also stored.
- the processor 401, the ROM 402, and the RAM 403 are connected to each other through a bus 404.
- An Input/Output (I/O) interface 405 is also connected to the bus 404 .
- the storage device is used for storing a program for executing the video processing methods described in the above method embodiments; the processor is configured to execute the program stored in the memory.
- the following devices can be connected to the I/O interface 405: input devices 406 including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a Liquid Crystal Display (LCD) 407 output means 407, speakers, vibrators, etc.; memory 408 including, eg, magnetic tape, hard disk, etc.
- Communication means 409 may allow electronic device 400 to communicate wirelessly or by wire with other devices to exchange data.
- FIG. 9 shows the electronic device 400 having various means, it should be understood that not all of the illustrated means are required to be implemented or provided, and more or fewer means may be implemented or provided instead.
- embodiments of the present disclosure include a computer-readable storage medium comprising a computer program carried on a non-transitory computer-readable medium, the computer program containing the video shown in the flowcharts for implementing the embodiments of the present disclosure
- the program code for the handler method may be downloaded and installed from a network via communication means 409, or from memory 408, or from ROM 402.
- the computer program is executed by the processor 401, the above-mentioned video processing functions defined in the methods of the embodiments of the present disclosure are executed.
- the above-mentioned computer-readable storage medium of the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two.
- the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, or any combination of the above.
- Computer readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable Erasable Programmable Read-Only Memory (EPROM or flash memory), optical fiber, Portable Compact Disk Read-Only Memory (CD-ROM), optical storage device, magnetic storage device, or the above any suitable combination.
- a computer-readable storage medium can be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
- a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave with computer-readable program code embodied thereon.
- Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
- a computer-readable signal medium can also be any computer-readable medium other than a computer-readable storage medium that can transmit, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device .
- the program code contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to: electric wire, optical cable, RF (Radio Frequency, radio frequency), etc., or any suitable combination of the above.
- the above-mentioned computer-readable storage medium may be included in the above-mentioned electronic device; or may exist alone without being assembled into the electronic device.
- the above-mentioned computer-readable storage medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device: in response to the trigger instruction, fix the target frame image of the target video; remove the target frame The target object in the image, and the completion operation is performed on the target area in the target frame image to generate and display the completion frame image, and the target area is the vacant area after the target object is removed from the target frame image.
- Computer program code for performing operations of the present disclosure may be written in one or more programming languages, including but not limited to object-oriented programming languages—such as Java, Smalltalk, C++, and This includes conventional procedural programming languages - such as the "C" language or similar programming languages.
- the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer can be connected to the user's computer through any kind of network—including a Local Area Network (LAN) or a Wide Area Network (WAN)—or, can be connected to an external computer ( For example, using an Internet service provider to connect via the Internet).
- LAN Local Area Network
- WAN Wide Area Network
- the client and server can use any currently known or future developed network protocol such as HTTP (HyperText Transfer Protocol) to communicate, and can communicate with digital data in any form or medium Communication (eg, a communication network) interconnects.
- HTTP HyperText Transfer Protocol
- Examples of communication networks include local area networks (“LAN”), wide area networks (“WAN”), the Internet (eg, the Internet), and peer-to-peer networks (eg, ad hoc peer-to-peer networks), as well as any currently known or future development network of.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of code that contains one or more logical functions for implementing the specified functions executable instructions.
- the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or operations , or can be implemented in a combination of dedicated hardware and computer instructions.
- the modules involved in the embodiments of the present disclosure may be implemented in software or hardware.
- the name of the module does not constitute a limitation of the unit itself in some cases, for example, the display module can also be described as "a unit that displays the face of the object and the sequence of face masks".
- exemplary types of hardware logic components include: Field-Programmable Gate Arrays (FPGAs), Application Specific Integrated Circuits (ASICs), Application Specific Standard Products (Application Specific Integrated Circuits) Specific Standard Product (ASSP), System-on-a-chip (SOC), Complex Programmable Logic Device (CPLD), etc.
- FPGAs Field-Programmable Gate Arrays
- ASICs Application Specific Integrated Circuits
- ASSP Application Specific Integrated Circuits
- SOC System-on-a-chip
- CPLD Complex Programmable Logic Device
- a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with the instruction execution system, apparatus or device.
- the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
- Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices, or devices, or any suitable combination of the foregoing.
- machine-readable storage media would include one or more wire-based electrical connections, portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), fiber optics, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
- RAM random access memory
- ROM read only memory
- EPROM or flash memory erasable programmable read only memory
- CD-ROM compact disk read only memory
- magnetic storage or any suitable combination of the foregoing.
- a video processing method including:
- the method further includes:
- the first special effect sequence frame is used to dynamically display special effect particles according to a preset path.
- the method further includes:
- a preset second special effect is displayed, and the preset second special effect is used to make the target frame image display a visual blur effect.
- the method further includes:
- a preset third special effect is displayed, and the preset third special effect is used to make the target frame image display a visual shaking effect.
- the method further includes:
- the completion operation is continuously performed on the target area in a subsequent frame image of the target video, and the subsequent frame image is located after the completion frame image in the target video.
- the method before acquiring the trigger instruction, the method further includes:
- the target object is determined to be a target type object.
- the trigger instruction includes at least one of a target gesture instruction, a target voice instruction, a target facial expression instruction, a target body instruction, and a target text instruction.
- removing the target object in the target frame image, and performing a completion operation on the target area in the target frame image further includes:
- the target area of the target object in the target frame image is determined according to the target binary image, wherein the target area in the target binary image includes each pixel whose pixel value is a target value.
- the performing background completion on the target area in the target frame image includes:
- the target binary image is fused with the target frame image, and the model input image is obtained;
- the target area in the target frame image is replaced with the target area in the processed frame image to generate the complement frame image.
- the image inpainting model is provided in a terminal device, and the terminal device is used to record the target video for the target object.
- inputting the model input image to an image inpainting model to generate a processing frame image model input image completion operation includes:
- Pixel truncation is performed on the third complementary image by using the preset pixel threshold to generate a fourth complementary image, wherein the processed frame image includes the fourth complementary image.
- a video processing apparatus including:
- an image acquisition module in response to a trigger instruction, to fix the target frame image of the target video
- the image processing module is used to remove the target object in the target frame image, and perform a completion operation on the target area in the target frame image, so as to generate and display the completion frame image, the target area Including the vacant area after the target object is removed from the target frame image.
- the image processing module is further configured to play a first special effect sequence frame with the complementary frame image as a background, and the first special effect sequence frame is used to dynamically move according to a preset path Display effect particles.
- the image processing module is further configured to display a preset second special effect, and the preset second special effect is used to make the target frame image display a visual blur effect.
- the image processing module is further configured to display a preset third special effect, where the preset third special effect is used to make the target frame image display a visual shaking effect.
- the image processing module is further configured to continuously perform a completion operation on the target area in a subsequent frame image of the target video, where the subsequent frame image is in the target video After the completion frame image in the video.
- the image processing module is further configured to determine that the target object is a target type object.
- the trigger instruction includes at least one of a target gesture instruction, a target voice instruction, a target facial expression instruction, a target body instruction, and a target text instruction.
- the image processing module is specifically configured to:
- the target area of the target object in the target frame image is determined according to the target binary image, wherein the target area in the target binary image includes each pixel whose pixel value is a target value.
- the image processing module is specifically configured to:
- the target binary image is fused with the target frame image, and the model input image is obtained;
- the model input image is input to an image inpainting model to generate the complement frame image.
- the image inpainting model is provided in a terminal device, and the terminal device is configured to record and/or play the target video for the target object.
- the image processing module is specifically configured to:
- Pixel truncation is performed on the third complementary image by using the preset pixel threshold to generate a fourth complementary image, and the processed frame image includes the fourth complementary image.
- an electronic device including:
- a memory for storing a computer program for the processor
- a display for displaying the video processed by the processor
- the processor is configured to implement the video processing method described above in the first aspect and various possible designs of the first aspect by executing the computer program.
- embodiments of the present disclosure provide a computer-readable storage medium, where computer-executable instructions are stored in the computer-readable storage medium, and when a processor executes the computer-executable instructions, the first aspect and the first Aspects of the video processing methods described in various possible designs.
- an embodiment of the present disclosure provides a computer program product, the program product includes: a computer program, which implements the steps of the method described in any embodiment of the present disclosure when the computer program is executed by a processing device.
- a computer program product includes: a computer program, which implements the steps of the method described in any embodiment of the present disclosure when the computer program is executed by a processing device.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Processing Or Creating Images (AREA)
- User Interface Of Digital Computer (AREA)
- Studio Devices (AREA)
Abstract
Description
Claims (15)
- 一种视频处理方法,其特征在于,包括:响应于触发指令,固定目标视频中的目标帧图像;移除所述目标帧图像中目标对象,并对所述目标帧图像中的目标区域进行补全操作,以生成并显示补全帧图像,所述目标区域包括将所述目标对象从所述目标帧图像中移除后的空缺区域。
- 根据权利要求1所述的视频处理方法,其特征在于,所述生成并显示补全帧图像,还包括:以所述补全帧图像为背景播放第一特效序列帧,所述第一特效序列帧用于按照预设路径动态显示特效粒子。
- 根据权利要求1所述的视频处理方法,其特征在于,在所述固定所述目标视频的目标帧图像之后,还包括:显示预设第二特效,所述预设第二特效用于使得所述目标帧图像展示视觉模糊效果。
- 根据权利要求1所述的视频处理方法,其特征在于,在所述固定所述目标视频的目标帧图像之后,还包括:显示预设第三特效,所述预设第三特效用于使得所述目标帧图像展示视觉晃动效果。
- 根据权利要求1所述的视频处理方法,其特征在于,在生成并显示补全帧图像之后,还包括:在所述目标视频的后续帧图像中持续对所述目标区域进行补全操作,所述后续帧图像在所述目标视频中位于所述补全帧图像之后。
- 根据权利要求1-5中任意一项所述的视频处理方法,其特征在于,在获取到所述触发指令之前,还包括:确定所述目标对象为目标类型对象。
- 根据权利要求1-5中任意一项所述的视频处理方法,其特征在于,所述触发指令包括:目标手势指令、目标语音指令、目标表情指令、目标肢体指令、目标文字指令中的至少一种。
- 根据权利要求1-5中任意一项所述的视频处理方法,其特征在于,所述移除所述目标帧图像中所述目标对象,并对所述目标帧图像中的目标区域进行补全操作,包括:利用预设对象分割模型对所述目标帧图像中各个像素点进行识别,以生成与所述目标帧图像同尺寸的目标二值图像;根据所述目标二值图像确定所述目标对象在所述目标帧图像中的所述目标区域,其中,所述目标二值图像中的所述目标区域包括像素值为目标值的各个像素点。
- 根据权利要求8所述的视频处理方法,其特征在于,所述对所述目标帧图像中的所述目标区域进行补全操作,包括:将所述目标二值图像与所述目标帧图像进行融合,并得到模型输入图像;将所述模型输入图像输入至图像修补模型,以生成处理帧图像;利用所述处理帧图像中的所述目标区域替换所述目标帧图像中所述目标区域,以生成所述补全帧图像。
- 根据权利要求9所述的视频处理方法,其特征在于,所述图像修补模型设置于终端设备,所述终端设备基于所述图像修补模型对所述目标视频进行处理。
- 根据权利要求9所述的视频处理方法,其特征在于,将所述模型输入图像输入至图像修补模型,以生成处理帧图像,包括:将所述模型输入图像输入至第一图像修补模型,以生成第一补全图像;利用预设像素阈值对所述第一补全图像进行像素截断,以生成第二补全图像;将所述第二补全图像输入至第二图像修补模型,以生成第三补全图像,其中,所述第二图像修补模型的补全精度大于所述第一图像修补模型;利用所述预设像素阈值对所述第三补全图像进行像素截断,以生成第四补全图像,其中,所述处理帧图像包括所述第四补全图像。
- 一种视频处理装置,其特征在于,包括:图像获取模块,用于响应于触发指令,固定目标视频中的目标帧图像;图像处理模块,用于移除所述目标帧图像中目标对象,并对目标帧图像中的目标区域进行补全操作,以生成并显示补全帧图像,所述目标区域包括将所述目标对象从所述目标帧图像中移除后的空缺区域。
- 一种电子设备,其特征在于,包括:处理器;以及存储器,用于存储计算机程序;显示器,用于显示经所述处理器处理后的视频;其中,所述处理器被配置为通过执行所述计算机程序来实现权利要求1-11任一项所述的视频处理方法。
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如权利要求1-11任一项所述的视频处理方法。
- 一种计算机程序产品,包括计算机程序,所述计算机程序在被处理装置执行 时实现权利要求1-11中任一项所述的视频处理方法。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/253,186 US20230421716A1 (en) | 2020-11-16 | 2021-11-15 | Video processing method and apparatus, electronic device and storage medium |
JP2023528594A JP2023549841A (ja) | 2020-11-16 | 2021-11-15 | ビデオ処理方法、装置、電子機器及び記憶媒体 |
EP21891255.8A EP4243398A4 (en) | 2020-11-16 | 2021-11-15 | VIDEO PROCESSING METHOD AND APPARATUS, ELECTRONIC DEVICE AND STORAGE MEDIUM |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011280804.3 | 2020-11-16 | ||
CN202011280804.3A CN112637517B (zh) | 2020-11-16 | 2020-11-16 | 视频处理方法、装置、电子设备及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022100735A1 true WO2022100735A1 (zh) | 2022-05-19 |
Family
ID=75303510
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/130708 WO2022100735A1 (zh) | 2020-11-16 | 2021-11-15 | 视频处理方法、装置、电子设备及存储介质 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20230421716A1 (zh) |
EP (1) | EP4243398A4 (zh) |
JP (1) | JP2023549841A (zh) |
CN (1) | CN112637517B (zh) |
WO (1) | WO2022100735A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115119014A (zh) * | 2022-06-15 | 2022-09-27 | 马上消费金融股份有限公司 | 视频处理方法、插帧数量模型的训练方法及装置 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112637517B (zh) * | 2020-11-16 | 2022-10-28 | 北京字节跳动网络技术有限公司 | 视频处理方法、装置、电子设备及存储介质 |
CN115914497A (zh) * | 2021-08-24 | 2023-04-04 | 北京字跳网络技术有限公司 | 视频处理方法、装置、设备、介质及程序产品 |
CN114040129B (zh) * | 2021-11-30 | 2023-12-05 | 北京字节跳动网络技术有限公司 | 视频生成方法、装置、设备及存储介质 |
CN114782284B (zh) * | 2022-06-17 | 2022-09-23 | 广州三七极耀网络科技有限公司 | 动作数据修正方法、装置、设备及存储介质 |
CN116503289B (zh) * | 2023-06-20 | 2024-01-09 | 北京天工异彩影视科技有限公司 | 一种视觉特效应用处理方法和系统 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140219555A1 (en) * | 2013-02-01 | 2014-08-07 | National Yunlin University Of Science And Technology | Method for detecting and removing scrolling texts during video communication |
CN110728639A (zh) * | 2019-09-29 | 2020-01-24 | 三星电子(中国)研发中心 | 图片修复方法及系统 |
CN111179159A (zh) * | 2019-12-31 | 2020-05-19 | 北京金山云网络技术有限公司 | 消除视频中目标影像的方法、装置、电子设备及存储介质 |
CN111260537A (zh) * | 2018-12-03 | 2020-06-09 | 珠海格力电器股份有限公司 | 一种图像隐私保护方法、装置、存储介质及摄像设备 |
CN111416939A (zh) * | 2020-03-30 | 2020-07-14 | 咪咕视讯科技有限公司 | 一种视频处理方法、设备及计算机可读存储介质 |
CN112188058A (zh) * | 2020-09-29 | 2021-01-05 | 努比亚技术有限公司 | 一种视频拍摄方法、移动终端以及计算机存储介质 |
CN112637517A (zh) * | 2020-11-16 | 2021-04-09 | 北京字节跳动网络技术有限公司 | 视频处理方法、装置、电子设备及存储介质 |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101105034B1 (ko) * | 2010-02-09 | 2012-01-16 | 주식회사 팬택 | 촬영 기능을 보유한 단말기 |
JP2014096661A (ja) * | 2012-11-08 | 2014-05-22 | International Business Maschines Corporation | 動画撮影中において動画中の動体オブジェクトをリアルタイムに隠消するための方法、並びに、その動画撮影機器及び当該動画撮影機器のためのプログラム |
US9898856B2 (en) * | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
US9514523B2 (en) * | 2014-11-18 | 2016-12-06 | Intel Corporation | Method and apparatus for filling images captured by array cameras |
CN104574311B (zh) * | 2015-01-06 | 2017-08-11 | 华为技术有限公司 | 图像处理方法和装置 |
CN104680487A (zh) * | 2015-01-21 | 2015-06-03 | 浙江大学 | 基于低秩矩阵恢复的非局部图像修复方法 |
CN106385591B (zh) * | 2016-10-17 | 2020-05-15 | 腾讯科技(上海)有限公司 | 视频处理方法及视频处理装置 |
US10613726B2 (en) * | 2017-12-22 | 2020-04-07 | Adobe Inc. | Removing and replacing objects in images according to a directed user conversation |
CN109215091B (zh) * | 2018-08-02 | 2021-09-14 | 浙江理工大学 | 一种基于图表示的服装流行色彩自动提取方法 |
CN111161275B (zh) * | 2018-11-08 | 2022-12-23 | 腾讯科技(深圳)有限公司 | 医学图像中目标对象的分割方法、装置和电子设备 |
CN111353946B (zh) * | 2018-12-21 | 2023-04-11 | 腾讯科技(深圳)有限公司 | 图像修复方法、装置、设备及存储介质 |
CN110225246B (zh) * | 2019-05-28 | 2021-07-20 | 北京字节跳动网络技术有限公司 | 事件脚本的生成方法、装置、电子设备及计算机可读存储介质 |
CN111444921A (zh) * | 2020-03-25 | 2020-07-24 | 浙江华睿科技有限公司 | 划痕缺陷检测方法、装置、计算设备和存储介质 |
CN111556278B (zh) * | 2020-05-21 | 2022-02-01 | 腾讯科技(深圳)有限公司 | 一种视频处理的方法、视频展示的方法、装置及存储介质 |
CN111754528B (zh) * | 2020-06-24 | 2024-07-12 | Oppo广东移动通信有限公司 | 人像分割方法、装置、电子设备和计算机可读存储介质 |
CN111832538A (zh) * | 2020-07-28 | 2020-10-27 | 北京小米松果电子有限公司 | 视频处理方法及装置、存储介质 |
-
2020
- 2020-11-16 CN CN202011280804.3A patent/CN112637517B/zh active Active
-
2021
- 2021-11-15 JP JP2023528594A patent/JP2023549841A/ja active Pending
- 2021-11-15 US US18/253,186 patent/US20230421716A1/en active Pending
- 2021-11-15 EP EP21891255.8A patent/EP4243398A4/en active Pending
- 2021-11-15 WO PCT/CN2021/130708 patent/WO2022100735A1/zh active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140219555A1 (en) * | 2013-02-01 | 2014-08-07 | National Yunlin University Of Science And Technology | Method for detecting and removing scrolling texts during video communication |
CN111260537A (zh) * | 2018-12-03 | 2020-06-09 | 珠海格力电器股份有限公司 | 一种图像隐私保护方法、装置、存储介质及摄像设备 |
CN110728639A (zh) * | 2019-09-29 | 2020-01-24 | 三星电子(中国)研发中心 | 图片修复方法及系统 |
CN111179159A (zh) * | 2019-12-31 | 2020-05-19 | 北京金山云网络技术有限公司 | 消除视频中目标影像的方法、装置、电子设备及存储介质 |
CN111416939A (zh) * | 2020-03-30 | 2020-07-14 | 咪咕视讯科技有限公司 | 一种视频处理方法、设备及计算机可读存储介质 |
CN112188058A (zh) * | 2020-09-29 | 2021-01-05 | 努比亚技术有限公司 | 一种视频拍摄方法、移动终端以及计算机存储介质 |
CN112637517A (zh) * | 2020-11-16 | 2021-04-09 | 北京字节跳动网络技术有限公司 | 视频处理方法、装置、电子设备及存储介质 |
Non-Patent Citations (1)
Title |
---|
See also references of EP4243398A4 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115119014A (zh) * | 2022-06-15 | 2022-09-27 | 马上消费金融股份有限公司 | 视频处理方法、插帧数量模型的训练方法及装置 |
CN115119014B (zh) * | 2022-06-15 | 2024-01-30 | 马上消费金融股份有限公司 | 视频处理方法、插帧数量模型的训练方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
CN112637517A (zh) | 2021-04-09 |
EP4243398A4 (en) | 2024-04-17 |
CN112637517B (zh) | 2022-10-28 |
EP4243398A1 (en) | 2023-09-13 |
JP2023549841A (ja) | 2023-11-29 |
US20230421716A1 (en) | 2023-12-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022100735A1 (zh) | 视频处理方法、装置、电子设备及存储介质 | |
WO2023051185A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
CN110898429B (zh) | 一种游戏剧情展现方法、装置、电子设备及存储介质 | |
CN112165632B (zh) | 视频处理方法、装置及设备 | |
CN110070496B (zh) | 图像特效的生成方法、装置和硬件装置 | |
WO2022048504A1 (zh) | 视频处理方法、终端设备及存储介质 | |
WO2021254502A1 (zh) | 目标对象显示方法、装置及电子设备 | |
US20230182028A1 (en) | Game live broadcast interaction method and apparatus | |
CN109600559B (zh) | 一种视频特效添加方法、装置、终端设备及存储介质 | |
US12019669B2 (en) | Method, apparatus, device, readable storage medium and product for media content processing | |
US20220159197A1 (en) | Image special effect processing method and apparatus, and electronic device and computer readable storage medium | |
US12041379B2 (en) | Image special effect processing method, apparatus, and electronic device, and computer-readable storage medium | |
WO2023226814A1 (zh) | 视频处理方法、装置、电子设备及存储介质 | |
WO2023103720A1 (zh) | 视频特效处理方法、装置、电子设备及程序产品 | |
WO2023040749A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
CN114445600A (zh) | 一种特效道具的展示方法、装置、设备及存储介质 | |
CN110069641B (zh) | 图像处理方法、装置和电子设备 | |
CN115278355B (zh) | 视频剪辑方法、装置、设备、计算机可读存储介质及产品 | |
WO2022237435A1 (zh) | 更换画面中的背景的方法、设备、存储介质及程序产品 | |
CN111696041B (zh) | 图像处理方法、装置及电子设备 | |
US11805219B2 (en) | Image special effect processing method and apparatus, electronic device and computer-readable storage medium | |
CN113837918A (zh) | 多进程实现渲染隔离的方法及装置 | |
WO2022213798A1 (zh) | 图像处理方法、装置、电子设备和存储介质 | |
US20240320256A1 (en) | Method, apparatus, device, readable storage medium and product for media content processing | |
CN112053450B (zh) | 文字的显示方法、装置、电子设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21891255 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023528594 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18253186 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2021891255 Country of ref document: EP Effective date: 20230616 |