WO2023011146A1 - 视频处理方法、装置、设备及存储介质 - Google Patents

视频处理方法、装置、设备及存储介质 Download PDF

Info

Publication number
WO2023011146A1
WO2023011146A1 PCT/CN2022/106110 CN2022106110W WO2023011146A1 WO 2023011146 A1 WO2023011146 A1 WO 2023011146A1 CN 2022106110 W CN2022106110 W CN 2022106110W WO 2023011146 A1 WO2023011146 A1 WO 2023011146A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
video
texture
processed
video processing
Prior art date
Application number
PCT/CN2022/106110
Other languages
English (en)
French (fr)
Inventor
任龙
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Priority to EP22851860.1A priority Critical patent/EP4344229A1/en
Publication of WO2023011146A1 publication Critical patent/WO2023011146A1/zh
Priority to US18/394,685 priority patent/US20240129576A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • H04N21/4316Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/761Proximity, similarity or dissimilarity measures
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4318Generation of visual interfaces for content selection or interaction; Content or additional data rendering by altering the content in the rendering process, e.g. blanking, blurring or masking an image region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • H04N21/4542Blocking scenes or portions of the received content, e.g. censoring scenes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Definitions

  • Embodiments of the present disclosure relate to the technical field of video processing, and in particular, to a video processing method, device, device, electronic device, storage medium, computer program product, and computer program.
  • Embodiments of the present disclosure provide a video processing method, device, device, storage medium, electronic device, computer program product, and computer program.
  • an embodiment of the present disclosure provides a video processing method, including: displaying the video to be processed and at least one texture in the video processing interface; receiving a selection operation for the target texture, and identifying the target part of the target object in the video to be processed ;Display the target texture on the target part in the video to be processed; receive the editing operation on the target texture, modify the target texture; generate the target video, and display the target texture on the target part in the target video.
  • an embodiment of the present disclosure provides a video processing device, including: a display module, which displays a video to be processed and at least one texture in a video processing interface; an identification module, which receives a selection operation for a target texture, and identifies the video to be processed The target part of the target object; the display module is also used to display the target texture on the target part in the video to be processed; the processing module is used to receive the editing operation acting on the target texture, modify the target texture, generate the target video, and target A target decal is shown on the target part in the video.
  • an embodiment of the present disclosure provides an electronic device, including: at least one processor and a memory; the memory stores computer-executable instructions; at least one processor executes the computer-executable instructions stored in the memory, so that at least one processor executes the first Aspects of video processing methods.
  • an embodiment of the present disclosure provides a computer-readable storage medium, in which computer-executable instructions are stored, and when the processor executes the computer-executable instructions, the video processing method in the first aspect is implemented.
  • an embodiment of the present disclosure provides a computer program product, including: a computer program, wherein when the computer program is executed by a processor, the video processing method in the first aspect is implemented.
  • an embodiment of the present disclosure provides a computer program, which implements the video processing method in the first aspect when the computer program is executed by a processor.
  • FIG. 1 is a schematic diagram of an application scenario of a video processing method provided by an embodiment of the present disclosure
  • FIG. 2 is a first schematic flowchart of a video processing method provided by an embodiment of the present disclosure
  • FIG. 3 is a first schematic diagram of a video processing interface provided by an embodiment of the present disclosure
  • FIG. 4 is a second schematic flow diagram of a video processing method provided by an embodiment of the present disclosure.
  • FIG. 5 is a second schematic diagram of a video processing interface provided by an embodiment of the present disclosure.
  • FIG. 6 is a third schematic flowchart of a video processing method provided by an embodiment of the present disclosure.
  • FIG. 7 is a third schematic diagram of a video processing interface provided by an embodiment of the present disclosure.
  • FIG. 8 is a fourth schematic diagram of a video processing interface provided by an embodiment of the present disclosure.
  • FIG. 9 is a fifth schematic diagram of a video processing interface provided by an embodiment of the present disclosure.
  • FIG. 10 is a schematic diagram of a video processing device provided by an embodiment of the present disclosure.
  • Fig. 11 is a schematic diagram of an electronic device provided by an embodiment of the present disclosure.
  • the technical concept of the present disclosure is to automatically identify the target part to be occluded in the video to be processed after the user selects a target texture for the target part during video processing, and display the texture on the target part , so as to achieve occlusion, and set the editing control on the target texture, so as to realize the editing of the added target texture. Since there is no need to add textures manually, the operation of textures can be simplified and the occlusion effect of textures can be improved. In addition, after adding textures, the textures can also be edited and adjusted, which can further improve the flexibility and convenience of video processing and improve user experience. .
  • FIG. 1 For the convenience of understanding, the application scenario of the embodiment of the present disclosure is first described in conjunction with FIG. 1:
  • FIG. 1 is a schematic diagram of an application scenario of a video processing method provided by an embodiment of the present disclosure.
  • the scenario includes: a terminal device 100 .
  • the terminal device 100 may be a mobile terminal (such as a smart phone, a handheld computer, a tablet computer and other mobile devices with video processing functions, and a wearable electronic device), a computer device (such as a desktop computer, a notebook computer, an all-in-one computer, a smart phone, etc.) Household equipment, etc.), etc.
  • a smartphone is taken as an example in FIG. 1 , but it is not limited thereto.
  • the user can import the video to be processed into the terminal device 100 , so that the video to be processed can be processed through the terminal device 100 .
  • the scene may further include a server 200 , and the terminal device 100 may communicate with the server 200 through a wireless or wired network to send data to the server 200 .
  • the wireless network may be a communication network such as 2G or 3G or 4G or 5G, or may be a wireless local area network, which is not limited here.
  • the user triggers operations such as video upload, video editing, and adding stickers through the terminal device 100, and then sends the operation data to the server 200, and the server 200 performs corresponding video processing operations.
  • Figure 1 is only a schematic diagram of an application scenario provided by the embodiment of the present disclosure, and the embodiment of the present disclosure does not limit the type and number of devices included in Figure 1 , nor does it limit the positions of the devices in Figure 1 relationship, for example, in the application scenario shown in FIG.
  • the server 200 may be an independent server, or may also be a service cluster or the like.
  • FIG. 2 is a first schematic flowchart of a video processing method provided by an embodiment of the present disclosure.
  • the video processing method provided in this embodiment can be applied to the terminal device or server shown in FIG. 1.
  • the video processing method provided in this embodiment includes the following steps:
  • S201 Display a video to be processed and at least one texture in a video processing interface.
  • the video to be processed may be uploaded by the user, or the video to be processed may also be collected by the user in real time through an image collection device of the terminal device, such as a camera and other components.
  • the embodiment of the present disclosure does not specifically limit the specific type of texture.
  • it can be a static texture or a dynamic texture, such as an animal image, an animation image, a still life image, etc.
  • the target texture is used for occlusion At least a partial area of the target site.
  • FIG. 3 is a first schematic diagram of a video processing interface provided by an embodiment of the present disclosure. It should be noted that the target part of the target object in FIG. 3 is shown as an example of a face, but it is not limited thereto. As shown in Figure 3(a), four types of stickers are displayed on the video processing interface: Sticker 1, Sticker 2, Sticker 3, and Sticker 4. Each sticker type contains multiple stickers. Users can slide up and down or Swipe left and right to select the target sticker from various stickers.
  • S202 Receive a selection operation on the target texture, and identify a target part of the target object in the video to be processed.
  • the target object can be any object that needs to add a texture, such as a person or an animal.
  • the target part can be any part on the target object.
  • the target object may be a face, hands, hair, etc., or a local area of the above-mentioned parts, such as one or more areas such as eyes and mouth.
  • the user can select the target texture by clicking on the texture.
  • the target part of the target object in the video to be processed is identified.
  • the target part of the target object is fixed, that is, during the video processing, all textures are used to block a certain type of part.
  • all the textures are used to block the target part of the person, such as the hand.
  • the hand in the image to be processed is automatically identified.
  • the user before entering the video processing interface, can set the target object and/or target part to be processed in this video processing process. For example, if the target object set by the user is a character, the target part is hands, all maps are used to occlude the hands during this video processing.
  • the first area of the video editing page displays one or more images to be processed of the target object
  • the second area of the video editing page displays multiple target textures.
  • the user can select the first target Map and drag the first target texture to the first target part of the first target object, then select the second target texture and drag the second target texture to the second target part of the first target object or the second target object.
  • the first target texture is added to all video frames including the first target part in the video, and the first target texture in each video frame blocks the first target part, add the second target texture to all video frames including the second target part in the video, and the second target texture in each video frame blocks the second target part.
  • the first target object and the second target object are any target objects in the plurality of target objects respectively.
  • it is not limited to adding two textures at the same time, so I won't repeat them here.
  • multiple textures can be added to the video at the same time, and each texture can be matched to the area of each frame in the video including the target part, which enriches video interaction methods and improves video editing efficiency.
  • the target part of the target object in the video to be processed can also be identified and added according to the type of the target map (shape type, color type, etc.), specifically, in response to receiving a selection operation for the target map, identify Pending the target part in the image that matches the target map type and adding the target map to the target part.
  • the type of the first target texture is the elongated style
  • the arm in the image to be processed is also the elongated style
  • the arm is determined as the first target part
  • the type of the second target texture is the circular style
  • the recognition is made If the face in the processed image is also in a circular pattern, the face is determined as the second target part.
  • sticker 1 is a sticker for covering hair
  • sticker 2 is a sticker for covering animals
  • sticker 4 is a sticker for covering hands.
  • the texture is an editable texture.
  • the editing control is a control for editing the attributes of the map elements, and the attributes of the map elements include brightness attributes, chroma attributes, contrast attributes, aspect ratio attributes, etc., and the corresponding editing controls include brightness adjustment controls and chroma adjustment controls.
  • a contrast adjustment control, a ratio adjustment control, and the texture element is a texture pattern, such as a dog head pattern, a flower pattern, a beverage bottle pattern, and the like.
  • this disclosure adopts the editable textures with exposed display editing controls around the texture elements, which greatly meets the user's needs for rapid adjustment of textures, and is conducive to improving the matching degree between the target texture and the target part.
  • the interactive path improves editing efficiency.
  • the target texture is displayed on the target part.
  • the faces of all characters in the video to be processed display "sticker No. 6".
  • corresponding prompt information can also be displayed in the video processing interface, wherein the prompt information is used to prompt at least one of the following information: prompting that the current identification of the target object fails, prompting the user to manually select the target part, prompting The user uploads a new video to be processed, and the user is prompted to adjust the progress of the video to be processed to a position including the target part.
  • S204 Receive an editing operation acting on the target texture, and modify the target texture.
  • the modification operation includes at least one of brightness adjustment operation, hue adjustment operation, contrast adjustment operation, scale adjustment operation, size adjustment operation, deletion operation, replacement operation and direction adjustment operation.
  • receiving an editing operation to act on a target texture may be a triggering operation to act on an editing control displayed on the target texture.
  • modification controls are displayed on the target texture.
  • the modification controls include: size adjustment controls, deletion controls, replacement controls, and direction adjustment controls, etc., and the user can adjust the target texture by modifying the controls. Be specific.
  • the target texture on the target part is deleted, and prompt information is displayed in the pending interface, wherein the information is used to prompt the user to re-add the target texture for the target object.
  • the modification controls as shown in Figure 3 (b) can also be displayed in the video processing interface, for example, delete controls and replace controls, etc., so that the user can click the interface
  • the modify controls in the target map are manipulated accordingly.
  • the target video is generated according to the video to be processed after the target texture is added, wherein the target part in the target video displays the target texture added by the user.
  • the method of generating the target video is not specifically limited in the embodiments of the present disclosure.
  • a control for generating the target video may be synchronously displayed on the video processing interface.
  • the user clicks on the control That is, the target video is generated, or the target video can be directly generated after the target texture is displayed on the target part.
  • step S204 is an optional step, that is, after step S203, if the user does not modify the target texture, the target video can be generated according to the target texture added by the user; if the user modifies the target texture In the modification operation, the target video is generated according to the modified target texture. At this time, the target part in the generated target video displays the modified target texture.
  • the user does not need to manually select, which simplifies the map operation and improves the occlusion effect of the map.
  • the map can also be edited Adjustment, so as to further improve the flexibility and convenience in the video processing process, and improve user experience.
  • the embodiment of the present disclosure does not specifically limit the solution of identifying the target part of the target object in the video to be processed.
  • image recognition technology may be used to detect the target part of the target object in the video to be processed.
  • the user can also manually select the target object to identify the target part of the target object.
  • the user can click to select the target object in the video to be processed.
  • receive The user selects the target object in the video to be processed to identify the target part of the target object.
  • the target part of the target object can also be directly manually selected by the user.
  • the user's selection operation of the target part in the video to be processed is received to detect the target part in the video to be processed.
  • the user can Click to select the target part in the video to be processed.
  • the user's selection operation of the target part in the video to be processed is received to identify the target part.
  • multiple methods for selecting target parts are provided, which can improve the flexibility of the video processing process and improve user experience.
  • FIG. 4 is a second schematic flowchart of a video processing method provided by an embodiment of the present disclosure. As shown in Figure 4, the video processing method includes the following steps:
  • S401 Display a video to be processed and at least one texture in a video processing interface.
  • S402. Receive a selection operation for multiple different target textures, and identify multiple target parts corresponding to at least one target object in the video to be processed.
  • the target part can be a part on the same target object, and can also be a part on different target objects.
  • the target part when the target part is a face, the target part can be the face of a different person; when the target part is a hand , the target part can be the hand of the same character.
  • FIG. 5 is a second schematic diagram of a video processing interface provided by an embodiment of the present disclosure.
  • target textures such as the "No. 6 texture” and “No. 8 texture” in sticker 1
  • FIG. 5 is shown as an example of a face, but it is not limited thereto.
  • the feature information includes: at least one of shape information, type information and color information.
  • the target part is a circle
  • an approximately circular target texture can be selected from multiple target textures
  • the feature information is type information
  • the target texture of the avatar can be selected from multiple target textures.
  • the target texture of the eye can be selected from multiple target textures.
  • the user can also modify the target texture.
  • the modification method is similar to the solution provided by the embodiment shown in FIG. 2 , and will not be repeated here.
  • target stickers when the user adds target stickers to multiple target parts by manually selecting the target object or target part, target stickers can be added to the multiple target parts one by one.
  • the user can first manually select a target object or target part in the video to be processed, and then select a target sticker for it; after the selection is successful, identify the selected target part in the video to be processed, and display the target part on the target part.
  • Target stickers further, add textures to other target parts in the same way.
  • the dynamic sticker when the target sticker is a dynamic sticker, the dynamic sticker can move correspondingly according to the movement of the target part.
  • the target map automatically follows the movement of the target part, the local area of the target part has a different motion frequency and/or motion mode from the overall target part, and the local area of the target map corresponding to the local area of the target part
  • the region has a different motion frequency and/or motion mode from the target texture
  • the local area of the target part has the same motion frequency and/or motion mode as the local area of the target texture
  • the motion frequency and/or motion mode of the target part and the target texture are the same.
  • the local area of the target sticker when the local area of the target part is recognized, the local area of the target sticker generates a style corresponding to the local area of the target sticker, and follows the local area of the target part.
  • the target part as a face
  • at least part of the area on the target part/partial area of the target part as the eye area, and the target map as a Coke bottle as an example
  • the Coke bottle also moves left and right accordingly
  • the partial area covering the eyes in the Coke bottle also blinks accordingly.
  • the partial area covering the eyes in the Coke bottle generates a pattern corresponding to the eyes and follows the eye movement.
  • a target texture is displayed on the target part in the target video.
  • step S404 is similar to the solution of step S205 in the embodiment shown in FIG. 2 , for details, reference may be made to the foregoing embodiments, and details are not repeated here.
  • a preview video may also be displayed on the video processing interface according to the currently added target texture, so that the user can check the current texture effect in time.
  • a first video with a first preset duration can be generated, and a target texture is displayed on the target part of the target object in the first video, wherein the first The preset duration is less than the duration of the video to be processed.
  • the length of the first preset duration for example, it may be any value shorter than the duration of the video to be processed, such as 3 seconds or 5 seconds.
  • the display information of the target texture on the target part can also be adjusted.
  • the scheme for adjusting the duration of the target texture will be described in detail below in conjunction with specific embodiments.
  • FIG. 6 is a third schematic flowchart of a video processing method provided by an embodiment of the present disclosure. As shown in Figure 6, the video processing method provided by the embodiment of the present disclosure includes the following steps:
  • S602. Receive a selection operation for multiple different target textures, and identify multiple target parts corresponding to at least one target object in the video to be processed.
  • steps S601-S603 are similar in scheme and principle to steps S401-S403 in the embodiment shown in FIG.
  • each duration control is used to display the display information of at least one target texture on the corresponding target part, wherein the display information includes: the display time of the target texture on the corresponding target part, the frame where the target texture is located in the video, etc. wait.
  • FIG. 7 is a third schematic diagram of a video processing interface provided by an embodiment of the present disclosure. As shown in (a) of FIG. 7 , duration control 1 and duration control 2 are used to display display information of target textures on target part 1 and target part 2 respectively, and each duration control can control at least one target texture.
  • S605. Receive an adjustment operation for the duration control, and adjust the display information of the target map corresponding to the duration control at the target location.
  • the adjustment operation includes: at least one of a deletion operation, a replacement operation, and a duration adjustment operation.
  • the adjustment control can also be displayed in the video processing interface, for example, delete the control, replace the control, etc., and the user can click on the duration control. Select the target texture to be deleted or replaced, and then delete or replace the corresponding target texture in the video to be processed by triggering the delete control or the replacement control.
  • the embodiment of the present disclosure does not specifically limit the duration adjustment operation.
  • the user can adjust the display period of each target map on the duration control by sliding left and right, thereby adjusting the target corresponding to the duration control in the target video.
  • duration controls corresponding to multiple parts is used as an example.
  • the display of duration controls corresponding to multiple parts can also be displayed and adjusted in the video processing interface in the same way.
  • a duration control corresponding to the target part can also be displayed and adjusted in the video processing interface in the same way.
  • S605a Receive a video segmentation and/or sequence adjustment operation for the video to be processed, divide and/or adjust multiple duration controls according to the video segmentation and/or sequence adjustment operation, and obtain multiple target duration controls.
  • S605b Display multiple target duration controls and the modified video to be processed on the video processing interface.
  • each target duration control is used to display the display information of the corresponding target texture in the target part in the modified video to be processed, wherein the display information includes: the display time of the target texture on the corresponding target part, the time of the target texture in the video The frame where in and so on.
  • FIG. 8 is a fourth schematic diagram of a video processing interface provided by an embodiment of the present disclosure. (a) in FIG. 8 is an undivided and/or adjusted video processing interface, and (b) in FIG. 8 is a divided and/or adjusted video processing interface.
  • the figure synchronously divides and switches the target textures in duration control 1 and duration control 2, and displays them in the video processing interface as shown in Figure 8 (b) The duration control shown in the figure.
  • FIG. 9 is a fifth schematic diagram of a video processing interface provided by an embodiment of the present disclosure. As shown in FIG. 9 , during the process of generating the target video, the status information of the generated target video may also be displayed on the video processing interface, wherein the status information includes the generation progress of the target video and the video picture corresponding to the current generation progress.
  • FIG. 9 the video generation progress in FIG. 9 is shown in the form of thumbnails, and in practical applications, it can also be shown in other forms, for example, progress bar, percentage and duration controls, etc.
  • the user may need to perform other operations, but these requirements may block the generation of the target video. Therefore, it is necessary to judge whether to continue to generate the target video according to the operation currently triggered by the user. .
  • the trigger operation on the functional control is received, and the operation corresponding to the functional control is performed according to the preset priority, or, the target video is continued to be generated.
  • the function control includes: at least one of subtitle recognition control, texture selection control and video processing control, and the preset priority is used to indicate the priority of the function control and generate the target video
  • the video processing control includes: video beautification, adding music , add subtitles or add special effects, etc.
  • the subtitle recognition process since the subtitle recognition process takes a certain amount of time and the time is not fixed, it will affect the generation of the target video. Therefore, in the process of generating the target video, the priority of subtitle recognition is lower than that of the target video generation process. When the user clicks on the subtitle recognition control, continue to generate the target video.
  • prompt information may also be displayed on the target video generation interface.
  • texture selection control and video processing control its priority can be the same as that of the target video processing process, so when the user triggers the texture selection control or video processing control, texture selection can be performed while the target video is being processed operation or video processing operation.
  • the region boundary of the target part can be identified, and the size of the texture can be automatically adjusted to completely cover the region boundary of the target part.
  • FIG. 10 is a schematic diagram of a video processing device provided by an embodiment of the present disclosure.
  • the video processing device 1000 includes:
  • the display module 1001 displays the video to be processed and at least one sticker in the video processing interface; the identification module 1002 receives a selection operation for the target sticker, and identifies the target part of the target object in the video to be processed; the display module 1001 is also used to A target texture is displayed on the target part in the video to be processed; the processing module 1003 is configured to receive editing operations on the target texture, modify the target texture, and generate a target video, in which the target part is displayed on the target texture.
  • the target map includes: at least one of a dynamic map and a static map, and the target map is used to block at least a partial area of the target part.
  • the identification module 1002 is specifically configured to: receive a selection operation for multiple different target textures, and identify multiple target parts corresponding to at least one target object in the video to be processed; the processing module 1003 is specifically used to : Display different target textures on multiple target parts.
  • the recognition module 1002 is specifically configured to: detect the target part of the target object in the video to be processed by image recognition technology; or receive a selection operation of the target object in the video to be processed, and identify The target part of the target object; or, receiving a selection operation of the target part in the video to be processed, and identifying the target part in the video to be processed.
  • the processing module 1003 is further configured to: generate a first video with a first preset duration, in which a target texture is displayed on the target part of the target object, and the first preset duration is shorter than the The duration of video processing; display the first video in the video processing interface.
  • the display module 1001 is further configured to: display a plurality of duration controls respectively corresponding to a plurality of target parts in the video processing interface, and each duration control is used to display at least one target texture on a corresponding target part information on the display.
  • the processing module 1003 is further configured to: receive an adjustment operation for the duration control, and adjust the display information of the target texture corresponding to the duration control at the target location; and generate a target video according to the adjustment operation.
  • the processing module 1003 is further configured to: receive a video segmentation and/or sequence adjustment operation for the video to be processed, and divide and/or adjust multiple duration controls according to the video segmentation and/or sequence adjustment operation , to obtain multiple target duration controls; multiple target duration controls and the modified video to be processed are displayed on the video processing interface, and each target duration control is used to display the display of the corresponding target texture in the modified video to be processed on the target part information.
  • the display module 1001 is further configured to: display the status information of the generated target video in the video processing interface, the status information includes the generation progress of the target video and the video picture corresponding to the current generation progress.
  • the processing module 1003 is further configured to: detect the action information of at least a part of the area on the target part in the video to be processed; control the corresponding area of the target map displayed on the target part to follow the action information to move accordingly, The corresponding area of the target map is used to block at least part of the area on the target part.
  • the processing module 1003 is further configured to: obtain feature information of the target part; determine the matching degree of each target map among multiple different target maps with the target part according to the feature information; Displays the target map that best matches the target part.
  • the above video processing device provided in this embodiment can be used to execute the technical solution of the embodiment shown in the above video processing method, and its implementation principle and technical effect are similar, and will not be repeated here.
  • FIG. 11 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure. It shows a schematic structural diagram of an electronic device 1100 suitable for implementing the embodiments of the present disclosure, and the electronic device 1100 may be the above-mentioned terminal device.
  • the terminal equipment may include but not limited to mobile phones, notebook computers, digital broadcast receivers, personal digital assistants (Personal Digital Assistant, PDA for short), tablet computers (Portable Android Device, PAD for short), portable multimedia players (Portable Media Player, referred to as PMP), mobile terminals such as vehicle-mounted terminals (such as vehicle-mounted navigation terminals), and fixed terminals such as digital TVs, desktop computers, etc.
  • FIG. 11 is only an example, and should not limit the functions and scope of use of the embodiments of the present disclosure.
  • an electronic device 1110 may include a processing device (such as a central processing unit, a graphics processing unit, etc.) 1108 is loaded into the program in the random access memory (Random Access Memory, referred to as RAM) 1103 to execute various appropriate actions and processes.
  • RAM Random Access Memory
  • various programs and data necessary for the operation of the electronic device 1100 are also stored.
  • the processing device 1101, the ROM 1102, and the RAM 1103 are connected to each other through a bus 1104.
  • An input/output (Input/Output, I/O for short) interface 1105 is also connected to the bus 1104 .
  • an input device 1106 including, for example, a touch screen, a touchpad, a keyboard, a mouse, a camera, a microphone, an accelerometer, a gyroscope, etc.; ), a speaker, a vibrator, etc.
  • a storage device 1108 including, for example, a magnetic tape, a hard disk, etc.
  • the communication means 1109 may allow the electronic device 1100 to perform wireless or wired communication with other devices to exchange data. While FIG. 11 shows electronic device 1100 having various means, it is to be understood that implementing or having all of the means shown is not a requirement. More or fewer means may alternatively be implemented or provided.
  • an embodiment of the present disclosure includes a computer program product, which includes a computer program carried on a computer-readable medium, where the computer program includes program code for executing the method shown in the flowchart.
  • the computer program may be downloaded and installed from a network via communication means 1109, or from storage means 1108, or from ROM 1102.
  • the processing device 1101 When the computer program is executed by the processing device 1101, the above-mentioned functions defined in the methods of the embodiments of the present disclosure are performed.
  • the above-mentioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the above two.
  • a computer readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof.
  • Computer-readable storage media may include, but are not limited to, electrical connections with one or more wires, portable computer diskettes, hard disks, random access memory (RAM), read-only memory (ROM), erasable Electrical Programmable Read Only Memory (EPROM or flash memory for short), optical fiber, compact disc read-only memory (CD-ROM for short), optical storage device, magnetic storage device, or the above-mentioned any suitable combination.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program that can be used by or in conjunction with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave carrying computer-readable program code therein.
  • Such propagated data signals may take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, which can transmit, propagate, or transmit a program for use by or in conjunction with an instruction execution system, apparatus, or device .
  • the program code contained on the computer readable medium can be transmitted by any appropriate medium, including but not limited to: electric wire, optical cable, radio frequency (Radio Frequency, RF for short), etc., or any suitable combination of the above.
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device, or may exist independently without being incorporated into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs, and when the above-mentioned one or more programs are executed by the electronic device, the electronic device is made to execute the methods shown in the above-mentioned embodiments.
  • Computer program code for carrying out the operations of the present disclosure can be written in one or more programming languages, or combinations thereof, including object-oriented programming languages—such as Java, Smalltalk, C++, and conventional Procedural Programming Language - such as "C" or a similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or it can be connected to an external A computer (connected via the Internet, eg, using an Internet service provider).
  • LAN Local Area Network
  • WAN Wide Area Network
  • Embodiments of the present disclosure also provide a computer program, which implements the methods shown in the above embodiments when the computer program is executed by a processor.
  • the video processing method, device, device, storage medium, electronic device, computer program product, and computer program provided by the embodiments of the present disclosure firstly display the video to be processed and at least one texture in the video processing interface; receive a selection operation for the target texture, Identify the target part of the target object in the video to be processed; display the target texture on the target part in the video to be processed; receive the editing operation on the target texture and modify the target texture; generate the target video, and place it on the target part in the target video Displays the target texture.
  • the target part of the target object in the video to be processed can be automatically identified without manual selection by the user, which can simplify the map operation and improve the occlusion effect of the map.
  • the map can also be Editing and adjustments can further improve the flexibility and convenience in the video processing process, meet the individual needs of users, enrich video interaction functions, and improve user experience.
  • each block in a flowchart or block diagram may represent a module, program segment, or portion of code that contains one or more logical functions for implementing specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented by a dedicated hardware-based system that performs the specified functions or operations , or may be implemented by a combination of dedicated hardware and computer instructions.
  • the units involved in the embodiments described in the present disclosure may be implemented by software or by hardware. Wherein, the name of the unit does not constitute a limitation of the unit itself under certain circumstances, for example, the first obtaining unit may also be described as "a unit for obtaining at least two Internet Protocol addresses".
  • exemplary types of hardware logic components include: Field Programmable Gate Array (Field Programmable Gate Array, FPGA for short), Application Specific Integrated Circuit (ASIC for short), Application Specific Standard Products ( Application Specific Standard Parts (ASSP for short), System on Chip (SOC for short), Complex Programmable Logic Device (CPLD for short), etc.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in conjunction with an instruction execution system, apparatus, or device.
  • a machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • a machine-readable medium may include, but is not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, apparatus, or devices, or any suitable combination of the foregoing.
  • machine-readable storage media would include one or more wire-based electrical connections, portable computer discs, hard drives, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM or flash memory), optical fiber, compact disk read only memory (CD-ROM), optical storage, magnetic storage, or any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read only memory
  • EPROM or flash memory erasable programmable read only memory
  • CD-ROM compact disk read only memory
  • magnetic storage or any suitable combination of the foregoing.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Marketing (AREA)
  • Business, Economics & Management (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

本公开实施例提供一种视频处理方法、装置、电子设备、存储介质、计算机程序产品及计算机程序,首先在视频处理界面中显示待处理视频以及至少一个贴图;接收针对目标贴图的选择操作,识别待处理视频中的目标对象的目标部位;在待处理视频中的目标部位上显示目标贴图;接收作用于目标贴图上的编辑操作,修改目标贴图;生成目标视频,目标视频中的目标部位上显示有目标贴图。

Description

视频处理方法、装置、设备及存储介质
相关申请的交叉引用
本申请要求于2021年08月03日提交中国专利局、申请号为202110886909.1、申请名称为“视频处理方法、装置、设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本公开实施例涉及视频处理技术领域,尤其涉及一种视频处理方法、装置、设备、电子设备、存储介质、计算机程序产品及计算机程序。
背景技术
随着互联网技术的发展,短视频、视频直播等业务也在逐渐兴起,现有技术中,视频处理过程复杂,影响用户体验。
发明内容
本公开实施例提供一种视频处理方法、装置、设备、存储介质、电子设备、计算机程序产品及计算机程序。
第一方面,本公开实施例提供一种视频处理方法,包括:在视频处理界面中显示待处理视频以及至少一个贴图;接收针对目标贴图的选择操作,识别待处理视频中的目标对象的目标部位;在待处理视频中的目标部位上显示目标贴图;接收作用于目标贴图上的编辑操作,修改目标贴图;生成目标视频,目标视频中的目标部位上显示有目标贴图。
第二方面,本公开实施例提供一种视频处理装置,包括:显示模块,在视频处理界面中显示待处理视频以及至少一个贴图;识别模块,接收针对目标贴图的选择操作,识别待处理视频中的目标对象的目标部位;显示模块,还用于在待处理视频中的目标部位上显示目标贴图;处理模块,用于接收作用于目标贴图上的编辑操作,修改目标贴图,生成目标视频,目标视频中的目标部位上显示有目标贴图。
第三方面,本公开实施例提供一种电子设备,包括:至少一个处理器和存储器;存储器存储计算机执行指令;至少一个处理器执行存储器存储的计算机执行指令,使得至少一个处理器执行如第一方面的视频处理方法。
第四方面,本公开实施例提供一种计算机可读存储介质,计算机可读存储介质中存储有计算机执行指令,当处理器执行计算机执行指令时,实现如第一方面的视频处理方法。
第五方面,本公开实施例提供一种计算机程序产品,包括:计算机程序,其特征在于,该计算机程序被处理器执行时,实现如第一方面的视频处理方法。
第六方面,本公开实施例提供一种计算机程序,该计算机程序在被处理器执行时实现如第一方面的视频处理方法。
附图说明
为了更清楚地说明本公开实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作一简单地介绍,显而易见地,下面描述中的附图是本公开的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。
图1为本公开实施例提供的视频处理方法的应用场景示意图;
图2为本公开实施例提供的视频处理方法的流程示意图一;
图3为本公开实施例提供的视频处理界面的示意图一;
图4为本公开实施例提供的视频处理方法的流程示意图二;
图5为本公开实施例提供的视频处理界面的示意图二;
图6为本公开实施例提供的视频处理方法的流程示意图三;
图7为本公开实施例提供的视频处理界面的示意图三;
图8为本公开实施例提供的视频处理界面的示意图四;
图9为本公开实施例提供的视频处理界面的示意图五;
图10为本公开实施例提供的视频处理装置的示意图;
图11为本公开实施例提供的电子设备的示意图。
具体实施方式
为使本公开实施例的目的、技术方案和优点更加清楚,下面将结合本公开实施例中的附图,对本公开实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例是本公开一部分实施例,而不是全部的实施例。基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本公开保护的范围。
相关技术中,用户在处理视频或者进行视频直播时,可以手动为视频中的人或物添加贴纸,从而通过贴纸对人或物进行遮挡,满足用户的个性化需求和丰富视频交互功能,进而提升用户体验。然而,此过程需要用户手动选中视频中被遮挡的位置,操作繁琐且遮挡效果不佳。
为了解决上述问题,本公开的技术构思是,在进行视频处理过程中,在用户为目标部位选择目标贴图后,自动识别待处理视频中的待遮挡的目标部位,并在目标部位上显示该贴图,从而实现遮挡,同时在目标贴图上设置编辑控件,从而实现对已添加的目标贴图进行编辑。由于无需手动添加贴图,可以简化贴图操作,提升贴图的遮挡效果,另外,在添加贴图后,还可以对贴图进行编辑调整,从而可以进一步提升视频处理过程中的灵活度和便捷性,提升用户体验。
为方便理解,首先结合图1对本公开实施例的应用场景进行说明:
图1为本公开实施例提供的视频处理方法的应用场景示意图。参考图1,该场景包括:终端设备100。
其中,终端设备100可以为移动终端(例如,智能手机、掌上电脑、平板电脑等具有视频处理功能的移动设备、可穿戴电子设备)、计算机设备(例如,台式机、笔记本电脑、一体机、智能家居设备等)等,图1中以智能手机为例示出,但不以此为限定。
在实际应用中,用户可以将待处理视频导入到该终端设备100中,从而通过终端设备100实现对待处理视频的处理。
一些实施例中,该场景还可以包括服务器200,终端设备100可以通过无线或有线网络与服务器200通信,用于发送数据给服务器200。其中,无线网络可以是2G或者3G或者4G或者5G等通信网络,也可以是无线局域网,在此不做限定。在视频处理过程中,用户通过终端设备100触发视频上传、视频剪辑、添加贴纸等操作,再将操作数据发送至服务器200,由服务器200执行相应的视频处理操作。
需要说明的是,本公开实施例提供的方案可以应用于多种视频处理场景,例如,视频的后期处理,直播过程中的视频画面处理以及视频拍摄过程中的视频处理等。
应理解,附图1仅是本公开实施例提供的一种应用场景的示意图,本公开实施例不对图1中包括的设备种类及设备个数进行限定,也不对图1中设备之间的位置关系进行限定,例如,在图1所示的应用场景中,还可以包括数据存储设备,该数据存储设备相对服务器200可以是外部存储器,也可以是集成在服务器200中的内部存储器。另外,服务器200可以是独立的服务器,或者,也可以是服务集群等。
下面以具体地实施例对本公开实施例的技术方案以及本公开的技术方案如何解决上述技术问题进行详细说明。下面这几个具体的实施例可以相互结合,对于相同或相似的概念或过程可能在某些实施例中不再赘述。下面将结合附图,对本公开实施例的实施例进行描述。
图2为本公开实施例提供的视频处理方法的流程示意图一。本实施例提供的视频处理方法可以应用于图1所示的终端设备或服务器,参考图2,本实施例提供的视频处理方法包括以下步骤:
S201、在视频处理界面中显示待处理视频以及至少一个贴图。
一些实施例中,该待处理视频可以是用户上传的,或者,该待处理视频也可以是用户通过终端设备的图像采集设备实时采集到的,比如摄像头等组件。
在实际应用中,对于贴图的具体类型,本公开实施例不做具体限定,示例性的,可以为静态贴图或者动态贴图,例如是动物形象、动漫形象、静物形象等等,目标贴图用于遮挡目标部位的至少部分区域。
图3为本公开实施例提供的视频处理界面的示意图一。需要说明的是,图3中的目标对象的目标部位以脸部为例示出,但不以此为限定。如图3(a)所示,在该视频处理界面共展示了贴纸1、贴纸2、贴纸3和贴纸4四种贴纸类型,每个贴纸类型中都包含多个贴图,用户可以通过上下滑动或左右滑动选择各种贴纸中的目标贴图。
S202、接收针对目标贴图的选择操作,识别待处理视频中的目标对象的目标部位。
需要说明的是,对于目标对象的类型本公开实施例也不做限定,例如,目标对象可以为人物、动物等任意需要添加贴图的对象,相应的,目标部位可以为目标对象上的任意部位,例如,对于人物来说,目标对象可以为脸部、手部、头发等部位,或者也可以为上述部位的局部区域,例如是眼睛、嘴巴等一个或多个区域。
如图3中的(a)图所示,用户可以通过点击贴图来选择目标贴图,相应的,在用户选择目标贴图之后,识别待处理视频中目标对象的目标部位。
一些实施例中,目标对象的目标部位是固定的,即在该视频处理过程中,所有贴图均用于遮挡某一类型的部位。
示例性的,在本次视频处理过程中,所有的贴图均用于遮挡人物的目标部位,如手部,当用户选择目标贴图之后,自动识别待处理图像中的手部。
另一些实施例中,用户可以在进入该视频处理界面之前,设定本次视频处理过程中待处理的目标对象和/或目标部位,例如,若用户设定的目标对象为人物,目标部位为手部时,则在该视频处理过程中,所有贴图均用于遮挡手部。
用户也可以设定多个目标部位,例如在视频编辑页面的第一区域展示一或多个目标对象的待处理图像,视频编辑页面的第二区域展示多个目标贴图,用户可以选中第一目标贴图并将第一目标贴图拖至第一目标对象的第一目标部位,再选中第二目标贴图并将第二目标贴图拖至第一目标对象或第二目标对象的第二目标部位。
添加完成之后,响应于接收到视频边界页面的添加完成确认操作,将第一目标贴图添加至视频中所有包括第一目标部位视频帧中,并且各视频帧中第一目标贴图均遮挡第一目标部位,将第二目标贴图添加至视频中所有包括第二目标部位的视频帧中,并且各视频帧中第二目标贴图均遮挡第二目标部位。第一目标对象和第二目标对象分别为所述多个目标对象中的任意目标对象。当然,并不仅限于同时添加两个贴图,此处不再赘述。由此,实现了针对视频同时添加多个贴图,每个贴图能够匹配至视频中的每一帧包括目标部位的区域,丰富了视频交互方式,提高了视频编辑效率。
在其他实施例中,还可以根据目标贴图的类型(形状类型、颜色类型等),识别待处理视频中目标对象的目标部位并添加,具体的,响应于接收到针对目标贴图的选中操作,识别待处理图像中与目标贴图类型匹配的目标部位,并将目标贴图添加至目标部位。示例性的,第一目标贴图的类型为瘦长样式,识别到待处理图像中胳膊也为瘦长样式,则将胳膊确定为第一目标部位,第二目标贴图的类型为圆形样式,识别到待处理图像中脸部也为圆形样式,则将脸部确定为第二目标部位。在用户选中第一目标贴图和第二目标贴图之后,无需将目标贴图移动至目标部位,第一目标贴图自动匹配至视频中包括第一目标部位的各视频帧中,以遮挡胳膊,第二目标贴图自动匹配至视频中包括第二目标部位的各视频帧中,以遮挡脸部。由此,简化了用户操作,提升了用户体验。再例如,贴纸1为用于遮挡头发的贴纸,贴图2为用于遮挡动物的贴纸,贴纸4为用于遮挡手部的贴纸。在识别过程中,如果选择了贴纸1中的目标贴图,则根据该目标贴图识别待处理视频中人物的头发,至于其他类型的贴纸与之类似,此处不再赘述。
在一些实施例中,该贴图为可编辑贴图,在进行视频编辑时,例如在将目标贴图移动至待处理图像的目标部位上后,在贴图周围展示有一个或多个编辑控件,以方便操作。所述编辑控件为对贴图元素的属性进行编辑的控件,所述贴图元素的属性包括亮度属性,色度属性、对比度属性长宽比属性等,相应的编辑控件包括亮度调节控件、色度调节控件,对比度调节控件、比例调节控件,所述贴图元素即贴图图案,例如狗头图案、花朵图案、饮料瓶图案等等。与现有贴图相比,本公开采用贴图元素周围外露展示编辑控件 的可编辑贴图极大的满足了用户对贴图的快速调节需求,有利于提高目标贴图和目标部位的匹配度,而且,缩短了交互路径,提高了编辑效率。
S203、在待处理视频中的目标部位上显示目标贴图。
相应的,在识别成功后,在目标部位上显示目标贴图。如图3中的(b)图所示,当用户现在贴纸1中的“6号贴图”时,在待处理视频中的所有人物的脸部显示“6号贴图”。
可选的,若识别失败,还可以在视频处理界面中显示相应的提示信息,其中,该提示信息用于提示以下至少一种信息:提示当前识别目标对象失败、提示用户手动选择目标部位、提示用户上传新的待处理视频,以及,提示用户将待处理视频的进度调节到包含目标部位的位置。
S204、接收作用于目标贴图上的编辑操作,修改目标贴图。
其中,修改操作包括亮度调节操作、色度调节操作、对比度调节操作、比例调节操作、尺寸调节操作、删除操作、替换操作和方向调节操作中的至少一种。
在一些实施例中,接收作用于目标贴图上的编辑操作可以是作用在所述目标贴图上展示的编辑控件上的触发操作。
可选的,在目标部位上显示目标贴图之后,在目标贴图上显示有修改控件。如图3中的(b)图所示,修改控件包括:尺寸调节控件、删除控件、替换控件和方向调节控件等,用户可以通过修改控件来调节目标贴图,至于调节方式,本公开实施例不做具体限定。
通过本方案,用户可以直接对目标贴图进行编辑调整,操作相对简单,可以进一步提升视频处理过程中的灵活度和便捷性,从而满足用户的个性化需求、丰富视频交互功能,进而提升用户体验。
一些实施例中,当用户触发删除控件或者替换控件后,删除目标部位上的目标贴图,并在待处理界面中展示提示信息,其中,信息用于提示用户重新为目标对象添加目标贴图。
可选的,在目标部位上显示目标贴图之后,也可以在视频处理界面中显示如图3中(b)图所示的修改控件,例如,删除控件和替换控件等,使得用户可以通过点击界面中的修改控件对目标贴图进行相应操作。
S205、生成目标视频。
本步骤中,用户在目标部位添加目标贴图之后,根据添加目标贴图后的待处理视频生成目标视频,其中,目标视频中的目标部位上显示有用户添加的目标贴图。
对于生成目标视频的方式,本公开实施例不做具体限定,例如,在目标部位上显示目标贴图后,可以同步在视频处理界面中显示用于生成目标视频的控件,当用户点击该控件时,即生成目标视频,或者,也可以在目标部位上显示目标贴图后,直接生成目标视频。
需要说明的是,上述步骤S204为可选步骤,也就是说,在步骤S203之后,用户若未对目标贴图进行修改,则可以根据用户添加的目标贴图生成目标视频;若用户对目标贴图进行了修改操作,则根据修改后的目标贴图生成目标视频,此时,生成的目标视频中的目标部位显示有修改后的目标贴图。
本公开实施例中,可以通过自动识别待处理视频中的目标对象的目标部位,无需用户手动选中,简化了贴图操作,提升贴图的遮挡效果,另外,在添加贴图后,还可以对贴图进行编辑调整,从而可以进一步提升视频处理过程中的灵活度和便捷性,提升用户体验。
需要说明的是,对于识别待处理视频中的目标对象的目标部位的方案,本公开实施例不做具体限定。例如,一些实施例中,可以通过图像识别技术检测待处理视频中目标对象的目标部位。
另一些实施例中,还可以由用户手动选择目标对象,从而识别该目标对象的目标部位,示例性的,用户可以点击选中待处理视频中的目标对象,相应的,在添加目标贴图时,接收用户对待处理视频中的目标对象的选择操作,识别该目标对象的目标部位。
在其实施例中,还可以由用户直接手动选择目标对象的目标部位,相应的,接收用户对待处理视频中的目标部位的选中操作,检测待处理视频中的目标部位,示例性的,用户可以点击选中待处理视频中的目标部位,相应的,在添加目标贴图时,接收用户对待处理视频中的目标部位的选择操作,识别该目标部位。
本公开实施例中,提供了多种目标部位的选择方式,可以提升视频处理过程的灵活性,提升用户体验。
在实际应用中,当待处理视频中包含多个目标对象时,还可以为不同的目标对象添加不同的目标贴图,图4为本公开实施例提供的视频处理方法的流程示意图二。如图4所示,该视频处理方法包括如下步骤:
S401、在视频处理界面中显示待处理视频以及至少一个贴图。
需要说明的是,在视频处理界面中显示待处理视频以及至少一个贴图的显示界面请参考图3中的(a)图,此处不做赘述。
S402、接收针对多个不同的目标贴图的选择操作,识别待处理视频中至少一个目标对象对应的多个目标部位。
其中,目标部位可以为同一目标对象上的部位,也可以为不同目标对象上的部位,例如,当目标部位为脸部时,目标部位可以为不同人物的脸部,当目标部位为手部时,目标部位可以为同一人物的手部。
图5为本公开实施例提供的视频处理界面的示意图二。如图5中的(a)图所示,用户在选择多个不同的目标贴图时,例如贴纸1中的“6号贴图”和“8号贴图”,在识别过程中,识别待处理视频中的与目标贴图数量相对应的脸部。需要说明的是,图5中的目标对象的目标部位以脸部为例示出,但不以此为限定。
S403、在多个目标部位上分别显示不同的目标贴图。
如图5中的(b)图所示,在不同目标部位上分别显示“6号贴图”和“8号贴图”。具体的,在目标部位上显示不同的目标贴图的方案,本公开实施例不做具体限定,例如,可以在多个目标部位上随机显示用户选择的目标贴图。
另一些实施例中,还可以根据目标部位的特征信息,为每个目标部位匹配更合适的目标贴图,具体包括如下步骤:
(1)获取目标部位的特征信息;
(2)根据特征信息,确定多个不同的目标贴图中每个目标贴图与目标部位的匹配度。
(3)在目标部位上显示与目标部位匹配度最高的目标贴图。
其中,特征信息包括:形状信息、类型信息和颜色信息中的至少一种。
具体的,当特征信息为形状信息时,若目标部位为圆形,则可以从多个目标贴图中选择近似圆形的目标贴图;当特征信息为类型信息时,若目标部位为脸部,则可以从多个目标贴图中选择头像类的目标贴图,若目标部位为眼部,则可以从多个目标贴图中选择眼部的目标贴图。
应理解,与上述实施例类似,在目标部位显示目标贴图之后,用户也可以对目标贴图进行修改,其修改方式与图2所示实施例提供的方案类似,此处不再赘述。
一些实施例中,当用户通过手动选择目标对象或目标部位,为多个目标部位添加目标贴图时,可以逐个为多个目标部位添加目标贴纸。示例性的,用户可以先手动选择待处理视频中的一个目标对象或者目标部位,再为其选择目标贴纸;选择成功后,识别待处理视频中选中的目标部位,并在该目标部位上显示该目标贴纸,进一步的,按照相同的方式为其他目标部位添加贴图。
一些实施例中,当目标贴图为动态贴纸时,动态贴图可以根据目标部位的运动而进行相应的运动。
具体包括如下步骤:(1)检测待处理视频中目标部位上至少部分区域的动作信息;
(2)控制目标部位上显示的目标贴图的对应区域跟随动作信息进行相应的运动,目标贴图的对应区域用于遮挡目标部位上至少部分区域。
在一些实施例中,目标贴图随目标部位的运动而自动跟随运动,目标部位的局部区域具有与目标部位整体不同的运动频率和/或运动方式,目标部位的局部区域所对应的目标贴图的局部区域与目标贴图具有不同的运动频率和/或运动方式,目标部位的局部区域与目标贴图的局部区域运动频率和/或运动方式相同,目标部位与目标贴图的运动频率和/或运动方式相同。进一步的,在识别到目标部位的局部区域动作的情况下,目标贴图的局部区域生成与目标贴纸的局部区域对应的样式,并跟随目标部位的局部区域动作。
示例性的,以目标部位为脸部,目标部位上至少部分区域/目标部位的局部区域为眼部区域、目标贴图是可乐瓶为例,当视频中脸部左右运动、同时眼部眨眼运动时,可乐瓶也相应的左右运动,同时可乐瓶中遮挡眼部的局部区域也相应的眨眼运动,具体的可乐瓶中遮挡眼部的局部区域生成与眼部对应的样式,跟随眼部运动。
S404、生成目标视频。
其中,目标视频中的目标部位上显示有目标贴图。
需要说明的是,步骤S404的方案与图2所示实施例中的步骤S205的方案类似,具体可参考上述实施例,此处不再赘述。
可选的,在生成目标视频之前,还可以根据当前添加的目标贴图,在视频处理界面展示一段预览视频,以使得用户及时查看当前的贴图效果。
具体的,在待处理视频中的目标部位上显示目标贴图后,可以生成第一预设时长的第一视频,并在第一视频中目标对象的目标部位上显示有目标贴图,其中,第一预设时长小于待处理视频的时长。需要说明的是,对于第一预设时长的长度不做具体限定,例如,可以为3秒、5秒等任意小于待处理视频的时长的数值。
在一些实施例中,在目标部位上显示目标贴图之后,还可以调节目标贴图在目标部位上的显示信息,下面结合具体实施例对目标贴图的时长调节方案进行详细说明。
图6为本公开实施例提供的视频处理方法的流程示意图三。如图6所示,本公开实施例提供的视频处理方法包括如下步骤:
S601、在视频处理界面中显示待处理视频以及至少一个贴图。
S602、接收针对多个不同的目标贴图的选择操作,识别待处理视频中至少一个目标对象对应的多个目标部位。
S603、在多个目标部位上分别显示不同的目标贴图。
需要说明的是,步骤S601~S603与图4所示实施例中的步骤S401~S403的方案和原理类似,具体可参考上述实施例,此处不再赘述。
S604、在视频处理界面中展示多个目标部位分别对应的多个时长控件。
其中,每个时长控件用于展示至少一个目标贴图在对应的目标部位上的显示信息,其中,显示信息包括:目标贴图在对应的目标部位上的显示时间、目标贴图在视频中所在的帧等等。
图7为本公开实施例提供的视频处理界面的示意图三。如图7中的(a)图所示,时长控件1和时长控件2分别用于展示目标部位1和目标部位2上的目标贴图的显示信息,每个时长控件可以控制至少一个目标贴图。
示例性的,在处理视频的过程中,在待处理视频的不同时刻,分别给目标部位1添加了“6号贴图”和“3号贴图”,给目标部位2分别添加了“8号贴图”和“5号贴图”,其中,以时长控件1为例,时段1和时段2分别为“6号贴图”和“3号贴图”在目标部位1上的显示时长。
S605、接收针对时长控件的调节操作,调节时长控件对应的目标贴图在目标部位的显示信息。
S606、根据调节操作,生成目标视频。
其中,调节操作包括:删除操作、替换操作、时长调节操作中的至少一种。
请继续参考图7中的(a)图,在视频处理界面中显示时长控件的同时,还可以在视频处理界面中显示调节控件,例如是,删除控件、替换控件等,用户可以在时长控件上选中待删除或者待替换的目标贴图,再通过触发删除控件或者替换控件来删除或替换待处理视频中对应的目标贴图。
需要说明的是,当用户触发替换按钮后,在视频处理界面中显示至少一个贴图和待处理视频,供用户为目标部位替换新的贴图。示例性的,如图7中的(b)图所示,当用户触发对“8号贴图”的替换操作时,在视频处理界面中展示待处理视频以及至少一个贴图,当用户选择“7号贴图”后,继续回到如图7中(a)图所示的界面,并将时长控件2中的“8号贴图”替换为“7号贴图”,同时将待处理视频中目标部位2的对应时段上显示“7号贴图”。
在实际应用中,对于时长调节操作,本公开实施例不做具体限定,例如,用户可以通过左右滑动来调节时长控件上每个目标贴图的显示时段,从而调节目标视频中,时长控件对应的目标贴图在目标部位的显示信息。
需要说明的是,本公开实施例中,以展示多个部位对应的时长控件为例示出,当只为一个目标部位添加目标贴图时,也可以按照相同的方法,在视频处理界面中展示并调节该目标部位对应的一个时长控件。
可选的,在待处理视频的目标部位添加目标贴图后,可能还会对添加目标贴图后的待处理视频进行视频分割和顺序调整等操作,此时需要同步调整时长控件,使得时长控件与分割或顺序调整后的待处理视频保持对应。下面结合步骤S606a~S606b进行详细说明:
S605a、接收针对待处理视频的视频分割和/或顺序调整操作,根据视频分割和/或顺序调整操作,分割和/或调整多个时长控件,获得多个目标时长控件。
S605b、在视频处理界面展示多个目标时长控件和修改后的待处理视频。
其中,每个目标时长控件用于展示修改后的待处理视频中对应的目标贴图在目标部位的显示信息,其中,显示信息包括:目标贴图在对应的目标部位上的显示时间、目标贴图在视频中所在的帧等等。
图8为本公开实施例提供的视频处理界面的示意图四。图8中的(a)图为未分割和/或调整的视频处理界面,图8中的(b)图为分割和/或调整的视频处理界面。
示例性的,以将时段1和时段2对应的视频进行分割并切换顺序为例,图同步分割和切换时长控件1和时长控件2中的目标贴图,并在视频处理界面中展示如图8中(b)图所示的时长控件。
S606a、在视频处理界面中展示生成目标视频的状态信息。
图9为本公开实施例提供的视频处理界面的示意图五。如图9所示,在生成目标视频的过程中,还可以在视频处理界面中展示生成目标视频的状态信息,其中,状态信息包括目标视频的生成进度和当前生成进度对应的视频画面。
需要说明的是,图9中的视频生成进度以缩略图的形式示出,在实际应用中,还可以以其他形式示出,例如,进度条、百分比和时长控件等。
一些实施例中,在视频生成过程中,用户可能还有进行其他操作的需求,但这些需求可能会阻断目标视频的生成,因此,需要根据用户当前触发的操作,来判断是否继续生成目标视频。
具体的,接收对功能控件的触发操作,根据预设优先级执行功能控件对应的操作,或者,继续生成目标视频。其中,功能控件包括:字幕识别控件、贴图选择控件和视频处理控件中的至少一种,预设优先级用于指示功能控件和生成目标视频的优先级,视频处理控件包括:视频美化、添加音乐、添加字幕或添加特效等。
以字幕识别控件为例,由于字幕识别过程中需要耗费一定时间,且时间不固定,会影响目标视频的生成,因此,在目标视频生成过程中,字幕识别的优先级低于目标视频生成过程,当用户点击字幕识别控件时,继续生成目标视频。
可选的,还可以在目标视频生成界面中展示提示信息。
对于贴图选择控件和视频处理控件来说,其优先级可以与目标视频处理过程的优先级相同,因此,当用户触发贴图选择控件或视频处理控件时,可以在目标视频处理的同时,进行贴图选择操作或视频处理操作。
在一些实施例中,在选中目标贴图并确认添加之后,可识别目标部位的区域边界,自动调整贴图大小,以完整遮挡目标部位的区域边界。
对应于上文实施例所示的视频处理方法,图10为本公开实施例提供的视频处理装置的示意图。为了便于说明,仅示出了与本公开实施例相关的部分。如图10所示,该视频处理装置1000包括:
显示模块1001,在视频处理界面中显示待处理视频以及至少一个贴图;识别模块1002,接收针对目标贴图的选择操作,识别待处理视频中的目标对象的目标部位;显示模块1001,还用于在待处理视频中的目标部位上显示目标贴图;处理模块1003,用于接收作用于目标贴图上的编辑操作,修改目标贴图,生成目标视频,目标视频中的目标部位上显示有目标贴图。
在本公开的一个实施例中,目标贴图包括:动态贴图和静态贴图中的至少一种,目标贴图用于遮挡目标部位的至少部分区域。
在本公开的一个实施例中,识别模块1002具体用于:接收针对多个不同的目标贴图的选择操作,识别待处理视频中至少一个目标对象对应的多个目标部位;处理模块1003具体用于:在多个目标部位上分别显示不同的目标贴图。
在本公开的一个实施例中,识别模块1002具体用于:通过图像识别技术检测待处理视频中目标对象的目标部位;或者,接收对待处理视频中的目标对象的选择操作,识别待处理视频中目标对象的目标部位;或者,接收对待处理视频中目标部位的选择操作,识别待处理视频中的目标部位。
在本公开的一个实施例中,处理模块1003还用于:生成第一预设时长的第一视频,在第一视频中目标对象的目标部位上显示有目标贴图,第一预设时长小于待处理视频的时长;在视频处理界面中展示第一视频。
在本公开的一个实施例中,显示模块1001还用于:在视频处理界面中展示多个目标部位分别对应的多个时长控件,每个时长控件用于展示至少一个目标贴图在对应的目标部位上的显示信息。
在本公开的一个实施例中,处理模块1003还用于:接收针对时长控件的调节操作,调节时长控件对应的目标贴图在目标部位的显示信息;根据调节操作,生成目标视频。
在本公开的一个实施例中,处理模块1003还用于:接收针对待处理视频的视频分割和/或顺序调整操作,根据视频分割和/或顺序调整操作,分割和/或调整多个时长控件,获得多个目标时长控件;在视频处理界面展示多个目标时长控件和修改后的待处理视频,每个目标时长控件用于展示修改后的待处理视频中对应的目标贴图在目标部位的显示信息。
在本公开的一个实施例中,显示模块1001还用于:在视频处理界面中展示生成目标视频的状态信息,状态信息包括目标视频的生成进度和当前生成进度对应的视频画面。
在本公开的一个实施例中,处理模块1003还用于:检测待处理视频中目标部位上至少部分区域的动作信息;控制目标部位上显示的目标贴图的对应区域跟随动作信息进行相应的运动,目标贴图的对应区域用于遮挡目标部位上至少部分区域。
在本公开的一个实施例中,处理模块1003还用于:获取目标部位的特征信息;根据特征信息,确定多个不同的目标贴图中每个目标贴图与目标部位的匹配度;在目标部位上显示与目标部位匹配度最高的目标贴图。
本实施例提供的上述视频处理装置,可用于执行上述视频处理方法所示实施例的技术方案,其实现原理和技术效果类似,此处不再赘述。
图11为本公开实施例提供的电子设备的结构示意图。其示出了适于用来实现本公开实施例的电子设备1100的结构示意图,该电子设备1100可以为上述终端设备。其中,终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、个人数字助理(Personal Digital Assistant,简称PDA)、平板电脑(Portable Android Device,简称PAD)、便携式多媒体播放器(Portable Media Player,简称PMP)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。
应理解,图11示出的电子设备仅仅是一个示例,不应对本公开实施、例的功能和使用范围带来任何限制。
如图11所示,电子设备1110可以包括处理装置(例如中央处理器、图形处理器等)1101,其可以根据存储在只读存储器(Read Only Memory,简称ROM)1102中的程序或者从存储装置1108加载到随机访问存储器(Random Access Memory,简称RAM)1103中的程序而执行各种适当的动作和处理。在RAM 1103中,还存储有电子设备1100操作所需的各种程序和数据。处理装置1101、ROM 1102以及RAM 1103通过总线1104彼此相连。输入/输出(Input/Output,简称I/O)接口1105也连接至总线1104。
通常,以下装置可以连接至I/O接口1105:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置1106;包括例如液晶显示器(Liquid Crystal Display,简称LCD)、扬声器、振动器等的输出装置1107;包括例如磁带、硬盘等的存储装置1108;以及通信装置1109。通信装置1109可以允许电子设备1100与其他设备进行无线或有线通信以交换数据。虽然图11示出了具有各种装置的电子设备1100,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置1109从网络上被下载和安装,或者从存储装置1108被安装,或者从ROM 1102被安装。在该计算机程序被处理装置1101执行时,执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(Electrical Programmable Read Only Memory,简称EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(compact disc read-only memory,简称CD-ROM)、 光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、射频(Radio Frequency,简称RF)等等,或者上述的任意合适的组合。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备执行上述实施例所示的方法。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(Local Area Network,简称LAN)或广域网(Wide Area Network,简称WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
本公开实施例还提供了一种计算机程序,该计算机程序在被处理器执行时实现如上述实施例所示的方法。
本公开实施例提供的视频处理方法、装置、设备、存储介质、电子设备、计算机程序产品及计算机程序,首先在视频处理界面中显示待处理视频以及至少一个贴图;接收针对目标贴图的选择操作,识别待处理视频中的目标对象的目标部位;在待处理视频中的目标部位上显示目标贴图;接收作用于目标贴图上的编辑操作,修改目标贴图;生成目标视频,目标视频中的目标部位上显示有目标贴图。本公开实施例提供的方案中,可以自动识别待处理视频中的目标对象的目标部位,无需用户手动选中,可以简化贴图操作,提升贴图的遮挡效果,另外,在添加贴图后,还可以对贴图进行编辑调整,从而可以进一步提升视频处理过程中的灵活度和便捷性,满足用户的个性化需求、丰富视频交互功能,进而提升用户体验。
附图中的流程图和框图,图示了按照本公开各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的 功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(Field Programmable Gate Array,简称FPGA)、专用集成电路(Application Specific Integrated Circuit,简称ASIC)、专用标准产品(Application Specific Standard Parts,简称ASSP)、片上系统(System on Chip,简称SOC)、复杂可编程逻辑设备(Complex Programmable Logic Device,简称CPLD)等等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦除可编程只读存储器(EPROM或快闪存储器)、光纤、便捷式紧凑盘只读存储器(CD-ROM)、光学储存设备、磁储存设备、或上述内容的任何合适组合。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。
此外,虽然采用特定次序描绘了各操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,虽然在上面论述中包含了若干具体实现细节,但是这些不应当被解释为对本公开的范围的限制。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的各种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。
尽管已经采用特定于结构特征和/或方法逻辑动作的语言描述了本主题,但是应当理解所附权利要求书中所限定的主题未必局限于上面描述的特定特征或动作。相反,上面所描述的特定特征和动作仅仅是实现权利要求书的示例形式。

Claims (16)

  1. 一种视频处理方法,包括:
    在视频处理界面中显示待处理视频以及至少一个贴图;
    接收针对目标贴图的选择操作,识别所述待处理视频中的目标对象的目标部位;
    在所述待处理视频中的所述目标部位上显示所述目标贴图;
    接收作用于所述目标贴图上的编辑操作,修改所述目标贴图;
    生成目标视频,所述目标视频中的所述目标部位上显示有所述目标贴图。
  2. 根据权利要求1所述的视频处理方法,其特征在于,所述目标贴图包括:动态贴图和静态贴图中的至少一种,所述目标贴图用于遮挡所述目标部位的至少部分区域。
  3. 根据权利要求1或2所述的视频处理方法,其中,所述接收针对目标贴图的选择操作,识别所述待处理视频中的目标对象的目标部位,包括:
    接收针对多个不同的目标贴图的选择操作,识别所述待处理视频中至少一个目标对象对应的多个目标部位;
    所述在所述目标部位上显示所述目标贴图,包括:
    在所述多个目标部位上分别显示所述不同的目标贴图。
  4. 根据权利要求1至3中任一项所述的视频处理方法,其中,所述识别所述待处理视频中的目标对象的目标部位,包括:
    通过图像识别技术检测所述待处理视频中目标对象的目标部位;
    或者,接收对所述待处理视频中的目标对象的选择操作,识别所述待处理视频中所述目标对象的目标部位;
    或者,接收对所述待处理视频中目标部位的选择操作,识别所述待处理视频中的所述目标部位。
  5. 根据权利要求1至4中任一项所述的视频处理方法,其中,在生成目标视频之前,还包括:
    生成第一预设时长的第一视频,在所述第一视频中目标对象的目标部位上显示有目标贴图,所述第一预设时长小于所述待处理视频的时长;
    在所述视频处理界面中展示所述第一视频。
  6. 根据权利要求3所述的视频处理方法,其中,在所述多个目标部位上分别显示所述多个不同的目标贴图之后,还包括:
    在所述视频处理界面中展示所述多个目标部位分别对应的多个时长控件,每个时长控件用于展示至少一个目标贴图在对应的目标部位上的显示信息。
  7. 根据权利要求6所述的视频处理方法,其中,所述生成目标视频,包括:
    接收针对所述时长控件的调节操作,调节所述时长控件对应的目标贴图在目标部位的显示信息;
    根据调节操作,生成目标视频。
  8. 根据权利要求6或7所述的视频处理方法,其中,在生成目标视频之前,还包括:
    接收针对所述待处理视频的视频分割和/或顺序调整操作,根据所述视频分割和/或所述顺序调整操作,分割和/或调整多个所述时长控件,获得多个目标时长控件;
    在所述视频处理界面展示多个所述目标时长控件和修改后的待处理视频,每个所述 目标时长控件用于展示修改后的待处理视频中对应的目标贴图在目标部位的显示信息。
  9. 根据权利要求1至8中任一项所述的视频处理方法,其中,还包括:
    在所述视频处理界面中展示生成所述目标视频的状态信息,所述状态信息包括所述目标视频的生成进度和当前生成进度对应的视频画面。
  10. 根据权利要求1至9中任一项所述的视频处理方法,其中,当所述目标贴图为动态贴图时,所述视频处理方法还包括:
    检测所述待处理视频中所述目标部位上至少部分区域的动作信息;
    控制所述目标部位上显示的目标贴图的对应区域跟随所述动作信息进行相应的运动,所述目标贴图的对应区域用于遮挡所述目标部位上至少部分区域。
  11. 根据权利要求3、6至8中任一项所述的视频处理方法,其中,在所述多个目标部位上分别显示所述多个不同的目标贴图,包括:
    获取目标部位的特征信息;
    根据所述特征信息,确定所述多个不同的目标贴图中每个目标贴图与所述目标部位的匹配度;
    在所述目标部位上显示与所述目标部位匹配度最高的目标贴图。
  12. 一种视频处理装置,包括:
    显示模块,在视频处理界面中显示待处理视频以及至少一个贴图;
    识别模块,接收针对目标贴图的选择操作,识别所述待处理视频中的目标对象的目标部位;
    所述显示模块,还用于在所述待处理视频中的所述目标部位上显示所述目标贴图;
    处理模块,用于接收作用于所述目标贴图上的编辑操作,修改所述目标贴图,生成目标视频,所述目标视频中的所述目标部位上显示有所述目标贴图。
  13. 一种电子设备,包括:至少一个处理器和存储器;
    所述存储器存储计算机执行指令;
    所述至少一个处理器执行所述存储器存储的计算机执行指令,使得所述至少一个处理器执行如权利要求1至11中任一项所述的视频处理方法。
  14. 一种计算机可读存储介质,其中,所述计算机可读存储介质中存储有计算机执行指令,当处理器执行所述计算机执行指令时,实现如权利要求1至11中任一项所述的视频处理方法。
  15. 一种计算机程序产品,包括:计算机程序,所述计算机程序被处理器执行时,实现如权利要求1至11中任一项所述的视频处理方法。
  16. 一种计算机程序,其中,所述计算机程序在被处理器执行时实现如权利要求1至11中任一项所述的视频处理方法。
PCT/CN2022/106110 2021-08-03 2022-07-15 视频处理方法、装置、设备及存储介质 WO2023011146A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP22851860.1A EP4344229A1 (en) 2021-08-03 2022-07-15 Video processing method and apparatus, device, and storage medium
US18/394,685 US20240129576A1 (en) 2021-08-03 2023-12-22 Video processing method, apparatus, device and storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110886909.1A CN113613067B (zh) 2021-08-03 2021-08-03 视频处理方法、装置、设备及存储介质
CN202110886909.1 2021-08-03

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US18/394,685 Continuation US20240129576A1 (en) 2021-08-03 2023-12-22 Video processing method, apparatus, device and storage medium

Publications (1)

Publication Number Publication Date
WO2023011146A1 true WO2023011146A1 (zh) 2023-02-09

Family

ID=78339300

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/106110 WO2023011146A1 (zh) 2021-08-03 2022-07-15 视频处理方法、装置、设备及存储介质

Country Status (4)

Country Link
US (1) US20240129576A1 (zh)
EP (1) EP4344229A1 (zh)
CN (1) CN113613067B (zh)
WO (1) WO2023011146A1 (zh)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113613067B (zh) * 2021-08-03 2023-08-22 北京字跳网络技术有限公司 视频处理方法、装置、设备及存储介质
CN114489897B (zh) * 2022-01-21 2023-08-08 北京字跳网络技术有限公司 一种对象处理方法、装置、终端设备及介质

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016028463A (ja) * 2014-07-10 2016-02-25 フリュー株式会社 写真シール作成装置および表示方法
CN105678686A (zh) * 2015-12-30 2016-06-15 北京金山安全软件有限公司 一种图片处理方法及装置
CN106777329A (zh) * 2017-01-11 2017-05-31 维沃移动通信有限公司 一种图像信息的处理方法及移动终端
CN109495790A (zh) * 2018-11-30 2019-03-19 北京字节跳动网络技术有限公司 基于编辑器的贴纸添加方法、装置、电子设备及可读介质
CN111556335A (zh) * 2020-04-15 2020-08-18 早安科技(广州)有限公司 一种视频贴纸处理方法及装置
CN112929683A (zh) * 2021-01-21 2021-06-08 广州虎牙科技有限公司 视频处理方法、装置、电子设备及存储介质
CN113613067A (zh) * 2021-08-03 2021-11-05 北京字跳网络技术有限公司 视频处理方法、装置、设备及存储介质

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2422973B (en) * 2005-02-04 2011-03-30 Quantel Ltd Multi-zonal video editing system
GB2495289A (en) * 2011-10-04 2013-04-10 David John Thomas Multimedia editing by string manipulation
CN106611412A (zh) * 2015-10-20 2017-05-03 成都理想境界科技有限公司 贴图视频生成方法及装置
CN108476289B (zh) * 2017-07-31 2021-02-02 深圳市大疆创新科技有限公司 一种视频处理方法、设备、飞行器及系统
CN107679497B (zh) * 2017-10-11 2023-06-27 山东新睿信息科技有限公司 视频面部贴图特效处理方法及生成系统
CN110062269A (zh) * 2018-01-18 2019-07-26 腾讯科技(深圳)有限公司 附加对象显示方法、装置及计算机设备
CN110675310B (zh) * 2019-07-02 2020-10-02 北京达佳互联信息技术有限公司 视频处理方法、装置、电子设备及存储介质
CN110582018B (zh) * 2019-09-16 2022-06-10 腾讯科技(深圳)有限公司 一种视频文件处理的方法、相关装置及设备
CN112822541B (zh) * 2019-11-18 2022-05-20 北京字节跳动网络技术有限公司 视频生成方法、装置、电子设备和计算机可读介质
CN111083354A (zh) * 2019-11-27 2020-04-28 维沃移动通信有限公司 一种视频录制方法及电子设备
CN111145308A (zh) * 2019-12-06 2020-05-12 北京达佳互联信息技术有限公司 一种贴纸获取方法和装置
CN111866592B (zh) * 2020-07-31 2022-09-20 掌阅科技股份有限公司 视频直播方法、计算设备及计算机存储介质
CN112199534A (zh) * 2020-10-10 2021-01-08 维沃移动通信有限公司 贴纸推荐方法、装置、电子设备及存储介质

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2016028463A (ja) * 2014-07-10 2016-02-25 フリュー株式会社 写真シール作成装置および表示方法
CN105678686A (zh) * 2015-12-30 2016-06-15 北京金山安全软件有限公司 一种图片处理方法及装置
CN106777329A (zh) * 2017-01-11 2017-05-31 维沃移动通信有限公司 一种图像信息的处理方法及移动终端
CN109495790A (zh) * 2018-11-30 2019-03-19 北京字节跳动网络技术有限公司 基于编辑器的贴纸添加方法、装置、电子设备及可读介质
CN111556335A (zh) * 2020-04-15 2020-08-18 早安科技(广州)有限公司 一种视频贴纸处理方法及装置
CN112929683A (zh) * 2021-01-21 2021-06-08 广州虎牙科技有限公司 视频处理方法、装置、电子设备及存储介质
CN113613067A (zh) * 2021-08-03 2021-11-05 北京字跳网络技术有限公司 视频处理方法、装置、设备及存储介质

Also Published As

Publication number Publication date
EP4344229A1 (en) 2024-03-27
CN113613067B (zh) 2023-08-22
CN113613067A (zh) 2021-11-05
US20240129576A1 (en) 2024-04-18

Similar Documents

Publication Publication Date Title
WO2023011146A1 (zh) 视频处理方法、装置、设备及存储介质
US11670339B2 (en) Video acquisition method and device, terminal and medium
US11037600B2 (en) Video processing method and apparatus, terminal and medium
US20150373281A1 (en) Systems and methods for identifying media portions of interest
US20230360184A1 (en) Image processing method and apparatus, and electronic device and computer-readable storage medium
CN112396679B (zh) 虚拟对象显示方法及装置、电子设备、介质
WO2021254502A1 (zh) 目标对象显示方法、装置及电子设备
WO2022170958A1 (zh) 基于增强现实的显示方法、设备、存储介质及程序产品
WO2020151491A1 (zh) 图像形变的控制方法、装置和硬件装置
US20230328197A1 (en) Display method and apparatus based on augmented reality, device, and storage medium
WO2023179346A1 (zh) 特效图像处理方法、装置、电子设备及存储介质
US20240119082A1 (en) Method, apparatus, device, readable storage medium and product for media content processing
US20230386001A1 (en) Image display method and apparatus, and device and medium
WO2023138548A1 (zh) 图像处理方法、装置、设备和存储介质
US11721046B2 (en) Customizing soundtracks and hairstyles in modifiable videos of multimedia messaging application
WO2023185671A1 (zh) 风格图像生成方法、装置、设备及介质
US20230133416A1 (en) Image processing method and apparatus, and device and medium
WO2024104333A1 (zh) 演播画面的处理方法、装置、电子设备及存储介质
WO2024046360A1 (zh) 媒体内容处理方法、装置、设备、可读存储介质及产品
WO2023202360A1 (zh) 视频预览方法、装置、可读介质及电子设备
WO2023138441A1 (zh) 视频生成方法、装置、设备及存储介质
JP7427786B2 (ja) 拡張現実に基づく表示方法、機器、記憶媒体及びプログラム製品
WO2023140787A2 (zh) 视频的处理方法、装置、电子设备、存储介质和程序产品
CN115499672B (zh) 图像显示方法、装置、设备及存储介质
WO2023226851A1 (zh) 三维效果形象的生成方法、装置、电子设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22851860

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2022851860

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2022851860

Country of ref document: EP

Effective date: 20231221

NENP Non-entry into the national phase

Ref country code: DE