WO2024016924A1 - 视频处理方法、装置、电子设备及存储介质 - Google Patents

视频处理方法、装置、电子设备及存储介质 Download PDF

Info

Publication number
WO2024016924A1
WO2024016924A1 PCT/CN2023/101016 CN2023101016W WO2024016924A1 WO 2024016924 A1 WO2024016924 A1 WO 2024016924A1 CN 2023101016 W CN2023101016 W CN 2023101016W WO 2024016924 A1 WO2024016924 A1 WO 2024016924A1
Authority
WO
WIPO (PCT)
Prior art keywords
target
control object
target control
display
display form
Prior art date
Application number
PCT/CN2023/101016
Other languages
English (en)
French (fr)
Inventor
赵双琳
杨启悦
高晶
周栩彬
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Publication of WO2024016924A1 publication Critical patent/WO2024016924A1/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/2621Cameras specially adapted for the electronic generation of special effects during image pickup, e.g. digital cameras, camcorders, video cameras having integrated special effects capability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects

Definitions

  • the embodiments of the present disclosure relate to the field of image processing technology, for example, to a video processing method, device, electronic device, and storage medium.
  • the present disclosure provides a video processing method, device, electronic equipment and storage medium to achieve the technical effect of improving the interactivity between special effect video content and users.
  • Embodiments of the present disclosure provide a video processing method.
  • the method includes: in response to a special effect triggering operation, sequentially collecting video frames to be processed; in the case of detecting a contact point on a display interface, determining whether the contact point is on the display interface.
  • the target display position in the video frame to be processed and add a target control object at the target display position; when it is detected that the special effects playback conditions are met, determine the display form of the target control object in turn; control the target control
  • the object is displayed in the video frame to be processed according to the corresponding display form.
  • Embodiments of the present disclosure also provide a video processing device, which includes: a video frame collection module, configured to sequentially collect video frames to be processed in response to a special effect triggering operation; an object adding module, configured to detect In the case of a touch point, determine the target display position of the touch point in the video frame to be processed, and add a target manipulation object at the target display position; the display form determination module is set to detect that the special effect playback is satisfied Under the conditions, the display form of the target control object is determined in turn; the special effects display module is configured to control the target control object according to the corresponding display form. The state is displayed in the video frame to be processed.
  • An embodiment of the present disclosure also provides an electronic device, wherein the electronic device includes: at least one processor; a storage device configured to store at least one program, and when the at least one program is executed by the at least one processor, such that The at least one processor implements the video processing method described in any one of the embodiments of the present disclosure.
  • Embodiments of the present disclosure also provide a storage medium containing computer-executable instructions, which when executed by a computer processor are used to perform the video processing method as described in any embodiment of the present invention.
  • Figure 1 is a schematic flow chart of a video processing method provided by an embodiment of the present disclosure
  • Figure 2 is a schematic flow chart of a video processing method provided by an embodiment of the present disclosure
  • Figure 3 is a schematic structural diagram of a video processing device provided by an embodiment of the present disclosure.
  • FIG. 4 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • the term “include” and its variations are open-ended, ie, “including but not limited to.”
  • the term “based on” means “based at least in part on.”
  • the term “one embodiment” means “at least one embodiment”; the term “another embodiment” means “at least one additional embodiment”; and the term “some embodiments” means “at least some embodiments”. Relevant definitions of other terms will be given in the description below.
  • a prompt message is sent to the user to clearly remind the user that the operation requested will require the acquisition and use of the user's personal information. Therefore, users can autonomously choose whether to provide personal information to software or hardware such as electronic devices, applications, servers or storage media that perform the operations of the technical solution of the present disclosure based on the prompt information.
  • the method of sending prompt information to the user may be, for example, a pop-up window, and the prompt information may be presented in the form of text in the pop-up window.
  • the pop-up window can also contain a selection control for the user to choose "agree” or "disagree” to provide personal information to the electronic device.
  • the technical solution provided by the embodiments of the present disclosure can be applied in any scene where special effects video needs to be generated.
  • the interactivity between the user and the display interface can be increased, and the personality of the special effects video can be improved.
  • ization and interactivity with users can also be a scene generated by any special effects video in the related art, and a special effects video in which dominoes or any blocky objects in the scene are continuously pushed down can be simulated and displayed.
  • FIG. 1 is a schematic flowchart of a video processing method provided by an embodiment of the present disclosure.
  • This embodiment can be applied to situations where users use special effects props to generate corresponding special effects videos.
  • This method can be executed by a video processing device, and the device can use software And/or implemented in the form of hardware.
  • the hardware can be an electronic device, such as a mobile terminal, a personal computer (Personal Computer, PC) or a server.
  • the scene of special effects video display is usually implemented by the cooperation of the client and the server.
  • the method provided in this embodiment can be executed by the server, the client, or the client and the server cooperate.
  • the method includes the following steps.
  • the device for executing the special effects video processing method provided by the embodiments of the present disclosure can be integrated into a device that supports special effects.
  • Video processing function application software and the software can be installed in an electronic device.
  • the electronic device can be a mobile terminal or a PC.
  • the application software may be a type of software for image/video processing.
  • the application software will not be described in detail here, as long as it can realize image/video processing. It can also be a specially developed application to add special effects and display the special effects in the software, or it can be integrated in the corresponding page. Users can process the special effects video through the page integrated in the PC.
  • a control for triggering the special effects can be developed in advance.
  • the special effects triggering operation can be responded to, thus starting Collect video frames to be processed in the scene to which the terminal device belongs. That is to say, special effects props corresponding to the embodiments of the present disclosure can be created, and when the triggering of the special effects props is detected, the video frames to be processed can be collected.
  • the camera device can be called to collect video frames to be processed based on the camera device. That is, the video frame to be processed is captured based on the camera device.
  • the touch point can be a touch point determined in the display interface after the user triggers a special effect in the application software to place the target control object.
  • the touch point is also a key point associated with a characteristic of the user, such as , the key point can be the tip of the nose.
  • the current position information of the touch point is the data determined by the application software or application program for the position of the touch point in the display interface, for example, the two-dimensional coordinates of the touch point in the display interface.
  • the user can use a feature of the face (such as the nose) to determine the touch point in the display interface that is used to determine the addition of the target manipulation object.
  • the user can also use the touch point based on the Facial features move as touch points in the display interface to determine the target display position of the touch points in the display interface; or, the user can determine the target control object in the display interface through touch operations with fingers on the display interface. The target display position, and then place the corresponding target control object at the target display position.
  • the target manipulation object can be understood as an object placed at the touch point, for example, it can be a block object, such as a domino, etc.
  • the target display position of the touch point in the video frame to be processed is determined, and a target manipulation object is added at the target display position.
  • determining the target display position of the touch point in the video frame to be processed includes: determining the touch point when a trigger operation on the display interface is detected; and determining the target display position based on the touch point.
  • the trigger operation can be a trigger operation of a finger on the display interface, or a key point's stay point on the display interface.
  • the target display position can be the position where the target control object is placed in the AR scene.
  • the user's nose is preset as the target key point. Based on this, when the user collects a facial image through the front camera of the mobile device, the features in the facial image can be identified based on a pre-trained facial recognition algorithm.
  • the nose feature which can be used as a target key point
  • the duration of the feature's stay in the display interface can be tracked. If the stay time reaches the preset stay duration threshold, optionally 200ms, the feature can be The coordinates of this feature in the display interface are used as the target display position.
  • the trigger operation of the touch point on the display interface can be detected.
  • the pressure values corresponding to multiple positions on the display interface are detected, and the positions corresponding to the pressure values higher than the preset pressure value threshold are used as touch points, and This location is displayed as the target location.
  • the user can click the display interface according to actual needs to determine the target display position of the touch point in the video frame to be processed on the click display interface, and place the preset target control object at the target display position. If the number of touch points includes multiple, it means that the number of target manipulation objects also includes multiple.
  • the special effects playback condition can be understood as the condition that requires controlling the placement of at least one target control object for animation playback.
  • the display form can include the upright state or the tiled state of the target control object.
  • the state corresponding to the target control object is regarded as the upright state, and the tiled state is a state different from the upright state.
  • the display form of the target control object can be determined in sequence, and after the display form is determined, the target control object can be sequentially controlled to be displayed according to its corresponding display form.
  • the display form of the target control object can be determined based on the display position of the target control object on the display interface.
  • the display form of the target control object can be determined based on the coordinate information of the target control object on the display interface. sequence information.
  • the display order information of the target control object can be determined based on the placement timestamp of the target control object in the display interface, and then the display form of the corresponding target control object can be determined.
  • the special effects playback conditions may be: the duration of adding the target control object reaches the second preset duration; and triggering a control that completes the addition of the target control object.
  • the duration threshold for adding the target control object can be set in advance, optionally, 30s, or 1 minute, etc., and this duration threshold can be used as the second preset duration.
  • the countdown module can be started to add the target manipulation object to the video frame to be processed within the second preset time period. When the countdown reaches 0, it means that the target control object has been added and the target operation can be controlled.
  • the controlled objects are displayed with special effects according to the determined display form.
  • the user can add the target control object in the video frame to be processed according to actual needs.
  • the completion control on the display interface can be triggered to determine that the target control object has been completed when the trigger completion control is detected.
  • the target control object can be controlled to be displayed according to the determined display form, thereby obtaining a special effects video.
  • the above method can place items within a limited time, or give the user sufficient time to place items, achieving the same controllable effect as the duration of item placement in actual scenes.
  • the target control object when adding a target control object to a video frame to be processed, can be used as foreground information, and the video frames to be processed in sequence can be used as background information to obtain the final special effects video.
  • control target manipulation object is displayed in the video frame to be processed according to the corresponding display form.
  • the target control object is building blocks or dominoes placed according to a certain pattern or randomly.
  • the display form of the building block can be determined starting from the first building block.
  • the display form is In the tiled form, control the adjustment of the first building block from the upright state to the tiled state, and then determine the display form of the second building block, and so on to obtain the special effects video, achieving the effect of simulating the display of dominoes in the scene.
  • the display form of the first building block is in the upright state, it is determined that the first building block will not have a corresponding impact on the second building block, and all subsequent building blocks will be processed in the upright state.
  • the initial building block does not start from the first one, but starts from the nth one, you can determine the display form of the nth building block, and determine n+1 and n+2 based on the display form of the nth building block. ...the display form of the building blocks, down to the last building block.
  • the technical solution provided by the embodiment of the present disclosure sequentially collects the video frames to be processed in response to the special effect triggering operation.
  • the target display position of the contact point in the video frame to be processed can be determined, and Add the target control object at the target display position.
  • the display form of the target control object can be determined in turn, and the target control object can be controlled to be displayed in the video frame to be processed according to the corresponding display form, realizing a system that can be based on
  • the user's trigger operation determines the placement of the target control object, that is, it achieves the effect of interactivity between the user and the display interface.
  • the target control object can be controlled to display accordingly in turn.
  • the display is carried out in different forms, which improves the effect of richness of the picture content.
  • FIG. 2 is a schematic flowchart of a video processing method provided by an embodiment of the present disclosure.
  • Special effects interaction can be realized not only in ordinary videos, but also in AR scenes.
  • Special effects interaction for example, when collecting the video frame to be processed, determine the target virtual plane located in the video frame to be processed, so as to add the target control object on the same target virtual plane based on the touch point, so as to realize the target control object and the target virtual plane to be processed.
  • the effect of video frame fusion please refer to the description of the embodiments of the present disclosure. The technical terms that are the same as or corresponding to the above embodiments will not be described again here.
  • the method includes the following steps.
  • the video frames to be processed can be collected in sequence. Or, after detecting the trigger start control, the video frames to be processed can be collected sequentially.
  • the process of sequentially collecting video frames to be processed also includes: displaying at least one trajectory to be selected; using the trajectory to be selected that is last triggered within the first preset time period as the target trajectory, and setting the target trajectory corresponding to the target trajectory.
  • the guidance map is displayed on the display interface, so that the contact point determines the target display position of the target control object based on the guidance map; or, if at least one trajectory to be selected is not triggered within the first preset time period, at least one trajectory to be selected is removed from the to-be-selected trajectory. Process video frame removal.
  • each track to be selected can show the display effect of the associated pattern.
  • the application can display the to-be-selected track 1 corresponding to the text pattern and the to-be-selected track 2 corresponding to the image pattern in the relevant display interface.
  • a template selection countdown is also displayed in the display interface, and the countdown time is the preset selection duration. It can be understood that the user can select from multiple trajectory templates to be selected by triggering the operation. When the selection duration reaches the first preset duration, that is, when the template selection countdown returns to zero, the trajectory to be selected selected by the user at this moment is the target. trajectory.
  • the user can also issue a confirmation instruction based on the determination control, thereby directly ending the template selection countdown process and using the currently selected trajectory to be selected as the target trajectory.
  • the user can select multiple templates in sequence according to their own wishes.
  • the application will only use the track to be selected last selected by the user as the target track.
  • the method of determining the target trajectory can be selected according to the actual situation, and the embodiments of the present disclosure are not limited here. Setting at least one track to be selected can guide the user to place content with a more graphic feel, further improving the effect of interactivity between the user and the display interface.
  • the trajectory to be selected is not selected within the first preset time period, it means that the user needs to display items in conjunction with the trajectory. At this time, the trajectory to be selected can be removed from the video frame to be processed, so that the user can customize it according to his or her needs. Play to determine the placement of the target control object.
  • a guidance map corresponding to the target trajectory may be displayed.
  • Guidance diagrams are used to guide users and help them place corresponding target control objects.
  • the guidance diagram can be displayed on the display interface in the form of a dotted line, and there can be corresponding arrows on the guidance diagram to guide the user to place the corresponding target control object based on the arrows.
  • a cover-up policy is also pre-deployed for the situation where the user does not make a selection. If at least one track to be selected is not selected within the preset time period, the track to be selected is removed from the video frame to be processed. It can be understood that if the user does not select any trajectory to be selected within the preset time period, the trajectory to be selected in the display interface will be cleared. In this case, the user can manipulate the target object as he wishes in the video frame to be processed. Arrange in any shape.
  • the terminal device refers to the device corresponding to the video frame to be processed.
  • the spatial position information may be position information in the world coordinate system.
  • the gyroscope and/or inertial measurement unit provided in the terminal device can determine the angle information of the target terminal device in the world coordinate system and multiple planes formed by the coordinate axis. According to the spatial position information and angle information of the terminal device, the virtual plane corresponding to the terminal device is determined and used as the target virtual plane.
  • the target manipulation object When a touch point is detected on the display interface, the target manipulation object can be placed on the target virtual plane according to the display position of the touch point in the video frame to be processed.
  • This setting can adjust all target control objects to be on the same horizontal plane, thereby determining the sequence information of the target control objects, and then determining the display form corresponding to the target control objects based on the sequence information.
  • the user can trigger any position on the display interface and use this trigger point as a contact point.
  • the target display position of the contact point on the target virtual plane in the video frame to be processed can be determined to add the target manipulation object at the display position.
  • the number of target control objects includes as many touch points as there are, that is, the number of touch points is consistent with the number of target control points.
  • a recall function can be set to remove the target control object from the video frame to be processed based on the recall function.
  • the target control object when it is detected that the triggering operation on the target control object satisfies the withdrawal condition, the target control object is removed from the video frame to be processed; where the withdrawal condition includes the triggering operation of clicking the withdrawal control, double-clicking the target control object and at least one of the operations of long-pressing the target control object.
  • the trigger operation on the target control object can be detected in real time. If the trigger operation on the target object meets the object withdrawal condition, the corresponding target control object can be removed from the video frame to be processed.
  • the withdrawal condition may be that the display interface includes a control corresponding to the withdrawal function.
  • the newly added target manipulation object will be removed from the pending processing. removed from the video frame. If the control is continuously clicked, the corresponding target control objects can be withdrawn in sequence based on the timestamp of adding the target control object.
  • the triggering target manipulation object is detected and the recall control is triggered, the newly triggered target manipulation object is removed from the video frame to be processed.
  • the corresponding target control object can be set according to actual needs, which improves the effect of interactivity with the display interface.
  • the target control object before sequentially determining the display form of the target control object, it may be: determining the preset motion attributes corresponding to the target control object, so as to sequentially determine the display form of the target control object based on the motion attributes; or, The sliding operation of the display interface is determined, the movement direction of the target control object is determined based on the sliding operation, and the display form of the target control object is determined based on the movement direction.
  • the motion attribute can be understood as: determining the display form of the target manipulation object includes at least two implementation methods. The first implementation method is determined based on the preset motion attributes, and the second implementation method is based on the user's trigger on the display interface. Operation to determine the display form of multiple target control objects.
  • the motion attributes may be the motion directions corresponding to a series of target control objects set in the development stage and the initial movement speed of the first target control object located at the starting position.
  • Display forms include upright or tiled forms. When placing corresponding target control objects at all target display positions, the initial form of all target control objects is an upright form.
  • the starting target control object among all target control objects can be determined.
  • the display form of the initial target control object changes from an upright state to a tiled state.
  • the second target control object can be determined based on the existing momentum conservation or energy conservation and the distance information between the first target control object and the second target control object.
  • the two goals control the display form of the object.
  • the display form of the third target control object is associated with the display form of the second target control object, that is, the display form of the next target control object can be determined based on the display form of the previous target control object.
  • the above method can be used to simulate the effect of dominoes falling one after another in the display environment.
  • the sliding operation on determines the display form of the corresponding target control object.
  • the sliding operation may be a finger sliding on the display interface. For example, swipe to the left, interact to the right, etc.
  • the sliding direction corresponding to the sliding operation is determined, and the initial speed corresponding to the sliding operation is determined based on the end point corresponding to the sliding operation.
  • a first target control object is determined from multiple target control objects.
  • the display forms corresponding to the multiple target control objects are determined in sequence.
  • the sliding direction is determined to be the left.
  • the rightmost target manipulation object in the video frame to be processed can be used as the first target manipulation object.
  • the movement speed of the first target control object is consistent with the speed at the end of the sliding operation, and the sliding direction is the direction of movement to the left. Based on this, the display form of all target control objects can be determined in sequence.
  • determining the starting control object can simulate the triggering of an object in the real environment.
  • the object can be used as the starting control point to sequentially control subsequent items to appear in a state of falling down, which improves the special effects video content. Realistic effect.
  • the target control object corresponding to the sliding operation can be determined, and the target control object can be determined as multiple objects in sequence.
  • the starting control object of the target control object can be determined.
  • the starting control object can be understood as the first object whose display form needs to be determined.
  • the target control object corresponding to the sliding operation can be used as the initial control object, and the display form corresponding to the corresponding target control object is determined in sequence according to the initial speed and sliding direction corresponding to the sliding operation.
  • determining the display form of the corresponding target control object based on the physics engine may be: if the target control object is the starting control object, then determining the motion information and display of the target control object based on the sliding operation of the starting control object. Form; according to the motion information and motion direction of the target control object, determine the motion information and display form of the next target control object adjacent to the target control object; among them, the motion information includes movement speed, and the display form includes an upright state or a flat state. shop status.
  • the starting control object is determined based on the physics engine, and the initial speed and sliding direction (movement direction) are determined based on the sliding operation of the starting control object, and the display of the starting control object is determined. form, and controls the initial control object to be displayed according to its display form.
  • the next target control object is determined according to the sliding direction, and based on the distance information between the next target control object and the starting control object, the quality information of the starting control object, and when the display form of the starting control object is in a tiled state
  • the corresponding motion information determines the display form of the next target control object, and controls the next target control object to display according to the display form determined at this time, and so on to determine the display form of multiple target control objects.
  • Using the above method to determine the display form of the corresponding target control object can simulate the effect of multiple dominoes falling continuously in the real environment, which not only improves the interaction between the display interface and the user, but also improves the authenticity of the display screen.
  • the next control object will The movement speed of the display form corresponding to the target control object is zero, and the display form is in an upright state. Multiple target control objects after the next target control object are consistent with the initial placement state.
  • the display form of the next target control object can be determined only by combining the movement speed of the previous target control object.
  • control target manipulation object is displayed in the video frame to be processed according to the corresponding display form.
  • the video frame to be processed is displayed as background information.
  • the technical solution provided by the embodiments of the present disclosure can determine the target virtual plane according to the spatial position information of the terminal device when collecting video frames to be processed, so that when a touch point is detected on the display interface, the location of the touch point can be determined.
  • the target display position on the target virtual plane, and the target control object is placed on the target virtual plane corresponding to the target display position, enabling the production of corresponding special effects videos in the AR scene, and improving the interaction between the user and the display interface sexual effect.
  • Figure 3 is a schematic structural diagram of a video processing device provided by an embodiment of the present disclosure. As shown in Figure 3, the device includes: a video frame collection module 310, an object addition module 320, a display form determination module 330, and a special effects display module 340. .
  • the video frame collection module 310 is configured to sequentially collect the video frames to be processed in response to the special effects triggering operation;
  • the object adding module 320 is configured to, when a contact point on the display interface is detected, determine that the contact point is in the video to be processed The target display position in the frame, and add the target control object at the target display position;
  • the display form determination module 330 is configured to determine the display form of the target control object in sequence when it is detected that the special effects playback conditions are met;
  • special effects display Module 340 is configured to control the target manipulation object to be displayed in the video frame to be processed according to the corresponding display form.
  • the technical solution provided by the embodiment of the present disclosure sequentially collects the video frames to be processed in response to the special effect triggering operation.
  • the target display position of the contact point in the video frame to be processed can be determined, and Add the target control object at the target display position.
  • the display form of the target control object can be determined in turn, and the target control object can be controlled to be displayed in the video frame to be processed according to the corresponding display form, realizing a system that can be based on
  • the user's trigger operation determines the placement of the target control object, that is, it achieves the effect of interactivity between the user and the display interface.
  • multiple target control objects can be controlled in sequence according to the corresponding display form. Display, improving the richness of the screen content.
  • the device further includes: a virtual plane creation module, configured to determine the target virtual plane located in the video to be processed based on the spatial location information of the terminal device, so as to based on the contact point, in The target manipulation object is added to the target virtual plane.
  • a virtual plane creation module configured to determine the target virtual plane located in the video to be processed based on the spatial location information of the terminal device, so as to based on the contact point, in The target manipulation object is added to the target virtual plane.
  • the object adding module includes: a contact point determination unit configured to determine the contact point when a trigger operation on the display interface is detected; a display position determination unit configured to determine the contact point based on The contact point determines the target display position.
  • the device further includes: a trajectory display module, configured to display at least one trajectory to be selected; a first trajectory selection module, configured to display the last triggered trajectory within the first preset time period to be selected.
  • the trajectory is used as the target trajectory, and the guidance map corresponding to the target trajectory is displayed on the display interface, so that the contact point determines the target display position of the target manipulation object based on the guidance map; or, a trajectory removal module, It is configured that if the at least one track to be selected is not triggered within the first preset time period, the at least one selected track is removed from the video frame to be processed.
  • the device further includes: a control object removal module configured to remove the target control object from the target control object when it is detected that the triggering operation on the target control object satisfies the withdrawal condition. removed from the video frame to be processed; wherein the withdrawal condition includes the triggering operation being at least one of an operation of clicking the withdrawal control, an operation of double-clicking the target manipulation object, and an operation of long-pressing the target manipulation object.
  • a control object removal module configured to remove the target control object from the target control object when it is detected that the triggering operation on the target control object satisfies the withdrawal condition. removed from the video frame to be processed; wherein the withdrawal condition includes the triggering operation being at least one of an operation of clicking the withdrawal control, an operation of double-clicking the target manipulation object, and an operation of long-pressing the target manipulation object.
  • the special effects playback conditions include: the duration of adding the target control object reaches the second preset duration; and triggering a control to complete the addition of the target control object.
  • the display form determination module before sequentially determining the display form of the target control object, includes: a first motion attribute determination unit configured to determine the preset and the target control object.
  • the motion attributes corresponding to the object are used to sequentially determine the display form of the target manipulation object based on the motion attributes; or, the second motion attribute determination unit is configured to determine the sliding operation of the display interface and determine the display form based on the sliding The operation is to determine the movement direction of the target control object, so as to sequentially determine the display form of the target control object based on the movement direction.
  • the device further includes: a starting control object determination module, configured to determine the target control object corresponding to the sliding operation, and use the target control object corresponding to the sliding operation as The initial control object of the display form of the target control object is determined in sequence.
  • a starting control object determination module configured to determine the target control object corresponding to the sliding operation, and use the target control object corresponding to the sliding operation as The initial control object of the display form of the target control object is determined in sequence.
  • the device further includes: if the target control object corresponding to the sliding operation is a starting control object, determining the sliding operation based on the sliding operation on the starting control object.
  • the movement information and display form of the target control object corresponding to the operation according to the movement information and movement direction of the target control object corresponding to the sliding operation, determine the next target adjacent to the target control object corresponding to the sliding operation.
  • the motion information and display form of the manipulation object wherein the motion information includes movement speed, and the display form includes an upright state or a tiled state.
  • the device further includes: determining, based on the movement information and movement direction of the target control object corresponding to the sliding operation, the target control object adjacent to the sliding operation.
  • the motion information and display form of the next target control object include: the motion information corresponding to the display form of the target control object corresponding to the sliding operation, the distance information corresponding to the next target control object, and the The quality information of the next target control object is determined, and the movement information and corresponding display form of the next target control object are determined.
  • the special effects display module displays the video frame to be processed as background information during the process of displaying the target manipulation object according to the corresponding display form.
  • the special effects processing device provided by the embodiments of the present disclosure can execute the special effects processing method provided by any embodiment of the present disclosure, and has functional modules corresponding to the execution method.
  • the multiple units and modules included in the above device are only divided according to functional logic, but are not limited to the above division, as long as the corresponding functions can be achieved; in addition, the names of the multiple functional units are also Just for easy differentiation from each other.
  • FIG. 4 is a schematic structural diagram of an electronic device provided by an embodiment of the present disclosure.
  • Terminal devices in the embodiments of the present disclosure may include, but are not limited to, mobile phones, notebook computers, digital broadcast receivers, personal digital assistants (Personal Digital Assistant, PDA), tablet computers (Portable Android Device, PAD), portable multimedia players Mobile terminals such as (Portable Media Player, PMP), vehicle-mounted terminals (such as vehicle-mounted navigation terminals), and fixed terminals such as digital television (TV), desktop computers, etc.
  • PDA Personal Digital Assistant
  • PMP Portable Multimedia Players
  • vehicle-mounted terminals such as vehicle-mounted navigation terminals
  • fixed terminals such as digital television (TV), desktop computers, etc.
  • the electronic device 400 shown in FIG. 4 is only an example.
  • the electronic device 400 may include a processing device (eg, central processing unit, graphics processing unit). processor, etc.) 401, which can execute multiple programs according to the program stored in the read-only memory (Read-Only Memory, ROM) 402 or loaded from the storage device 408 into the random access memory (Random Access Memory, RAM) 403. appropriate actions and handling.
  • ROM read-only memory
  • RAM random access memory
  • various programs and data required for the operation of the electronic device 400 are also stored.
  • the processing device 401, ROM 402 and RAM 403 are connected to each other via a bus 404.
  • An input/output (I/O) interface 405 is also connected to bus 404.
  • the following devices can be connected to the I/O interface 405: input devices 406 including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; including, for example, a Liquid Crystal Display (LCD) , an output device 407 such as a speaker, a vibrator, etc.; a storage device 408 including a magnetic tape, a hard disk, etc.; and a communication device 409.
  • the communication device 409 may allow the electronic device 400 to communicate wirelessly or wiredly with other devices to exchange data.
  • FIG. 4 illustrates electronic device 400 with various means, it should be understood that implementation or availability of all illustrated means is not required. More or fewer means may alternatively be implemented or provided.
  • embodiments of the present disclosure include a computer program product including a computer program carried on a non-transitory computer-readable medium, the computer program containing program code for performing the method illustrated in the flowchart.
  • the computer program may be downloaded and installed from the network via communication device 409, or from storage device 408, or from ROM 402.
  • the processing device 401 When the computer program is executed by the processing device 401, the above-mentioned functions defined in the method of the embodiment of the present disclosure are performed.
  • Embodiments of the present disclosure provide a computer storage medium on which a computer program is stored.
  • the program is executed by a processor, the video processing method provided by the above embodiments is implemented.
  • the computer-readable medium mentioned above in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the above two.
  • the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device or device, or any combination thereof.
  • Examples of computer-readable storage media may include, but are not limited to: electrical connections having one or more wires, portable computer disks, hard drives, RAM, ROM, Erasable Programmable Read-Only Memory (EPROM) ) Or flash memory, optical fiber, portable compact disk read-only memory (Compact Disc Read-Only Memory, CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
  • a computer-readable storage medium may be any tangible medium that contains or stores a program for use by or in connection with an instruction execution system, apparatus, or device.
  • a computer-readable signal medium may include a data signal propagated in baseband or as part of a carrier wave, carrying computer-readable program code therein. Such propagated data signals may take a variety of forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above.
  • a computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium that can send, propagate, or transmit a program for use by or in connection with an instruction execution system, apparatus, or device .
  • Program code contained on a computer-readable medium can be transmitted using any appropriate medium, including but not limited to: wires, optical cables, radio frequency (Radio Frequency, RF), etc., or any suitable combination of the above.
  • the client and server can communicate using any currently known or future developed network protocol, such as HyperText Transfer Protocol (HTTP), and can communicate with digital data in any form or medium.
  • HTTP HyperText Transfer Protocol
  • Communications e.g., communications network
  • Examples of communication networks include Local Area Networks (LANs), Wide Area Networks (WANs), the Internet (e.g., the Internet), and end-to-end networks (e.g., ad hoc end-to-end networks), as well as any current network for knowledge or future research and development.
  • LANs Local Area Networks
  • WANs Wide Area Networks
  • the Internet e.g., the Internet
  • end-to-end networks e.g., ad hoc end-to-end networks
  • the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; it may also exist independently without being assembled into the electronic device.
  • the above-mentioned computer-readable medium carries one or more programs.
  • the electronic device responds to the special effect triggering operation, sequentially collects the video frames to be processed; when detecting the display When the interface is touched, determine the target display position of the touch point in the video frame to be processed, and add a target control object at the target display position; when it is detected that the special effects playback conditions are met, determine the The display form of the target control object; controlling the target control object to be displayed in the video frame to be processed according to the corresponding display form.
  • Computer program code for performing the operations of the present disclosure may be written in one or more programming languages, including but not limited to object-oriented programming languages—such as Java, Smalltalk, C++, and Includes conventional procedural programming languages—such as "C" or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computers can be connected through any kind of network, including LAN or WAN. to the user's computer, or may be connected to an external computer (such as via an Internet connection using an Internet service provider).
  • each block in the flowchart or block diagram may represent a module, segment, or portion of code that contains one or more logic functions that implement the specified executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown one after another may actually execute substantially in parallel, or they may sometimes execute in the reverse order, depending on the functionality involved.
  • each block of the block diagram and/or flowchart illustration, and combinations of blocks in the block diagram and/or flowchart illustration can be implemented by special purpose hardware-based systems that perform the specified functions or operations. , or can be implemented using a combination of specialized hardware and computer instructions.
  • the units involved in the embodiments of the present disclosure can be implemented in software or hardware.
  • the name of the unit does not constitute a limitation on the unit itself.
  • the first acquisition unit may also be described as "a unit that acquires at least two Internet Protocol addresses.”
  • exemplary types of hardware logic components include: field programmable gate array (Field Programmable Gate Array, FPGA), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), application specific standard product (Application Specific Standard Parts (ASSP), System on Chip (SOC), Complex Programmable Logic Device (CPLD), etc.
  • a machine-readable medium may be a tangible medium that may contain or store a program for use by or in connection with an instruction execution system, apparatus, or device.
  • the machine-readable medium may be a machine-readable signal medium or a machine-readable storage medium.
  • Machine-readable media may include, but are not limited to, electronic, magnetic, optical, electromagnetic, infrared, or semiconductor systems, devices or devices, or any suitable combination of the foregoing. More specific examples of machine-readable storage media would include an electrical connection based on one or more wires, a portable computer disk, a hard disk, RAM, ROM, EPROM or flash memory, optical fiber, CD-ROM, optical storage device, magnetic storage device , or any suitable combination of the above.

Abstract

本公开实施例提供了一种视频处理方法、装置、电子设备及存储介质,该方法包括:响应于特效触发操作,依次采集待处理视频帧;在检测到于显示界面的触点的情况下,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;在检测到满足特效播放条件的情况下,依次确定所述目标操控对象的展示形态;控制所述目标操控对象依据相应的展示形态在所述待处理视频帧中进行展示。

Description

视频处理方法、装置、电子设备及存储介质
本申请要求在2022年07月20日提交中国专利局、申请号为202210862523.1的中国专利申请的优先权,该申请的全部内容通过引用结合在本申请中。
技术领域
本公开实施例涉及图像处理技术领域,例如涉及一种视频处理方法、装置、电子设备及存储介质。
背景技术
随着网络技术的发展,越来越多的应用程序进入了用户的生活,例如,一系列可以拍摄短视频的软件,深受用户的喜爱。
为了提高视频拍摄的趣味性,软件开发商可以研发各种各样的特效道具,然而,为用户提供的特效道具十分有限,视频内容的丰富性有待进一步提升,例如,在增强现实(Augmented Reality,AR)领域,基于相关的特效道具所生成的特效视频效果不佳,同时,与用户之间的交互性也不强,从而引起用户使用体验不佳的问题。
发明内容
本公开提供一种视频处理方法、装置、电子设备及存储介质,以实现提高特效视频内容与用户之间交互性的技术效果。
本公开实施例提供了一种视频处理方法,该方法包括:响应于特效触发操作,依次采集待处理视频帧;在检测到于显示界面的触点的情况下,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;在检测到满足特效播放条件的情况下,依次确定所述目标操控对象的展示形态;控制所述目标操控对象依据相应的展示形态在待处理视频帧中进行展示。
本公开实施例还提供了一种视频处理装置,该装置包括:视频帧采集模块,设置为响应于特效触发操作,依次采集待处理视频帧;对象添加模块,设置为在检测到于显示界面的触点的情况下,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;展示形态确定模块,设置为在检测到满足特效播放条件的情况下,依次确定所述目标操控对象的展示形态;特效展示模块,设置为控制所述目标操控对象依据相应的展示形 态在待处理视频帧中进行展示。
本公开实施例还提供了一种电子设备,其所述电子设备包括:至少一个处理器;存储装置,设置为存储至少一个程序,当所述至少一个程序被所述至少一个处理器执行,使得所述至少一个处理器实现如本公开实施例任一所述的视频处理方法。
本公开实施例还提供了一种包含计算机可执行指令的存储介质,所述计算机可执行指令在由计算机处理器执行时用于执行如本发明实施例任一所述的视频处理方法。
附图说明
图1为本公开实施例所提供的一种视频处理方法流程示意图;
图2为本公开实施例所提供的一种视频处理方法流程示意图;
图3为本公开实施例所提供的一种视频处理装置结构示意图;
图4为本公开实施例所提供的一种电子设备的结构示意图。
具体实施方式
下面将参照附图描述本公开的实施例。虽然附图中显示了本公开的一些实施例,然而应当理解的是,本公开可以通过多种形式来实现,而且不应该被解释为限于这里阐述的实施例。。
应当理解,本公开的方法实施方式中记载的多个步骤可以按照不同的顺序执行,和/或并行执行。此外,方法实施方式可以包括附加的步骤和/或省略执行示出的步骤。
本文使用的术语“包括”及其变形是开放性包括,即“包括但不限于”。术语“基于”是“至少部分地基于”。术语“一个实施例”表示“至少一个实施例”;术语“另一实施例”表示“至少一个另外的实施例”;术语“一些实施例”表示“至少一些实施例”。其他术语的相关定义将在下文描述中给出。
需要注意,本公开中提及的“第一”、“第二”等概念仅用于对不同的装置、模块或单元进行区分,并非用于限定这些装置、模块或单元所执行的功能的顺序或者相互依存关系。
需要注意,本公开中提及的“一个”、“多个”的修饰是示意性而非限制性的,本领域技术人员应当理解,除非在上下文另有明确指出,否则应该理解为“一个或多个”。
本公开实施方式中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的。
可以理解的是,在使用本公开多个实施例公开的技术方案之前,均应当依据相关法律法规通过恰当的方式对本公开所涉及个人信息的类型、使用范围、使用场景等告知用户并获得用户的授权。
例如,在响应于接收到用户的主动请求时,向用户发送提示信息,以明确地提示用户,其请求执行的操作将需要获取和使用到用户的个人信息。从而,使得用户可以根据提示信息来自主地选择是否向执行本公开技术方案的操作的电子设备、应用程序、服务器或存储介质等软件或硬件提供个人信息。
作为一种可选的但非限定性的实现方式,响应于接收到用户的主动请求,向用户发送提示信息的方式例如可以是弹窗的方式,弹窗中可以以文字的方式呈现提示信息。此外,弹窗中还可以承载供用户选择“同意”或者“不同意”向电子设备提供个人信息的选择控件。
可以理解的是,上述通知和获取用户授权过程仅是示意性的,不对本公开的实现方式构成限定,其它满足相关法律法规的方式也可应用于本公开的实现方式中。
可以理解的是,本技术方案所涉及的数据(包括但不限于数据本身、数据的获取或使用)应当遵循相应法律法规及相关规定的要求。
在介绍本技术方案之前,可以先对应用场景进行示例性说明。可以将本公开实施例所提供的技术方案应用在任意需要生成特效视频的场景中,例如,在AR特效视频生成的场景中,可以增加用户与显示界面之间的交互性,提高了特效视频个性化以及与用户之间互动性的效果。还可以是相关技术中任意特效视频生成的场景中,可以模拟显示场景中的骨牌或者任意块状物品被连续推倒的特效视频。
图1为本公开实施例所提供的一种视频处理方法流程示意图,本实施例可适用于用户利用特效道具,生成相应特效视频情况,该方法可以由视频处理装置来执行,该装置可以通过软件和/或硬件的形式实现,该硬件可以是电子设备,如移动终端、个人计算机(Personal Computer,PC)端或服务器等。特效视频展示的场景通常是由客户端和服务器来配合实现的,本实施例所提供的方法可以由服务端来执行,客户端来执行,或者是客户端和服务端的配合来执行。
如图1所示,所述方法包括如下步骤。
S110、响应于特效触发操作,依次采集待处理视频帧。
执行本公开实施例提供的特效视频处理方法的装置,可以集成在支持特效 视频处理功能的应用软件中,且该软件可以安装至电子设备中,可选的,电子设备可以是移动终端或者PC端等。应用软件可以是对图像/视频处理的一类软件,其应用软件在此不再一一赘述,只要可以实现图像/视频处理即可。还可以是专门研发的应用程序,来实现添加特效并将特效进行展示的软件中,亦或是集成在相应的页面中,用户可以通过PC端中集成的页面来实现对特效视频的处理。
在本实施例中,在支持特效视频处理功能的应用软件或应用程序中,可以预先开发用于触发特效的控件,当检测到用户触发该控件时,即可对特效触发操作进行响应,从而开始采集终端设备所属场景的待处理视频帧。也就是说,可以创建与本公开实施例所对应的特效道具,当检测到触发该特效道具后,可以采集待处理视频帧。
用户可以触发相应的特效道具后,可以调用摄像装置,以基于摄像装置采集待处理视频帧。即,基于摄像装置拍摄待处理视频帧。
S120、当检测到于显示界面的触点时,确定触点于待处理视频帧中的目标显示位置,并在目标显示位置处添加目标操控对象。
触点可以是用户在应用软件中触发一种特效后,在显示界面中确定出来的、用于放置目标操控对象的触摸点,该触摸点同时也是与用户自身一特征相关联的关键点,例如,关键点可以是鼻尖。触点的当前位置信息,即是应用软件或应用程序针对触点在显示界面中的位置所确定出的数据,例如,触点在显示界面中的二维坐标。
可选的,针对于所触发的特效来说,用户可以利用面部的一特征(如鼻子)确定出显示界面中作为确定添加目标操控对象的触点,在后续的过程中,用户还可以基于该面部特征作为触点在显示界面中进行移动,从而确定触点于显示界面中的目标显示位置;或者,用户可以通过手指在显示界面上的触控操作,确定出显示界面中摆放目标操控对象的目标显示位置,进而在目标显示位置处摆放相应的目标操控对象。
目标操控对象可以理解为摆放在触摸点的对象,例如,可以是块状物体,如,骨牌等。
在检测到显示界面中包括触点时,确定该触点在待处理视频帧中的目标显示位置,并在目标显示位置处添加目标操控对象。
在本实施例中,确定触点于待处理视频帧中的目标显示位置,包括:当检测到于显示界面的触发操作时,确定所述触点;基于所述触点,确定目标显示位置。
触发操作可以是手指于显示界面上的触发操作,或者是,关键点于显示界面的停留点。目标显示位置可以是AR场景中摆放目标操控对象的位置。
示例性的,预先设置用户的鼻子作为目标关键点,基于此,当用户通过移动设备的前置摄像头采集到面部图像时,可以基于预先训练好的面部识别算法对面部图像中的特征进行识别,当在面部图像中识别出可以作为目标关键点的鼻子这一特征时,可以追踪该特征于显示界面中的停留时长,若停留时长达到预设停留时长阈值,可选的,200ms,即可将该特征在显示界面中的坐标作为目标显示位置。可以检测触摸点于显示界面的触发操作,可选的,检测显示界面上多个位置所对应的压力值,并将高于预设压力值阈值的压力值所对应的位置作为触摸点,并将该位置作为目标显示位置。
可以理解为,用户可以根据实际需求点击显示界面,以在点击显示界面上确定该触摸点于待处理视频帧中的目标显示位置,并将预先设定的目标操控对象摆放在目标显示位置。如果触摸点的数量包括多个,则说明目标操控对象的数量也包括多个。
S130、在检测到满足特效播放条件时,依次确定目标操控对象的展示形态。
特效播放条件可以理解为需要控制摆放的至少一个目标操控对象进行动画播放的条件。展示形态可以包括目标操控对象直立状态或者平铺状态。将基于目标显示位置摆放目标操控对象时目标操控对象所对应的状态作为直立状态,平铺状态为区别于直立状态的状态。
可以理解为,在检测到满足特效播放条件时,可以依次确定目标操控对象的展示形态,以在确定其展示形态后,依次控制目标操控对象按照其相应的展示形态进行展示。
需要说明的是,依次确定目标操控对象的展示形态,可以依据目标操控对象于显示界面的展示位置来确定,如,可以结合目标操控对象于显示界面中的坐标信息,确定目标操控对象的展示形态的顺序信息。还可以是,依据目标操控对象于显示界面中的摆放时间戳,确定目标操控对象展示的顺序信息,进而确定相应目标操控对象的展示形态。
在本公开实施例中,特效播放条件可以是:添加所述目标操控对象的时长达到第二预设时长;触发所述目标操控对象添加完成的控件。
可以理解为:可以预先设置添加目标操控对象的时长阈值,可选的,30s,或者1min等,可以将此时长阈值作为第二预设时长。在采集待处理视频帧的过程中,可以启动倒计时模块,以在第二预设时长内在待处理视频帧中添加目标操控对象。当倒计时为0时,说明目标操控对象已添加完成,可以控制目标操 控对象按照确定出的展示形态进行特效展示。另一种方式为:用户可以根据实际需求在待处理视频帧中添加目标操控对象,在添加完毕之后,可以触发显示界面上的完成控件,以在检测到触发完成控件时,确定目标操控对象已添加完成,可以控制目标操控对象按照确定出的展示形态进行展示,从而得到特效视频。
上述方式,可以在有限的时间内进行物品的摆放,或者是,给予用户充足的时间进行物品摆放,实现了与实际场景中物品摆放时长可控性的效果。
还需要说明的是,在待处理视频帧中添加目标操控对象时,可以将目标操控对象作为前景信息,将依次采集的待处理视频帧作为背景信息,得到最终的特效视频。
S140、控制目标操控对象依据相应的展示形态在待处理视频帧中进行展示。
可以理解为:在确定相应目标操控对象所对应的展示形态时,可以将其按照确定出的展示形态进行展示。
示例性的,目标操控对象为按照一定规律或者随机摆放的积木或者骨牌,当检测到满足特效播放条件时,可以从第一个积木开始确定该积木的展示形态,可选的,展示形态为平铺形态,则控制第一个积木由直立状态调整为平铺状态,接下来确定第二个积木的展示形态,依次类推得到特效视频,达到了模拟出显示场景中多米诺骨牌的效果。当然,如果第一个积木的展示形态为直立状态,则确定第一个积木不会对第二个积木产生相应的影响,后续所有的积木均处理直立状态。当然,如果初始积木不是从第一个开始的,而是从第n个开始,则可以确定第n个积木的展示形态,并根据第n个积木的展示形态,确定n+1、n+2...积木的展示形态,直至最后一个积木。
本公开实施例所提供的技术方案,响应于特效触发操作,依次采集待处理视频帧,在检测到用户触发显示界面时,可以确定该触点于待处理视频帧中的目标显示位置,并在目标显示位置处添加目标操控对象,在检测到满足特效播放条件时,可以依次确定目标操控对象的展示形态,并控制目标操控对象依据相应的展示形态在待处理视频帧中展示,实现了可以基于用户的触发操作确定目标操控对象的摆放位置,即,达到了用户与显示界面之间交互性的效果,进一步的,在满足特效播放条件的基础上,可以控制目标操控对象依次按照相应的展示形态进行展示,提高了画面内容丰富性的效果。
图2为本公开实施例所提供的一种视频处理方法流程示意图,在前述实施例的基础上,不仅可以在普通的视频中实现特效互动,还可以在AR场景中实现 特效互动,例如,在采集待处理视频帧时,确定位于待处理视频帧中的目标虚拟平面,以基于触控点将目标操控对象添加在同一目标虚拟平面上,以实现目标操控对象与待处理视频帧相融合的效果,其实施方式可以参见本公开实施例的阐述,其中,与上述实施例相同或者相应的技术术语在此不再赘述。
如图2所示,所述方法包括如下步骤。
S210、响应于特效触发操作,依次采集待处理视频帧。
可以理解为,当检测触发特效道具所对应的控件时,可以依次采集待处理视频帧。或者是,在检测到触发开始控件后,可以依次采集待处理视频帧。
在本实施例中,在依次采集待处理视频帧的过程中,还包括:显示至少一个待选择轨迹;将第一预设时长内最后触发的待选择轨迹作为目标轨迹,并将目标轨迹对应的引导图展示于显示界面上,以使触点基于引导图确定目标操控对象的目标显示位置;或,若在第一预设时长内未触发至少一个待选择轨迹,则将至少一个选择轨迹从待处理视频帧中移除。
待选择轨迹可以有一个或多个,每个待选择轨迹都可以呈现出所关联图案的展示效果。当用户触发一种特效后,应用可以在相关显示界面中显示出对应文字图案的待选择轨迹1,以及对应图像图案的待选择轨迹2。在本实施例中,向用户提供特效相关联的待选择轨迹的同时,还会在显示界面中展示模板选择倒计时,倒计时时间即是预先设置的选择时长。可以理解,用户可以通过触发操作在多个待选择轨迹模版进行选择,当选择时长达到第一预设时长,即,模板选择倒计时归零时,用户在这一时刻选择的待选择轨迹即是目标轨迹。当然,在未达到预设时长时,用户也可以基于确定控件下发确认指令,从而直接结束模板选择倒计时的计时过程,并将当前选择的待选择轨迹作为目标轨迹,本领域技术人员应当理解,一方面,当待选择轨迹模版有多个时,用户可以按照自身意愿对多个模板依次进行选择,但选择的时间达到预设时长后,应用只会将用户最后选择的待选择轨迹作为目标轨迹;另一方面,确定目标轨迹的方式可以根据实际情况进行选择,本公开实施例在此不做的限定。设置至少一个待选择轨迹可以引导用户摆放处更具有画面感的内容,进一步,提高用户与显示界面之间交互性的效果。
当然,如果在第一预设时长内未选择待选择轨迹时,则说明用户需要结合轨迹进行物品展示,此时,可以将待选择轨迹从待处理视频帧中移除,以便于用户根据需求自己发挥,来确定目标操控对象的摆放位置。
如果用户选择了目标轨迹,可以显示与目标轨迹所对应的引导图。
引导图是用于引导用户以及帮助用户如何摆放相应的目标操控对象的导引 图。引导图可以是虚线的形式展示于显示界面上,并且引导图上可以有相应的箭头,以基于箭头引导用户去摆放相应的目标操控对象。
需要说明的是,在本实施例中,还针对用户未进行选择的情况预先部署了兜底策略。若预设时长内未选择至少一个待选择轨迹,将待选择轨迹从待处理视频帧中移除。可以理解,如果用户在预设时长内没有选择任何待选择轨迹,则会清空显示界面中的待选择轨迹,在这种情况下,用户可以在待处理视频帧中按照自己的意愿将目标操控对象摆放为任何形状。
S220、基于所获得的终端设备的空间位置信息,确定位于待处理视频中的目标虚拟平面,以基于触点,在目标虚拟平面上添加目标操控对象。
终端设备是指采集待处理视频帧所对应的设备。空间位置信息可以是在世界坐标系下的位置信息。终端设备中设置的陀螺仪和/或惯性测量单元,可以确定目标终端设备在世界坐标系下与坐标轴构成的多个平面的夹角信息。根据终端设备的空间位置信息,夹角信息,确定终端设备所对应的虚拟平面,并将其作为目标虚拟平面。
在检测到于显示界面上存在触点时,可以根据触点于待处理视频帧中的显示位置,并将目标操控对象放置在目标虚拟平面上。这样设置可以调整所有目标操控对象位于同一水平面上,从而确定目标操控对象的先后顺序信息,进而依据先后顺序信息确定出目标操控对象所对应的展示形态。
S230、当检测到于显示界面的触点时,确定触点于待处理视频帧中目标虚拟平面的目标显示位置,并在目标显示位置处添加目标操控对象。
可以理解为:用户可以触发显示界面上的任意位置,将此触发点作为触点。可以确定该触点于待处理视频帧中目标虚拟平面的目标显示位置,以在显示位置处添加目标操控对象。触摸点存在多少个,目标操控对象的数量就包括多少个,即触摸点与目标操控点的数量相一致。
在实际应用中,可能存在目标操控对象的摆放位置与实际需求不相符的问题,为了解决此类问题,可以设置撤回功能,以基于撤回功能将目标操控对象从待处理视频帧中移除。
可选的,当检测到对目标操控对象的触发操作满足撤回条件时,将目标操控对象从待处理视频帧中移除;其中,撤回条件包括触发操作为点击撤回控件的操作、双击目标操控对象的操作以及长按目标操控对象的操作中的至少一种。
可以理解为,可以实时检测对目标操控对象的触发操作,如果对目标对象的触发操作满足对象撤回条件,则可以将相应的目标操控对象从待处理视频帧中移除。
在本公开实施例中,撤回条件可以是,显示界面上包括与撤回功能相对应的控件,当检测到添加完目标操控对象之后,触发了撤回控件,则将刚添加的目标操控对象从待处理视频帧中移除。若连续点击率撤回控件,则可以依据添加目标操控对象的时间戳,依次撤回相应目标操控对象。还可以是,当检测到触发目标操控对象之后,又触发了撤回控件,则将刚触发的目标操控对象从待处理视频帧中移除。还可以是:当检测到双击目标操控对象或者对目标操控对象的触发时长达到相应的预设按压时长,则说明要将此时触发的目标操控对象从待处理视频帧中移除。设置此种方式,可以按照实际需求设置相应的目标操控对象,提高了与显示界面交互性的效果。
S240、在检测到满足特效播放条件时,依次确定目标操控对象的展示形态。
在本公开实施例中,在依次确定目标操控对象的展示形态之前,可以是:确定预先设置的与目标操控对象相对应的运动属性,以基于运动属性依次确定目标操控对象的展示形态;或,确定于显示界面的滑动操作,并基于滑动操作确定目标操控对象的运动方向,以基于运动方向依次确定目标操控对象的展示形态。
运动属性可以理解为:确定目标操控对象的展示形态包括至少两种实现方式,第一种实现方式是基于预先设置的运动属性来确定的,第二种实现方式为根据用户于显示界面上的触发操作来确定多个目标操控对象的展示形态的。
接下来分别介绍两种实现方式是如何来确定展示形态的。
运动属性可以是在开发阶段设置的一系列目标操控对象所对应的运动方向以及位于起始位置处第一个目标操控对象的初始运动速度。展示形态包括直立形态或平铺形态。在所有目标显示位置摆放相应目标操控对象时,所有目标操控对象的初始形态均为直立形态。
可以理解为:当检测到满足特效播放条件时,可以确定所有目标操控对象中的起始目标操控对象。依据预先设置的初始运动速度和运动方向,确定起始目标操控对象的展示形态由直立状态转变为平铺状态。接下来,确定第二个目标操控对象的展示形态,此时,可以基于现有的动量守恒或能量守恒、以及第一个目标操控对象和第二个目标操控对象之间的间距信息,确定第二个目标操控对象的展示形态。相应的,第三个目标操控对象的展示形态与第二个目标操控对象的展示形态相关联,即可以依据前一个目标操控对象的展示形态,确定下一目标操控对象的展示形态。
采用上述方式可以模拟出展示环境中骨牌依次倒下的效果。
为了进一步提高用户与显示界面之间的交互性,可以基于用户与显示界面 上的滑动操作确定相应目标操控对象的展示形态。
滑动操作可以是手指于显示界面的滑动。例如,偏向左侧滑动、偏向右侧互动等。
当检测到满足特效播放条件,且检测到于显示界面的滑动操作时,确定所述滑动操作所对应的滑动方向,以及根据滑动操作所对应的结束点,确定与滑动操作相对应的初始速度。基于滑动方向,从多个目标操控对象中确定出第一个目标操控对象。基于第一个目标操控对象、滑动方向和初始速度,依次确定多个目标操控对象所对应的展示形态,其确定方式可以参见上述表述,在此不再赘述。
示例性的,如果滑动操作对应的为偏向左侧的滑动操作,则确定滑动方向为左侧,此时可以将待处理视频帧中位于最右侧的目标操控对象作为第一个目标操作对象。此时,第一个目标操控对象的运动速度于滑动操作结束时的速度相一致,滑动方向为向左侧运动的方向,基于此,可以依次确定所有目标操控对象的展示形态。
在本实施例中,确定起始操控对象,可以模拟出真实环境中触发一个物体后,可以以该物体为起始控制点,依次控制后续的物品出现依次倒下的状态,提高了特效视频内容真实感的效果。
在实际应用中,可能存在用户触发的是所有目标操控对象中的任意一个,为了模拟出此时的效果,可以确定与滑动操作所对应的目标操控对象,并将该目标操控对象作为依次确定多个目标操控对象的起始操控对象。
起始操控对象可以理解为第一个需要确定其展示形态的对象。
可以理解为,可以将滑动操作所对应的目标操控对象作为起始操控对象,并根据滑动操作所对应的初始速度和滑动方向,依次确定相应目标操控对象所对应的展示形态。
需要确定的是,在确定起始操控对象之后,可以基于物理引擎确定相应目标操控对应的展示形态,进而控制目标操控对象按照相应的展示形态进行展示。在本实施例中,基于物理引擎确定相应目标操控对象的展示形态可以是:若目标操控对象为起始操控对象,则基于对起始操控对象的滑动操作,确定目标操控对象的运动信息和展示形态;根据目标操控对象的运动信息和运动方向,确定与目标操控对象相邻的下一目标操控对象的运动信息和展示形态;其中,运动信息中包括运动速度,展示形态中包括直立状态或平铺状态。
可以理解为,基于物理引擎确定起始操控对象,并根据对起始操控对象的滑动操作,确定初始速度和滑动方向(运动方向),确定起始操控对象的展示 形态,并控制该起始操控对象按照其展示形态进行展示。接下来,根据滑动方向确定下一目标操控对象,并根据下一目标操控对象的与起始操控对象的间距信息、起始操控对象的质量信息以及起始操控对象的展示形态为平铺状态时所对应的运动信息,确定下一目标操控对象的展示形态,并控制下一目标操控对象按照此时确定出的展示形态进行展示,以此类推确定多个目标操控对象的展示形态。采用上述方式确定相应目标操控对象展示形态,可以模拟出现实环境中多个骨牌连续倒下的效果,不仅提高了显示界面与用户之间的互动性,也提高了显示画面真实性的效果。
需要说明的是,如果初始操控对象与下一目标操控对象之间的间距信息大于目标操控对象本身的高度信息,则不管初始操控对象转换为平铺状态是所对应的运动速度有多大,下一目标操控对象所对应的展示形态的运动速度为零,展示形态为直立状态,下一目标操控对象之后的多个目标操控对象均与初始摆放的状态相一致。
还需要说明的是,如果不考虑多个目标操控对象之间的距离信息,则可以仅结合前一目标操控对象的运动速度,来确定下一目标操控对象的展示形态。
S250、控制目标操控对象依据相应的展示形态在待处理视频帧中进行展示。
可以理解为:将目标操控对象依据其相应的展示形态进行展示的过程中,将待处理视频帧作为背景信息进行显示。
本公开实施例所提供的技术方案,在采集待处理视频帧时,可以根据终端设备的空间位置信息,确定目标虚拟平面,以在检测到显示界面上包括触点时,可以确定该触点于目标虚拟平面上的目标显示位置,并将目标操控对象摆放在与目标显示位置相对应的目标虚拟平面上,实现了在AR场景中可以制作相应的特效视频,并且提高了用户与显示界面互动性的效果。
图3为本公开实施例所提供的一种视频处理装置结构示意图,如图3所示,所述装置包括:视频帧采集模块310、对象添加模块320、展示形态确定模块330以及特效展示模块340。
视频帧采集模块310,设置为响应于特效触发操作,依次采集待处理视频帧;对象添加模块320,设置为当检测到于显示界面的触点时,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;展示形态确定模块330,设置为在检测到满足特效播放条件时,依次确定所述目标操控对象的展示形态;特效展示模块340,设置为控制所述目标操控对象依据相应的展示形态在所述待处理视频帧中进行展示。
本公开实施例所提供的技术方案,响应于特效触发操作,依次采集待处理视频帧,在检测到用户触发显示界面时,可以确定该触点于待处理视频帧中的目标显示位置,并在目标显示位置处添加目标操控对象,在检测到满足特效播放条件时,可以依次确定目标操控对象的展示形态,并控制目标操控对象依据相应的展示形态在待处理视频帧中展示,实现了可以基于用户的触发操作确定目标操控对象的摆放位置,即,达到了用户与显示界面之间交互性的效果,在满足特效播放条件的基础上,可以控制多个目标操控对象依次按照相应的展示形态进行展示,提高了画面内容丰富性的效果。
在上述技术方案的基础上,所述装置还包括:虚拟平面创建模块,设置为基于终端设备的空间位置信息,确定位于所述待处理视频中的目标虚拟平面,以基于所述触点,在所述目标虚拟平面上添加所述目标操控对象。
在上述多个技术方案的基础上,所述对象添加模块中包括:触点确定单元,设置为当检测到于显示界面的触发操作时,确定所述触点;显示位置确定单元,设置为基于所述触点,确定所述目标显示位置。
在上述多个技术方案的基础上,所述装置还包括:轨迹显示模块,设置为显示至少一种待选择轨迹;第一轨迹选择模块,设置为将第一预设时长内最后触发的待选择轨迹作为目标轨迹,并将所述目标轨迹对应的引导图展示于显示界面上,以使所述触点基于所述引导图确定所述目标操控对象的目标显示位置;或,轨迹移除模块,设置为若在所述第一预设时长内未触发所述至少一个待选择轨迹,则将所述至少一个选择轨迹从所述待处理视频帧中移除。
在上述多个技术方案的基础上,所述装置还包括:操控对象移除模块,设置为当检测到对所述目标操控对象的触发操作满足撤回条件时,将所述目标操控对象从所述待处理视频帧中移除;其中,所述撤回条件包括所述触发操作为点击撤回控件的操作、双击所述目标操控对象的操作以及长按所述目标操控对象的操作中的至少一种。
在上述多个技术方案的基础上,所述特效播放条件包括:添加所述目标操控对象的时长达到第二预设时长;触发所述目标操控对象添加完成的控件。
在上述多个技术方案的基础上,所述展示形态确定模块,在依次确定所述目标操控对象的展示形态之前,包括:第一运动属性确定单元,设置为确定预先设置的与所述目标操控对象相对应的运动属性,以基于所述运动属性依次确定所述目标操控对象的展示形态;或,第二运动属性确定单元,设置为确定于所述显示界面的滑动操作,并基于所述滑动操作确定所述目标操控对象的运动方向,以基于所述运动方向依次确定所述目标操控对象的展示形态。
在上述多个技术方案的基础上,所述装置还包括:起始操控对象确定模块,设置为确定所述滑动操作所对应的目标操控对象,并将所述滑动操作所对应的目标操控对象作为依次确定所述目标操控对象的展示形态的起始操控对象。
在上述多个技术方案的基础上,所述装置还包括:若所述滑动操作所对应的目标操控对象为起始操控对象,则基于对所述起始操控对象的滑动操作,确定所述滑动操作所对应的目标操控对象的运动信息和展示形态;根据所述滑动操作所对应的目标操控对象的运动信息和运动方向,确定与所述滑动操作所对应的目标操控对象相邻的下一目标操控对象的运动信息和展示形态;其中,所述运动信息中包括运动速度,所述展示形态中包括直立状态或平铺状态。
在上述多个技术方案的基础上,所述装置还包括:所述根据所述滑动操作所对应的目标操控对象的运动信息和运动方向,确定与所述滑动操作所对应的目标操控对象相邻的下一目标操控对象的运动信息和展示形态,包括:根据所述滑动操作所对应的目标操控对象的展示形态所对应的运动信息、与所述下一目标操控对象相对应的间距信息以及所述下一目标操控对象的质量信息,确定所述下一目标操控对象的运动信息以及相应的展示形态。
在上述多个技术方案的基础上,所述特效展示模块:将所述目标操控对象依据相应的展示形态进行展示的过程中,将所述待处理视频帧作为背景信息进行显示。
本公开实施例所提供的特效处理装置可执行本公开任意实施例所提供的特效处理方法,具备执行方法相应的功能模块。
值得注意的是,上述装置所包括的多个单元和模块只是按照功能逻辑进行划分的,但并不局限于上述的划分,只要能够实现相应的功能即可;另外,多个功能单元的名称也只是为了便于相互区分。
图4为本公开实施例所提供的一种电子设备的结构示意图。下面参考图4,其示出了适于用来实现本公开实施例的电子设备(例如图4中的终端设备或服务器)400的结构示意图。本公开实施例中的终端设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、个人数字助理(Personal Digital Assistant,PDA)、平板电脑(Portable Android Device,PAD)、便携式多媒体播放器(Portable Media Player,PMP)、车载终端(例如车载导航终端)等的移动终端以及诸如数字电视(Television,TV)、台式计算机等的固定终端。图4示出的电子设备400仅仅是一个示例。
如图4所示,电子设备400可以包括处理装置(例如中央处理器、图形处 理器等)401,其可以根据存储在只读存储器(Read-Only Memory,ROM)402中的程序或者从存储装置408加载到随机访问存储器(Random Access Memory,RAM)403中的程序而执行多种适当的动作和处理。在RAM 403中,还存储有电子设备400操作所需的多种程序和数据。处理装置401、ROM 402以及RAM 403通过总线404彼此相连。输入/输出(Input/Output,I/O)接口405也连接至总线404。
通常,以下装置可以连接至I/O接口405:包括例如触摸屏、触摸板、键盘、鼠标、摄像头、麦克风、加速度计、陀螺仪等的输入装置406;包括例如液晶显示器(Liquid Crystal Display,LCD)、扬声器、振动器等的输出装置407;包括例如磁带、硬盘等的存储装置408;以及通信装置409。通信装置409可以允许电子设备400与其他设备进行无线或有线通信以交换数据。虽然图4示出了具有多种装置的电子设备400,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在非暂态计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置409从网络上被下载和安装,或者从存储装置408被安装,或者从ROM 402被安装。在该计算机程序被处理装置401执行时,执行本公开实施例的方法中限定的上述功能。
本公开实施方式中的多个装置之间所交互的消息或者信息的名称仅用于说明性的目的。
本公开实施例提供的电子设备与上述实施例提供的视频处理方法属于同一发明构思,未在本实施例中详尽描述的技术细节可参见上述实施例。
本公开实施例提供了一种计算机存储介质,其上存储有计算机程序,该程序被处理器执行时实现上述实施例所提供的视频处理方法。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的系统、装置或器件,或者任意以上的组合。计算机可读存储介质的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、RAM、ROM、可擦式可编程只读存储器(Erasable Programmable Read-Only Memory,EPROM) 或闪存、光纤、便携式紧凑磁盘只读存储器(Compact Disc Read-Only Memory,CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机可读信号介质可以发送、传播或者传输用于由指令执行系统、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、射频(Radio Frequency,RF)等等,或者上述的任意合适的组合。
在一些实施方式中,客户端、服务器可以利用诸如超文本传输协议(HyperText Transfer Protocol,HTTP)之类的任何当前已知或未来研发的网络协议进行通信,并且可以与任意形式或介质的数字数据通信(例如,通信网络)互连。通信网络的示例包括局域网(Local Area Network,LAN),广域网(Wide Area Network,WAN”,网际网(例如,互联网)以及端对端网络(例如,ad hoc端对端网络),以及任何当前已知或未来研发的网络。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:响应于特效触发操作,依次采集待处理视频帧;当检测到于显示界面的触点时,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;在检测到满足特效播放条件时,依次确定所述目标操控对象的展示形态;控制所述目标操控对象依据相应的展示形态在待处理视频帧中进行展示。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括但不限于面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括LAN或WAN—连 接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开多种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。单元的名称在一种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际协议地址的单元”。
本文中以上描述的功能可以至少部分地由一个或多个硬件逻辑部件来执行。例如,非限制性地,可以使用的示范类型的硬件逻辑部件包括:现场可编程门阵列(Field Programmable Gate Array,FPGA)、专用集成电路(Application Specific Integrated Circuit,ASIC)、专用标准产品(Application Specific Standard Parts,ASSP)、片上系统(System on Chip,SOC)、复杂可编程逻辑设备(Complex Programmable Logic Device,CPLD)等。
在本公开的上下文中,机器可读介质可以是有形的介质,其可以包含或存储以供指令执行系统、装置或设备使用或与指令执行系统、装置或设备结合地使用的程序。机器可读介质可以是机器可读信号介质或机器可读储存介质。机器可读介质可以包括但不限于电子的、磁性的、光学的、电磁的、红外的、或半导体系统、装置或设备,或者上述内容的任何合适组合。机器可读存储介质的更具体示例会包括基于一个或多个线的电气连接、便携式计算机盘、硬盘、RAM、ROM、EPROM或快闪存储器、光纤、CD-ROM、光学储存设备、磁储存设备、或上述内容的任何合适组合。
以上描述仅为本公开的实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技 术方案。
此外,虽然采用特定次序描绘了多个操作,但是这不应当理解为要求这些操作以所示出的特定次序或以顺序次序执行来执行。在一定环境下,多任务和并行处理可能是有利的。同样地,在上面论述中包含了多个实现细节。在单独的实施例的上下文中描述的某些特征还可以组合地实现在单个实施例中。相反地,在单个实施例的上下文中描述的多种特征也可以单独地或以任何合适的子组合的方式实现在多个实施例中。

Claims (14)

  1. 一种视频处理方法,包括:
    响应于特效触发操作,依次采集待处理视频帧;
    在检测到于显示界面的触点的情况下,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;
    在检测到满足特效播放条件的情况下,依次确定所述目标操控对象的展示形态;
    控制所述目标操控对象依据相应的展示形态在所述待处理视频帧中进行展示。
  2. 根据权利要求1所述的方法,还包括:
    基于所获得的终端设备的空间位置信息,确定位于所述待处理视频中的目标虚拟平面,以基于所述触点,在所述目标虚拟平面上添加所述目标操控对象。
  3. 根据权利要求1所述的方法,其中,所述确定所述触点于所述待处理视频帧中的目标显示位置,包括:
    在检测到于所述显示界面的触发操作的情况下,确定所述触点;
    基于所述触点,确定所述目标显示位置。
  4. 根据权利要求1所述的方法,在依次采集所述待处理视频帧的过程中,还包括:
    显示至少一个待选择轨迹;
    将第一预设时长内最后触发的待选择轨迹作为目标轨迹,并将所述目标轨迹对应的引导图展示于所述显示界面上,以基于所述引导图确定所述目标操控对象的目标显示位置;或,在所述第一预设时长内未触发所述至少一个待选择轨迹的情况下,将所述至少一个待选择轨迹从所述待处理视频帧中移除。
  5. 根据权利要求1所述的方法,还包括:
    在检测到对所述目标操控对象的触发操作满足撤回条件的情况下,将所述目标操控对象从所述待处理视频帧中移除;
    其中,所述撤回条件包括所述触发操作为点击撤回控件的操作、双击所述目标操控对象的操作以及长按所述目标操控对象的操作中的至少一种。
  6. 根据权利要求1所述的方法,其中,所述特效播放条件包括下述至少一种:
    添加所述目标操控对象的时长达到第二预设时长;
    触发所述目标操控对象添加完成的控件。
  7. 根据权利要求1所述的方法,在所述依次确定所述目标操控对象的展示形态之前,还包括:
    确定预先设置的与所述目标操控对象相对应的运动属性,以基于所述运动属性依次确定所述目标操控对象的展示形态;或,
    确定于所述显示界面的滑动操作,并基于所述滑动操作确定所述目标操控对象的运动方向,以基于所述运动方向依次确定所述目标操控对象的展示形态。
  8. 根据权利要求7所述的方法,还包括:
    确定所述滑动操作所对应的目标操控对象,并将所述滑动操作所对应的目标操控对象作为依次确定所述目标操控对象的展示形态的起始操控对象。
  9. 根据权利要求1所述的方法,其中,所述目标操控对象的数量为多个,所述依次确定所述目标操控对象的展示形态,包括:
    在所述目标操控对象中的一个目标操控对象为起始操控对象的情况下,基于对所述起始操控对象的滑动操作,确定所述一个目标操控对象的运动信息和展示形态;根据所述一个目标操控对象的运动信息和运动方向,确定与所述一个目标操控对象相邻的下一目标操控对象的运动信息和展示形态,直至依次确定所述多个目标操控对象的展示形态;
    其中,所述运动信息中包括运动速度,所述展示形态中包括直立状态或平铺状态。
  10. 根据权利要求9所述的方法,其中,所述根据所述滑动操作所对应的目标操控对象的运动信息和运动方向,确定与所述滑动操作所对应的目标操控对象相邻的下一目标操控对象的运动信息和展示形态,包括:
    根据所述滑动操作所对应的目标操控对象的展示形态所对应的运动信息、与所述下一目标操控对象相对应的间距信息以及所述下一目标操控对象的质量信息,确定所述下一目标操控对象的运动信息以及相应的展示形态。
  11. 根据权利要求1所述的方法,其中,所述控制所述目标操控对象依据相应的展示形态在所述待处理视频帧中进行展示,包括:
    将所述目标操控对象依据相应的展示形态进行展示的过程中,将所述待处理视频帧作为背景信息进行显示。
  12. 一种视频处理装置,包括:
    视频帧采集模块,设置为响应于特效触发操作,依次采集待处理视频帧;
    对象添加模块,设置为在检测到于显示界面的触点的情况下,确定所述触点于所述待处理视频帧中的目标显示位置,并在所述目标显示位置处添加目标操控对象;
    展示形态确定模块,设置为在检测到满足特效播放条件的情况下,依次确定所述目标操控对象的展示形态;
    特效展示模块,设置为控制所述目标操控对象依据相应的展示形态在待处理视频帧中进行展示。
  13. 一种电子设备,包括:
    至少一个处理器;
    存储装置,设置为存储至少一个程序,
    当所述至少一个程序被所述至少一个处理器执行,使得所述至少一个处理器实现如权利要求1-11中任一所述的视频处理方法。
  14. 一种包含计算机可执行指令的存储介质,所述计算机可执行指令在由计算机处理器执行时用于执行如权利要求1-11中任一所述的视频处理方法。
PCT/CN2023/101016 2022-07-20 2023-06-19 视频处理方法、装置、电子设备及存储介质 WO2024016924A1 (zh)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210862523.1A CN115278107A (zh) 2022-07-20 2022-07-20 视频处理方法、装置、电子设备及存储介质
CN202210862523.1 2022-07-20

Publications (1)

Publication Number Publication Date
WO2024016924A1 true WO2024016924A1 (zh) 2024-01-25

Family

ID=83767974

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/101016 WO2024016924A1 (zh) 2022-07-20 2023-06-19 视频处理方法、装置、电子设备及存储介质

Country Status (2)

Country Link
CN (1) CN115278107A (zh)
WO (1) WO2024016924A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115278107A (zh) * 2022-07-20 2022-11-01 北京字跳网络技术有限公司 视频处理方法、装置、电子设备及存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021217398A1 (zh) * 2020-04-28 2021-11-04 深圳市大疆创新科技有限公司 图像的处理方法及装置、可移动平台及其控制终端、计算机可读存储介质
CN114401443A (zh) * 2022-01-24 2022-04-26 脸萌有限公司 特效视频处理方法、装置、电子设备及存储介质
CN114531553A (zh) * 2022-02-11 2022-05-24 北京字跳网络技术有限公司 生成特效视频的方法、装置、电子设备及存储介质
CN114567805A (zh) * 2022-02-24 2022-05-31 北京字跳网络技术有限公司 确定特效视频的方法、装置、电子设备及存储介质
CN115278107A (zh) * 2022-07-20 2022-11-01 北京字跳网络技术有限公司 视频处理方法、装置、电子设备及存储介质

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021217398A1 (zh) * 2020-04-28 2021-11-04 深圳市大疆创新科技有限公司 图像的处理方法及装置、可移动平台及其控制终端、计算机可读存储介质
CN114401443A (zh) * 2022-01-24 2022-04-26 脸萌有限公司 特效视频处理方法、装置、电子设备及存储介质
CN114531553A (zh) * 2022-02-11 2022-05-24 北京字跳网络技术有限公司 生成特效视频的方法、装置、电子设备及存储介质
CN114567805A (zh) * 2022-02-24 2022-05-31 北京字跳网络技术有限公司 确定特效视频的方法、装置、电子设备及存储介质
CN115278107A (zh) * 2022-07-20 2022-11-01 北京字跳网络技术有限公司 视频处理方法、装置、电子设备及存储介质

Also Published As

Publication number Publication date
CN115278107A (zh) 2022-11-01

Similar Documents

Publication Publication Date Title
CN111857923B (zh) 特效展示方法、装置、电子设备及计算机可读介质
WO2021259301A1 (zh) 直播互动方法、装置、可读介质及电子设备
WO2021159792A1 (zh) 虚拟物品交互方法、装置、计算机设备及存储介质
WO2020107908A1 (zh) 多用户视频特效添加方法、装置、终端设备及存储介质
WO2024016924A1 (zh) 视频处理方法、装置、电子设备及存储介质
WO2023116653A1 (zh) 元素展示方法、装置、电子设备及存储介质
WO2023169305A1 (zh) 特效视频生成方法、装置、电子设备及存储介质
WO2023185647A1 (zh) 媒体内容的显示方法、装置、设备、存储介质和程序产品
WO2023202590A1 (zh) 页面切换方法、装置以及终端设备的交互方法
US11869195B2 (en) Target object controlling method, apparatus, electronic device, and storage medium
WO2023138504A1 (zh) 图像渲染方法、装置、电子设备及存储介质
WO2021104130A1 (zh) 在视频中显示对象的方法、装置、电子设备及计算机可读存储介质
WO2023078069A1 (zh) 一种直播互动方法、系统及相关装置
CN111467791A (zh) 目标对象的控制方法及装置、系统
WO2023140786A2 (zh) 特效视频处理方法、装置、电子设备及存储介质
US20230267664A1 (en) Animation processing method and apparatus, electronic device and storage medium
WO2024027819A1 (zh) 图像处理方法、装置、设备及存储介质
WO2023241427A1 (zh) 图像处理方法、装置、设备及存储介质
WO2023185393A1 (zh) 图像处理方法、装置、设备及存储介质
WO2023207989A1 (zh) 虚拟对象的控制方法、装置、设备及存储介质
WO2023116562A1 (zh) 图像展示方法、装置、电子设备及存储介质
WO2023179375A1 (zh) 布局对象的操控方法、装置、电子设备以及存储介质
CN111897437A (zh) 跨终端的交互方法、装置、电子设备以及存储介质
CN111857482A (zh) 一种互动方法、装置、设备和可读介质
CN108829329B (zh) 一种操作对象展示方法、装置和可读介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23841990

Country of ref document: EP

Kind code of ref document: A1