CN115049574A - Video processing method and device, electronic equipment and readable storage medium - Google Patents

Video processing method and device, electronic equipment and readable storage medium Download PDF

Info

Publication number
CN115049574A
CN115049574A CN202210874585.4A CN202210874585A CN115049574A CN 115049574 A CN115049574 A CN 115049574A CN 202210874585 A CN202210874585 A CN 202210874585A CN 115049574 A CN115049574 A CN 115049574A
Authority
CN
China
Prior art keywords
video
video frame
character model
input
frame
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210874585.4A
Other languages
Chinese (zh)
Inventor
胡鹏翔
刘文强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vivo Mobile Communication Co Ltd
Original Assignee
Vivo Mobile Communication Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vivo Mobile Communication Co Ltd filed Critical Vivo Mobile Communication Co Ltd
Priority to CN202210874585.4A priority Critical patent/CN115049574A/en
Publication of CN115049574A publication Critical patent/CN115049574A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/74Browsing; Visualisation therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Computer Hardware Design (AREA)
  • Computer Graphics (AREA)
  • Architecture (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The application discloses a video processing method, a video processing device, an electronic device and a readable storage medium, which belong to the technical field of image processing, and the method comprises the following steps: receiving a first input of a user to a first 3D character model and a first video, the first video comprising a first video frame and a second video frame; in response to the first input, fusing the first 3D character model with the first video frame to obtain a third video frame, and fusing the first 3D character model with the second video frame to obtain a fourth video frame; generating a second video; wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.

Description

Video processing method and device, electronic equipment and readable storage medium
Technical Field
The present application belongs to the field of image processing technologies, and in particular, to a video processing method and apparatus, an electronic device, and a readable storage medium.
Background
In the prior art, a 3D model generation technology and an ar (augmented reality) technology are applied as image processing technologies in the field of image processing. The 3D model generation technique is a technique of constructing a model having three-dimensional data from a virtual three-dimensional space by using three-dimensional creation software, describing a spatial relationship between objects mathematically by using a computer, and editing each part of the model. The AR technology is a new technology for integrating real world information and virtual world information seamlessly, and the technology aims to fit a virtual world in the real world on a screen and interact with the virtual world.
At present, the mode of generating the AR video based on the 3D model is complex, the operation difficulty is high, and the method is used for manufacturing the required AR video which is difficult to rapidly.
Disclosure of Invention
An embodiment of the present application provides a video processing method, an apparatus, an electronic device, and a readable storage medium, which can solve the problem that an AR video is generated based on a 3D model and is relatively complex.
In a first aspect, an embodiment of the present application provides a video processing method, where the method includes:
receiving a first input of a user to a first 3D character model and a first video, the first video comprising a first video frame and a second video frame;
in response to the first input, fusing the first 3D character model with the first video frame to obtain a third video frame, and fusing the first 3D character model with the second video frame to obtain a fourth video frame;
generating a second video;
wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.
In a second aspect, an embodiment of the present application provides a video processing apparatus, including:
a receiving module, configured to receive a first input of a first 3D character model and a first video from a user, where the first video includes a first video frame and a second video frame;
a fusion module, configured to fuse the first 3D character model and the first video frame in response to the first input to obtain a third video frame, and fuse the first 3D character model and the second video frame to obtain a fourth video frame;
the generating module is used for generating a second video;
wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.
In a third aspect, embodiments of the present application provide an electronic device, which includes a processor and a memory, where the memory stores a program or instructions executable on the processor, and the program or instructions, when executed by the processor, implement the steps of the method according to the first aspect.
In a fourth aspect, embodiments of the present application provide a readable storage medium, on which a program or instructions are stored, which when executed by a processor implement the steps of the method according to the first aspect.
In a fifth aspect, an embodiment of the present application provides a chip, where the chip includes a processor and a communication interface, where the communication interface is coupled to the processor, and the processor is configured to execute a program or instructions to implement the method according to the first aspect.
In a sixth aspect, embodiments of the present application provide a computer program product, which is stored in a storage medium and executed by at least one processor to implement the method according to the first aspect.
In the embodiment of the application, after receiving a first input of a user to a first 3D character model and a first video, an electronic device may, in response to the first input, fuse the first 3D character model with a first video frame of the first video to obtain a third video frame, fuse the first 3D character model with a second video frame of the first video to obtain a fourth video frame, and generate a second video based on the first video, where the second video includes the third video frame and the fourth video frame. Namely, the method can fuse the first 3D character model on different video frames of the pre-recorded first video to generate the second video, and the purpose of fusing the virtual 3D character model with the real video and generating the second video with the story line is achieved.
Drawings
Fig. 1 is a flowchart of a video processing method according to an embodiment of the present application;
fig. 2 is one of interface schematic diagrams of a video playing interface provided in an embodiment of the present application;
fig. 3 is a second schematic interface diagram of a video playing interface according to an embodiment of the present application;
fig. 4 is a third schematic interface diagram of a video playing interface according to the embodiment of the present application;
fig. 5 is a fourth schematic interface diagram of a video playing interface provided in the embodiment of the present application;
fig. 6 is a fifth schematic interface diagram of a video playing interface provided in the embodiment of the present application;
fig. 7 is a sixth schematic interface diagram of a video playing interface provided in the embodiment of the present application;
fig. 8 is a seventh schematic interface diagram of a video playing interface provided in the embodiment of the present application;
fig. 9 is an eighth schematic interface diagram of a video playing interface according to an embodiment of the present application;
fig. 10 is a ninth schematic interface diagram of a video playing interface provided in the embodiment of the present application;
fig. 11 is a tenth of an interface schematic diagram of a video playing interface provided in the embodiment of the present application;
fig. 12 is an eleventh schematic interface diagram of a video playing interface according to an embodiment of the present application;
fig. 13 is a twelfth schematic interface diagram of a video playing interface provided in the embodiment of the present application;
fig. 14 is a thirteen schematic interface diagrams of a video playing interface according to an embodiment of the present application;
fig. 15 is a fourteenth schematic interface diagram of a video playing interface according to an embodiment of the present application;
fig. 16 is a fifteen-level interface schematic diagram of a video playing interface provided in the embodiment of the present application;
fig. 17 is a sixteenth schematic interface diagram of a video playing interface according to an embodiment of the present application;
fig. 18 is a schematic structural diagram of a video processing apparatus according to an embodiment of the present application;
fig. 19 is a schematic structural diagram of an electronic device according to an embodiment of the present application;
fig. 20 is a schematic structural diagram of an electronic device according to another embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described clearly below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments that can be derived by one of ordinary skill in the art from the embodiments given herein are intended to be within the scope of the present disclosure.
The terms first, second and the like in the description and in the claims of the present application are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It will be appreciated that the data so used may be interchanged under appropriate circumstances such that embodiments of the application may be practiced in sequences other than those illustrated or described herein, and that the terms "first," "second," and the like are generally used herein in a generic sense and do not limit the number of terms, e.g., the first term can be one or more than one. In addition, "and/or" in the specification and claims means at least one of connected objects, a character "/" generally means that a preceding and succeeding related objects are in an "or" relationship.
In the video processing method provided by the embodiment of the present application, an execution main body of the video processing method may be the video processing apparatus provided by the embodiment of the present application, or an electronic device integrated with the video processing apparatus, where the video processing apparatus may be implemented in a hardware or software manner.
Through the video processing method provided by the embodiment of the application, the user can add the virtual character in the video and interact with the virtual character in the video, such as running together, kicking together and the like.
By the video processing method provided by the embodiment of the application, the user can also add virtual characters in the video, and use the video as a script to complete a story line of the script video to obtain the script video created by the user. The user can perform according to the script video.
The video processing method provided by the embodiment of the present application is described in detail below with reference to the accompanying drawings through specific embodiments and application scenarios thereof.
Fig. 1 shows a video processing method provided in an embodiment of the present application, and the method is applied to an electronic device. As shown in fig. 1, the video processing method may include the following steps 1100 to 1300, which are described in detail below.
At step 1100, a first input from a user to a first 3D character model and a first video is received.
The first 3D character model may be a character model that is created by a user editing a 3D character model provided by the electronic device to be edited. Referring to fig. 4, a user may edit a 3D character model to be edited provided by an electronic device, resulting in a "thief" character model as a first 3D character model 401.
The first 3D character model may have different actions. Referring to FIG. 4, a first 3D character model 401, a "thief" character model, may have different actions, e.g., first 3D character model 401 may have action 1, action 2, and action 3.
The first video may be a video recorded in advance by the user. Referring to fig. 3, a first video 301 is a video of a user's running. The first video frame and the second video frame are two different video frames in the first video.
In this embodiment, before receiving a first input of a first 3D character model and a first video by a user in step 1100, referring to fig. 2, in a case where the electronic device starts a camera application, the electronic device displays a shooting preview interface 201. Wherein, a photographing control, a beauty control, a story video control and a video recording control are displayed on the photographing preview interface 201. The electronic equipment can enter a story video mode by clicking the story video control by the user, and the user can select the first video and edit the 3D character model to be edited to obtain the first 3D character model under the condition that the electronic equipment enters the story video mode.
Specifically, under the condition that the electronic device enters a story video mode, a scene video control 202 and an editing model control 203 are provided on the shooting preview interface 201, wherein a first video can be selected through the scene video control 202, and a 3D character model to be edited can be edited through the editing model control 203 to obtain a first 3D character model. With the scene of a police thief in a movie as a storyline, when a user clicks the scene video control 202 shown in fig. 2, different videos recorded in advance by the user in the electronic device, such as video 1, video 2, and video 3 … …, video 6, can be loaded and displayed on the shooting preview interface 201, and referring to fig. 3, the user can select video 1, that is, a video that the user runs, as the first video 301.
Meanwhile, when the user clicks the edit model control 203 shown in fig. 2, a model edit interface is displayed, and referring to fig. 4, by clicking the "add" control, the 3D character model to be edited can be imported, and the 3D character model to be edited is named as "thief". Referring to fig. 4, in the case of importing a 3D character model to be edited, a user may adjust the 3D character model to be edited by dragging, and put out different escape actions and save them as different actions, such as action 1, action 2, and action 3, respectively. Clicking on the "Save" control may then generate a "thief" character model with a different action, i.e., first 3D character model 401. That is, based on the story video control in the camera application, the first video 301 may be selected and the first 3D character model 401 may be generated. Referring to fig. 5, a happy citizen character model can be generated as the first 3D character model 501 in the same manner.
The first input may be: the click input of the user on the first video, or the voice instruction input by the user, or the specific gesture input by the user may be specifically determined according to actual use requirements, which is not limited in the embodiment of the present application.
The specific gesture in the embodiment of the application can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure identification gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input in the embodiment of the application can be click input, double-click input, click input of any number of times and the like, and can also be long-time press input or short-time press input.
Referring to fig. 6, in the case of playing the first video 301, i.e., the user running video, the user may directly drag the playing progress bar 601 of the first video 301 to the first video frame that needs to be inserted into the first 3D character model, for example, the first video frame 602 may be the video frame displayed at the playing time 15:05 of the first video 301. Here, the user clicks on the insert personality control 603 shown in fig. 6, and a personality selection window 604 is displayed on the first video frame 602, where the user may click on the thief 3D personality model, first 3D personality model 401.
After receiving a first input of the first 3D character model and the first video by the user in performing step S1100, the following steps are entered:
step 1200, in response to the first input, fusing the first 3D character model with the first video frame to obtain a third video frame, and fusing the first 3D character model with the second video frame to obtain a fourth video frame.
In one example, the step S1200 of fusing the first 3D character model and the first video frame to obtain a third video frame in response to the first input may further include the steps S1210 to S1220 of:
step S1210, in response to the first input, displaying a first selection window.
The first selection window includes at least one motion control for indicating a motion of the character model. Illustratively, when the user clicks on the thief 3D character model shown in fig. 6, i.e., the first 3D character model 401, a first selection window 605 is displayed on the first video frame 602, the first selection window being an action setting window. The first selection window 605 includes an action control 1, an action control 2, and an action control 3 thereon, where the action control 1 is used to indicate an action 1 of the first 3D character model 401, the action control 2 is used to indicate an action 2 of the first 3D character model 401, and the action control 3 is used to indicate an action 3 of the first 3D character model 401.
Step S1220, in response to a second input to a target action control of the at least one action control, fusing the first 3D character model with the first video frame to obtain a third video frame.
In step S1220, the electronic device may receive a second input of a target motion control of the at least one motion control, and fuse the first 3D character model and the first video frame to obtain a third video frame in response to the second input.
The second input may be: the click input of the user to the target action control in the at least one action control, or the voice instruction input by the user, or the specific gesture input by the user may be specifically determined according to the actual use requirement, which is not limited in the embodiment of the present application.
The specific gesture in the embodiment of the application can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure identification gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input in the embodiment of the application can be click input, double-click input, click input of any number of times and the like, and can also be long-time press input or short-time press input.
Referring to fig. 6, when the user clicks the action control 1 on the first selection window 605, the electronic device may fuse the first 3D character model 401 having the action 1 with the first video frame to obtain a third video frame.
In one example, the step S1200 of fusing the first 3D character model and the second video frame to obtain a fourth video frame in response to the first input may further include the following steps S1230 to S1240:
step S1230, in response to the first input, displaying a second selection window.
The second selection window includes at least one video frame control. Referring to fig. 6, in the case of playing a first video, a second selection window 606 is displayed to the left of the play progress bar 601 of the first video, and the second selection window 606 is used to select a second video frame in the first video. The user clicks on the second selection window 606 shown in fig. 6 to display at least one video frame control (not shown).
Step S1240, in response to a third input to a target video frame control of the at least one video frame control, fusing the first 3D character model with the second video frame to obtain a fourth video frame.
The target video frame control is to indicate the second video frame.
In step S1240, the electronic device may receive a second input of the target video frame control of the at least one video frame control, and in response to the second input, fuse the first 3D character model with the second video frame to obtain a fourth video frame.
The second input may be: the click input of the user to the target video frame control in the at least one video frame control, or the voice instruction input by the user, or the specific gesture input by the user may be specifically determined according to the actual use requirement, which is not limited in the embodiment of the present application.
The specific gesture in the embodiment of the application can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure identification gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input in the embodiment of the application can be single click input, double click input, click input of any number of times and the like, and can also be long-press input or short-press input.
In a specific embodiment, the above fusing the first 3D character model with the second video frame to obtain the fourth video frame may further include: and responding to the editing input of the first 3D character model to obtain a second 3D character model, and fusing the second 3D character model and the second video frame to obtain a fourth video frame.
Referring to fig. 6, when the user clicks the target video frame control, the electronic device may use the video frame in the first video indicated by the target video frame control as the second video frame, and referring to fig. 7, the second video frame 701 may be the video frame displayed at the playing time 15:15 of the first video 301. Here, the user clicks the edit character control 601 shown in fig. 6, edits the motion of the first 3D character model 401, which is the thief 3D character model shown in fig. 6, and for example, obtains the thief 3D character model having the motion 4 as the second 3D character model 702 by editing, and obtains a fourth video frame by fusing the second video frame and the second 3D character model, referring to fig. 7.
It will be appreciated that the 3D character model 501 for the happy citizen may be fused to the video frames following the second video frame, and referring to fig. 8, the video frame 801 following the second video frame 701 may be the video frame displayed at the play time 35:15 of the first video 301, which is fused with the 3D character model 501 for the happy citizen in order to create a scenario where the user (police) and the happy citizen catch the thief together.
Step 1300, generating a second video.
Wherein the second video has an association relation with the first video, that is, the second video is generated based on the first video.
The second video includes the third video frame and the fourth video frame.
In this embodiment, the generating the second video in step 1300 may further include: interpolating the frame according to the actions of the first 3D character model in the third video frame and the fourth video frame to generate at least one fifth video frame; and generating a second video according to the third video frame, the fourth video frame and the at least one fifth video frame.
Illustratively, according to the thief 3D character model with action 1 in the third video frame and the thief 3D character model with action 4 in the fourth video frame, at least one fifth video frame is generated by interpolating between the third video frame and the fourth video frame, and the second video shown in fig. 9 is generated according to the third video frame, the fourth video frame and the at least one fifth video frame, so as to ensure the consistency of the thief character model when the second video is played.
According to the embodiment, after receiving a first input of a user to a first 3D character model and a first video, an electronic device, in response to the first input, fuses the first 3D character model and a first video frame of the first video to obtain a third video frame, fuses the first 3D character model and a second video frame of the first video to obtain a fourth video frame, and further generates a second video based on the first video, wherein the second video includes the third video frame and the fourth video frame. Namely, the method can fuse the first 3D character model on different video frames of the pre-recorded first video to generate the second video, and the purpose of fusing the virtual 3D character model with the real video and generating the second video with the story line is achieved.
In one embodiment, before performing step 1300 to generate the second video, the video processing method of the embodiment of the present disclosure further includes steps 2100 to 2300 as follows:
at step 2100, a fourth input is received for a third 3D character model.
The third 3D character model may be a character model obtained by a user editing another 3D character to be edited provided by the electronic device. Referring to fig. 12, a user may edit another 3D character model to be edited provided by the electronic device, resulting in a "sports avatar" character model as a third 3D character model 1201.
The third 3D character model may have a different motion. Referring to fig. 12, the third 3D character model 1201, i.e., the "moving split" character model, may have different actions, e.g., the third 3D character model 1201 may have a ball-taking action, a targeting action.
In this embodiment, before step 2100 is executed to receive a fourth input to the third 3D character model, referring to fig. 10, in a case where the electronic device starts a camera application, the electronic device displays a shooting preview interface 1001. A photographing control, a beauty control, a story video control, and a video recording control are displayed on the photographing preview interface 1001. Here, the electronic device may enter a story video mode by clicking the story video control by the user, and the third 3D character model 1201 may be obtained by editing the 3D character model to be edited by the user by selecting the first video when the electronic device enters the story video mode.
Specifically, when the electronic device enters a story video mode, a scene video control 1002 and an editing model control 1003 are provided on the shooting preview interface 1001, wherein the scene video can be selected through the scene video control 1002, and the 3D character model to be edited can be edited through the editing model control 1003 to obtain a third 3D character model. With the basketball highlight as a story line, when the user clicks the scene video control 1002 shown in fig. 10, different pre-downloaded scene videos in the electronic device, such as video 1, video 2, and video 3 … …, for example, video 6, may be loaded and displayed on the shooting preview interface 1001, and referring to fig. 11, the user may select video 2, that is, basketball court video, as the scene video.
Meanwhile, in the case where the user clicks the edit model control 1003 shown in fig. 10, the model edit interface is displayed, and referring to fig. 12, by clicking the "add" control, the 3D character model to be edited is imported, and the 3D character model to be edited is named "sports avatar". Referring to fig. 12, when a 3D character model to be edited is imported, a user may adjust the 3D character model to be edited by dragging, and put out different shooting motions and save the different motions, such as a ball-taking motion and a aiming motion, respectively. Clicking on the "save" control generates a "sports split" character model, i.e. a third 3D character model 1201 with a different shooting action. That is, based on the story video control in the camera application, the third 3D character model 1201 may be generated.
The fourth input may be: the click input of the user on the third 3D character model, or the voice instruction input by the user, or the specific gesture input by the user may be specifically determined according to the actual use requirement, which is not limited in this embodiment of the application.
The specific gesture in the embodiment of the application can be any one of a single-click gesture, a sliding gesture, a dragging gesture, a pressure identification gesture, a long-press gesture, an area change gesture, a double-press gesture and a double-click gesture; the click input in the embodiment of the application can be single click input, double click input, click input of any number of times and the like, and can also be long-press input or short-press input.
Referring to fig. 13, in a case of playing a scene video, that is, a basketball court video, a user may directly drag a playing progress bar 1301 of the basketball court video to a video frame that needs to be inserted into the third 3D character model, for example, a video frame 1302 may be a video frame displayed at a playing time 10:05 of the basketball court video. Here, the user clicks the insert character control 1303 shown in fig. 13, and a character selection window 1304 is displayed on the video frame 1302, where the user can click the third 3D character model 1201 which is the sports avatar character model, and select the ball-taking action control in the first selection window 1305, so that the third 3D character model 1201 having the ball-taking action can be fused with the video frame 1302.
Referring to fig. 14, for example, video frame 1401 may be a video frame displayed at playing time 20:05 of basketball court video. Here, the third 3D character model 1201 having the targeting action may be fused with the video frame 1401 based on the same manner.
Step 2200, in response to the fourth input, obtaining at least two frames of video frames, where actions of the third 3D character model in the at least two frames of video frames are different.
The at least two video frames include a sixth video frame including at least two third 3D character models.
Continuing with the above example, one of the resulting at least two frames of video frames may be: the resulting video frames are fused with third 3D character model 1201 having a ball-holding motion and video frames 1302. Another video frame of the resulting at least two video frames may be: the generated video frame is fused with the third 3D character model 1201 with the targeting action and the video frame 1401. Here, in order to facilitate the user to simulate the action of the scenario, it is necessary to present a moving separate 3D character model of the previous frame or several previous frames in the current video frame, and with reference to fig. 15, realize the effect of a third 3D character model in which a plurality of actions appear in one video frame, and further generate a video shown in fig. 16, which is an AR scenario in which the story line is at the moment of shooting a highlight. The video can be a virtual video template, and the user can shoot the first video based on the virtual video template, namely the user shoots the first video by imitating shooting motions in the virtual video template.
Step 2300, adjusting the action of the target object in the video frame of the first video according to the at least two video frames.
Continuing with the above example, the motion of the target object in the video frame of the corresponding first video may be adjusted according to the motion of the third 3D character model in at least two video frames in the virtual video template, so as to generate a video that is the same as the scenario and motion of the virtual video template, i.e. the video shown in fig. 17.
According to the embodiment of the disclosure, the user can imitate the action of the third 3D character model in the virtual video template, complete the shooting story line and record the real video. And after the user records the real video, the real video can be made into the video with the body separating effect based on the virtual video template.
Corresponding to the above embodiments, referring to fig. 18, an embodiment of the present application further provides a video processing apparatus 2100, where the apparatus 2100 includes a receiving module 2110, a fusing module 2120, and a generating module 2130.
A receiving module 2110 is configured to receive a first input from a user to a first 3D character model and a first video, the first video including a first video frame and a second video frame.
A fusion module 2120, configured to, in response to the first input, fuse the first 3D character model with the first video frame to obtain a third video frame, and fuse the first 3D character model with the second video frame to obtain a fourth video frame.
A generating module 2130 for generating a second video.
Wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.
In one embodiment, the fusion module 2120 is specifically configured to display, in response to the first input, a first selection window, where the first selection window includes at least one action control, and the action control is used to indicate an action of the character model; and in response to a second input to a target action control in the at least one action control, fusing the first 3D character model and the first video frame to obtain a third video frame.
In one embodiment, the fusion module 2120 is specifically configured to, in response to the first input, display a second selection window, where the second selection window includes at least one video frame control; and in response to a third input to a target video frame control in the at least one video frame control, fusing the first 3D character model with the second video frame to obtain a fourth video frame, wherein the target video frame control is used for indicating the second video frame.
In an embodiment, the fusion module 2120 is specifically configured to obtain a second 3D character model in response to an editing input to the first 3D character model, and fuse the second 3D character model with the second video frame to obtain a fourth video frame.
In one embodiment, the apparatus 2100 further comprises an adjustment module (not shown in the figures).
The receiving module 2110 is further configured to receive a fourth input to the third 3D character model.
The generating module 2130 is further configured to, in response to the fourth input, obtain at least two video frames, where actions of a third 3D character model in the at least two video frames are different.
And the adjusting module is used for adjusting the action of the target object in the video frame of the first video according to the at least two frames of video frames.
In one embodiment, the at least two video frames include a sixth video frame including at least two third 3D character models.
In one embodiment, the generating module 2130 is specifically configured to insert a frame to generate at least one fifth video frame according to the actions of the first 3D character model in the third video frame and the fourth video frame; and generating a second video according to the third video frame, the fourth video frame and the at least one fifth video frame.
In an embodiment of the application, after receiving a first input of a user to a first 3D character model and a first video, an electronic device, in response to the first input, fuses the first 3D character model and a first video frame of the first video to obtain a third video frame, fuses the first 3D character model and a second video frame of the first video to obtain a fourth video frame, and further generates a second video based on the first video, where the second video includes the third video frame and the fourth video frame. Namely, the method can fuse the first 3D character model on different video frames of the pre-recorded first video to generate the second video, and the purpose of fusing the virtual 3D character model with the real video and generating the second video with the story line is achieved.
The video processing apparatus in the embodiment of the present application may be an electronic device, and may also be a component in the electronic device, such as an integrated circuit or a chip. The electronic device may be a terminal, or may be a device other than a terminal. The electronic Device may be, for example, a Mobile phone, a tablet computer, a notebook computer, a palm computer, a vehicle-mounted electronic Device, a Mobile Internet Device (MID), an Augmented Reality (AR)/Virtual Reality (VR) Device, a robot, a wearable Device, an ultra-Mobile personal computer (UMPC), a netbook or a Personal Digital Assistant (PDA), and the like, and may also be a server, a Network Attached Storage (Network Attached Storage, NAS), a personal computer (personal computer, PC), a television (television, TV), an assistant, or a self-service machine, and the embodiments of the present application are not limited in particular.
The video processing apparatus in the embodiment of the present application may be an apparatus having an operating system. The operating system may be an Android (Android) operating system, an ios operating system, or other possible operating systems, and embodiments of the present application are not limited specifically.
The video processing apparatus provided in the embodiment of the present application can implement each process implemented in the method embodiment of fig. 1, and is not described here again to avoid repetition.
Optionally, as shown in fig. 19, an electronic device 2200 is further provided in this embodiment of the present application, and includes a processor 2201 and a memory 2202, where the memory 2202 stores a program or an instruction that can be executed on the processor 2201, and when the program or the instruction is executed by the processor 2201, the steps of the above-described embodiment of the video processing method are implemented, and the same technical effects can be achieved, and are not repeated here to avoid repetition.
It should be noted that the electronic device in the embodiment of the present application includes the mobile electronic device and the non-mobile electronic device described above.
Fig. 20 is a schematic hardware structure diagram of an electronic device implementing an embodiment of the present application.
The electronic device 2300 includes, but is not limited to: radio frequency unit 2301, network module 2302, audio output unit 2303, input unit 2304, sensor 2305, display unit 2306, user input unit 2307, interface unit 2308, memory 2309, processor 2310 and the like.
Those skilled in the art will appreciate that the electronic device 2300 may further include a power supply (e.g., a battery) to power the various components, which may be logically coupled to the processor 2310 via a power management system to manage charging, discharging, and power consumption management functions via the power management system. The electronic device structure shown in fig. 20 does not constitute a limitation of the electronic device, and the electronic device may include more or less components than those shown, or combine some components, or arrange different components, and thus, the description thereof is omitted.
Wherein the user input unit 2307 is configured to receive a first input of a first 3D character model and a first video, the first video including a first video frame and a second video frame.
A processor 2310 is configured to, in response to the first input, fuse the first 3D character model with the first video frame to obtain a third video frame, and fuse the first 3D character model with the second video frame to obtain a fourth video frame.
Processor 2310 is further configured to generate a second video.
Wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.
According to the embodiment, after receiving a first input of a user to a first 3D character model and a first video, an electronic device, in response to the first input, fuses the first 3D character model and a first video frame of the first video to obtain a third video frame, fuses the first 3D character model and a second video frame of the first video to obtain a fourth video frame, and further generates a second video based on the first video, wherein the second video includes the third video frame and the fourth video frame. Namely, the method can fuse the first 3D character model on different video frames of the pre-recorded first video to generate the second video, and the purpose of fusing the virtual 3D character model with the real video and generating the second video with the story line is achieved.
In one embodiment, the display unit 2307 is configured to display a first selection window in response to the first input, the first selection window including at least one action control for indicating an action of the character model.
The processor 2310 is further configured to fuse the first 3D character model with the first video frame to obtain a third video frame in response to a second input to a target motion control of the at least one motion control.
In one embodiment, the display unit 2307 is configured to display a second selection window in response to the first input, the second selection window including at least one video frame control.
The processor 2310 is further configured to fuse the first 3D character model with the second video frame to obtain a fourth video frame in response to a third input to a target video frame control of the at least one video frame control, where the target video frame control is used to indicate the second video frame.
In one embodiment, user input unit 2307 is used to receive a fourth input to the third 3D character model.
A processor 2310 configured to obtain at least two video frames in response to the fourth input, where actions of a third 3D character model in the at least two video frames are different.
A processor 2310 configured to adjust an action of a target object in the video frame of the first video according to the at least two frames of video frames.
In one embodiment, the at least two video frames include a sixth video frame including at least two third 3D character models.
In one embodiment, the processor 2310 is further configured to interpolate to generate at least one fifth video frame according to the motion of the first 3D character model in the third video frame and the fourth video frame; and generating a second video according to the third video frame, the fourth video frame and the at least one fifth video frame.
It should be understood that, in the embodiment of the present application, the input Unit 2304 may include a Graphics Processing Unit (GPU) 23041 and a microphone 23042, and the Graphics processor 23041 processes image data of still pictures or videos obtained by an image capturing device (such as a camera) in a video capturing mode or an image capturing mode. The display unit 2306 may include a display panel 23061, and the display panel 23061 may be configured in the form of a liquid crystal display, an organic light emitting diode, or the like. The user input unit 2307 includes a touch panel 23071 and at least one of other input devices 23072. The touch panel 23071 is also referred to as a touch screen. The touch panel 23071 may include two parts of a touch detection device and a touch controller. Other input devices 23072 may include, but are not limited to, a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, and a joystick, which are not described in detail herein.
The memory 2309 may be used to store software programs as well as various data. The memory 2309 may mainly include a first storage area storing a program or instructions and a second storage area storing data, wherein the first storage area may store an operating system, an application program or instructions required for at least one function (such as a sound playing function, an image playing function, and the like), and the like. Further, the memory 2309 can include volatile memory or nonvolatile memory, or the memory 2309 can include both volatile and nonvolatile memory. The non-volatile Memory may be a Read-Only Memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an Electrically Erasable PROM (EEPROM), or a flash Memory. The volatile Memory may be a Random Access Memory (RAM), a Static Random Access Memory (Static RAM, SRAM), a Dynamic Random Access Memory (Dynamic RAM, DRAM), a Synchronous Dynamic Random Access Memory (Synchronous DRAM, SDRAM), a Double Data Rate Synchronous Dynamic Random Access Memory (Double Data Rate SDRAM, ddr SDRAM), an Enhanced Synchronous SDRAM (ESDRAM), a Synchronous Link DRAM (SLDRAM), and a Direct Memory bus RAM (DRRAM). The memory 2309 in embodiments of the present application includes, but is not limited to, these and any other suitable types of memory.
The processor 2310 may include one or more processing units; optionally, the processor 2310 integrates an application processor, which primarily handles operations involving the operating system, user interface, and applications, and a modem processor, which primarily handles wireless communication signals, such as a baseband processor. It is to be appreciated that the modem processor can be separate from and integrated with the processor 2310.
The embodiments of the present application further provide a readable storage medium, where a program or an instruction is stored on the readable storage medium, and when the program or the instruction is executed by a processor, the program or the instruction implements each process of the video processing method embodiment, and can achieve the same technical effect, and in order to avoid repetition, details are not repeated here.
The processor is the processor in the electronic device described in the above embodiment. The readable storage medium includes a computer readable storage medium, such as a computer read only memory ROM, a random access memory RAM, a magnetic or optical disk, and the like.
The embodiment of the present application further provides a chip, where the chip includes a processor and a communication interface, the communication interface is coupled to the processor, and the processor is configured to run a program or an instruction to implement each process of the above video processing method embodiment, and can achieve the same technical effect, and the details are not repeated here to avoid repetition.
It should be understood that the chips mentioned in the embodiments of the present application may also be referred to as system-on-chip, system-on-chip or system-on-chip, etc.
Embodiments of the present application provide a computer program product, where the program product is stored in a storage medium, and the program product is executed by at least one processor to implement the processes of the foregoing video processing method embodiments, and can achieve the same technical effects, and in order to avoid repetition, details are not repeated here.
It should be noted that, in this document, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises the element. Further, it should be noted that the scope of the methods and apparatus of the embodiments of the present application is not limited to performing the functions in the order illustrated or discussed, but may include performing the functions in a substantially simultaneous manner or in a reverse order based on the functions involved, e.g., the methods described may be performed in an order different than that described, and various steps may be added, omitted, or combined. In addition, features described with reference to certain examples may be combined in other examples.
Through the above description of the embodiments, those skilled in the art will clearly understand that the method of the above embodiments can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware, but in many cases, the former is a better implementation manner. Based on such understanding, the technical solutions of the present application may be embodied in the form of a computer software product, which is stored in a storage medium (such as ROM/RAM, magnetic disk, optical disk) and includes instructions for enabling a terminal (such as a mobile phone, a computer, a server, or a network device) to execute the method according to the embodiments of the present application.
While the present embodiments have been described with reference to the accompanying drawings, it is to be understood that the invention is not limited to the precise embodiments described above, which are meant to be illustrative and not restrictive, and that various changes may be made therein by those skilled in the art without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (10)

1. A method of video processing, the method comprising:
receiving a first input of a user to a first 3D character model and a first video, the first video comprising a first video frame and a second video frame;
in response to the first input, fusing the first 3D character model with the first video frame to obtain a third video frame, and fusing the first 3D character model with the second video frame to obtain a fourth video frame;
generating a second video;
wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.
2. The method of claim 1, wherein said fusing the first 3D character model with the first video frame to obtain a third video frame in response to the first input comprises:
in response to the first input, displaying a first selection window, the first selection window including at least one action control for indicating an action of a character model;
and in response to a second input to a target action control in the at least one action control, fusing the first 3D character model and the first video frame to obtain a third video frame.
3. The method of claim 1, wherein said fusing the first 3D character model with the second video frame to obtain a fourth video frame in response to the first input comprises:
in response to the first input, displaying a second selection window, the second selection window comprising at least one video frame control;
and in response to a third input to a target video frame control in the at least one video frame control, fusing the first 3D character model with the second video frame to obtain a fourth video frame, wherein the target video frame control is used for indicating the second video frame.
4. The method of claim 1, wherein said fusing the first 3D character model with the second video frame to obtain a fourth video frame comprises:
and responding to the editing input of the first 3D character model to obtain a second 3D character model, and fusing the second 3D character model and the second video frame to obtain a fourth video frame.
5. The method of claim 1, prior to said generating the second video, further comprising:
receiving a fourth input to a third 3D character model;
responding to the fourth input, obtaining at least two video frames, wherein actions of a third 3D character model in the at least two video frames are different;
and adjusting the action of the target object in the video frame of the first video according to the at least two video frames.
6. The method of claim 5, wherein the at least two video frames include a sixth video frame including at least two third 3D character models.
7. The method of any of claims 1-6, wherein the generating the second video comprises:
interpolating a frame to generate at least one fifth video frame according to the actions of the first 3D character model in the third video frame and the fourth video frame;
and generating a second video according to the third video frame, the fourth video frame and the at least one fifth video frame.
8. A video processing apparatus, characterized in that the apparatus comprises:
a receiving module, configured to receive a first input of a first 3D character model and a first video from a user, where the first video includes a first video frame and a second video frame;
a fusion module, configured to fuse the first 3D character model and the first video frame in response to the first input to obtain a third video frame, and fuse the first 3D character model and the second video frame to obtain a fourth video frame;
the generating module is used for generating a second video;
wherein the second video has an association relation with the first video, and the second video comprises the third video frame and the fourth video frame.
9. An electronic device comprising a processor and a memory, the memory storing a program or instructions executable on the processor, the program or instructions when executed by the processor implementing the steps of the video processing method according to any of claims 1-7.
10. A computer-readable storage medium, on which a program or instructions are stored, which, when executed by a processor, carry out the steps of the video processing method according to any one of claims 1 to 7.
CN202210874585.4A 2022-07-21 2022-07-21 Video processing method and device, electronic equipment and readable storage medium Pending CN115049574A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210874585.4A CN115049574A (en) 2022-07-21 2022-07-21 Video processing method and device, electronic equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210874585.4A CN115049574A (en) 2022-07-21 2022-07-21 Video processing method and device, electronic equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN115049574A true CN115049574A (en) 2022-09-13

Family

ID=83166488

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210874585.4A Pending CN115049574A (en) 2022-07-21 2022-07-21 Video processing method and device, electronic equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN115049574A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115761114A (en) * 2022-10-28 2023-03-07 如你所视(北京)科技有限公司 Video generation method and device and computer readable storage medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115761114A (en) * 2022-10-28 2023-03-07 如你所视(北京)科技有限公司 Video generation method and device and computer readable storage medium
CN115761114B (en) * 2022-10-28 2024-04-30 如你所视(北京)科技有限公司 Video generation method, device and computer readable storage medium

Similar Documents

Publication Publication Date Title
CN108566519B (en) Video production method, device, terminal and storage medium
CN109120866B (en) Dynamic expression generation method and device, computer readable storage medium and computer equipment
KR101831802B1 (en) Method and apparatus for producing a virtual reality content for at least one sequence
CN114387400A (en) Three-dimensional scene display method, display device, electronic equipment and server
US20140282000A1 (en) Animated character conversation generator
CN113918522A (en) File generation method and device and electronic equipment
CN114222076B (en) Face changing video generation method, device, equipment and storage medium
CN115049574A (en) Video processing method and device, electronic equipment and readable storage medium
CN114466232A (en) Video processing method, video processing device, electronic equipment and medium
CN112309449A (en) Audio recording method and device
CN114357554A (en) Model rendering method, rendering device, terminal, server and storage medium
CN114025237B (en) Video generation method and device and electronic equipment
CN112367487B (en) Video recording method and electronic equipment
KR101806922B1 (en) Method and apparatus for producing a virtual reality content
CN114125297A (en) Video shooting method and device, electronic equipment and storage medium
CN113873319A (en) Video processing method and device, electronic equipment and storage medium
CN112261483A (en) Video output method and device
TWM560053U (en) Editing device for integrating augmented reality online
CN114173178A (en) Video playing method, video playing device, electronic equipment and readable storage medium
CN113873162A (en) Shooting method, shooting device, electronic equipment and readable storage medium
CN115278378A (en) Information display method, information display device, electronic apparatus, and storage medium
CN117395462A (en) Method and device for generating media content, electronic equipment and readable storage medium
CN114527906A (en) Display method, display device, electronic equipment and readable storage medium
CN115842953A (en) Shooting method and device thereof
CN114785949A (en) Video object processing method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination