CN114401451A - Video editing method and device, electronic equipment and readable storage medium - Google Patents

Video editing method and device, electronic equipment and readable storage medium Download PDF

Info

Publication number
CN114401451A
CN114401451A CN202111629147.3A CN202111629147A CN114401451A CN 114401451 A CN114401451 A CN 114401451A CN 202111629147 A CN202111629147 A CN 202111629147A CN 114401451 A CN114401451 A CN 114401451A
Authority
CN
China
Prior art keywords
original
video
dimensional
video data
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111629147.3A
Other languages
Chinese (zh)
Inventor
周礼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
You Peninsula Beijing Information Technology Co ltd
Original Assignee
You Peninsula Beijing Information Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by You Peninsula Beijing Information Technology Co ltd filed Critical You Peninsula Beijing Information Technology Co ltd
Priority to CN202111629147.3A priority Critical patent/CN114401451A/en
Publication of CN114401451A publication Critical patent/CN114401451A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The embodiment of the application provides a video editing method, a video editing device, electronic equipment and a readable storage medium, wherein the method comprises the following steps: acquiring original video data and original three-dimensional scene information, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data; reconstructing a three-dimensional video scene according to the original video data and the original three-dimensional scene information; and receiving the editing operation of the user for the three-dimensional video scene, and responding to the editing operation to generate target video data. The method can solve the problems of poor universality and inconvenience of the conventional video editing process, so that a user can conveniently edit the video in a three-dimensional scene.

Description

Video editing method and device, electronic equipment and readable storage medium
Technical Field
The embodiment of the disclosure relates to the technical field of videos, in particular to a video editing method, a video editing device, electronic equipment and a computer-readable storage medium.
Background
In recent years, with the rise of short videos (Instant Music videos), in order to meet the requirements of people for making short videos, various Video acquisition tools and Video editing tools are developed, and based on the tools, users can create rich and diverse short Video contents.
The existing video editing method generally performs editing processing on a two-dimensional video scene, and the method generally includes acquiring two-dimensional video data based on a camera, loading the two-dimensional video data based on a video editing tool, and implementing video editing processing by adding virtual elements to the two-dimensional video data.
Therefore, the video editing method in the prior art is only directed at a two-dimensional video scene, so that the method is generally not applicable when a user needs to add a virtual three-dimensional element in video data, or even if the virtual three-dimensional element can be added, the method is inconvenient and has large limitations.
Disclosure of Invention
An object of the present disclosure is to provide an image processing method to solve the problems of poor universality and inconvenience in the conventional video editing process.
In a first aspect of the present disclosure, a video editing method is provided, where the method includes:
acquiring original video data and original three-dimensional scene information, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data;
reconstructing a three-dimensional video scene according to the original video data and the original three-dimensional scene information;
and receiving the editing operation of the user for the three-dimensional video scene, and responding to the editing operation to generate target video data.
In a second aspect of the present disclosure, there is also provided a video capturing method, including:
receiving input operation of a user;
and responding to the input operation, acquiring original video data and original three-dimensional scene information based on a preset augmented reality technology, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data.
In a third aspect of the present disclosure, there is also provided a video editing apparatus, the apparatus including:
the system comprises an original data acquisition module, a data acquisition module and a data acquisition module, wherein the original data acquisition module is used for acquiring original video data and original three-dimensional scene information, and the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data;
the three-dimensional video scene reconstruction module is used for reconstructing a three-dimensional video scene according to the original video data and the original three-dimensional scene information;
and the target video data generation module is used for receiving the editing operation of the user on the three-dimensional video scene and generating target video data in response to the editing operation.
In a fourth aspect of the present disclosure, there is also provided a video capturing apparatus, the apparatus including:
the receiving module is used for receiving input operation of a user;
and the acquisition module is used for responding to the input operation and acquiring original video data and original three-dimensional scene information based on a preset augmented reality technology, wherein the original three-dimensional scene information is used for representing a real scene in the process of acquiring the original video data.
In a fifth aspect of the present disclosure, there is also provided an electronic device, including:
a memory for storing executable instructions;
a processor configured to execute the electronic device according to the control of the instruction to perform the method of the first aspect or the second aspect of the disclosure.
In a sixth aspect of the present disclosure, there is also provided a computer-readable storage medium storing a computer program readable and executable by a computer, the computer program being configured to perform the method according to the first or second aspect of the present disclosure when the computer program is read and executed by the computer.
One advantageous effect of the embodiments of the present disclosure is that, according to the embodiments of the present disclosure, a video editing apparatus obtains original video data and original three-dimensional scene information representing a real scene in a process of acquiring the original video data; reconstructing a three-dimensional video scene based on the original video data and the original three-dimensional scene information; then, the device can provide the three-dimensional video scene which can conveniently and directly carry out editing operation on the video data in the three-dimensional scene for the user, and based on the three-dimensional video scene, the device can generate the target video data by receiving the editing operation of the user and responding to the editing operation.
Other features of the present description and advantages thereof will become apparent from the following detailed description of exemplary embodiments thereof, which proceeds with reference to the accompanying drawings.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the specification and together with the description, serve to explain the principles of the specification.
Fig. 1 is a schematic flowchart of a video editing method according to an embodiment of the present disclosure.
Fig. 2 is a schematic diagram of triggering data acquisition provided by an embodiment of the present disclosure.
Fig. 3 is a schematic flowchart of a three-dimensional scene reconstruction process provided in an embodiment of the present disclosure.
Fig. 4 is a schematic flowchart of a video capture method according to an embodiment of the present disclosure.
Fig. 5 is a schematic block diagram of a video editing apparatus according to an embodiment of the present disclosure.
Fig. 6 is a schematic block diagram of a video capture device according to an embodiment of the present disclosure.
Fig. 7 is a schematic diagram of a hardware structure of an electronic device according to an embodiment of the present disclosure.
Detailed Description
Various exemplary embodiments of the present disclosure will now be described in detail with reference to the accompanying drawings. It should be noted that: the relative arrangement of the components and steps, the numerical expressions and numerical values set forth in these embodiments do not limit the scope of the present invention unless specifically stated otherwise.
The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the invention, its application, or uses.
Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail but are intended to be part of the specification where appropriate.
In all examples shown and discussed herein, any particular value should be construed as merely illustrative, and not limiting. Thus, other examples of the exemplary embodiments may have different values.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, further discussion thereof is not required in subsequent figures.
< method embodiment I >
In recent years, with the continuous development of the augmented reality technology, the traditional method for editing video based on two-dimensional video scenes is increasingly insufficient in universality and convenience. In order to solve the problems, in the related art, a video editing method for adding a virtual three-dimensional element in real time based on an augmented reality technology is provided for recording video data containing the virtual three-dimensional element in the process of acquiring video data, but the method needs to perform video editing processing in real time, and has relatively large limitation.
To solve the above problems, an embodiment of the present disclosure provides a video editing method with strong universality and capable of conveniently performing video editing processing in an offline scene, please refer to fig. 1, which is a flowchart of the video editing method provided by the embodiment of the present disclosure. For convenience of description, in the embodiments of the present disclosure, if there is no specific description, a video editing apparatus for implementing the method is illustrated as a mobile phone.
As shown in fig. 1, the method of the present embodiment may include the following steps S1100-S1300, which are described in detail below.
Step S1100, acquiring original video data and original three-dimensional scene information, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data.
Raw video data is video data captured using an image capture device, such as a camera.
The original three-dimensional scene information may be each frame of camera attitude data corresponding to each video frame of the original video data and three-dimensional scene data representing a real scene, wherein the three-dimensional scene data may be at least one of scene model data, object occlusion information, scene recognition plane information, scene anchor point information, and the like; the scene model data may be scene mesh data; the object occlusion information may be information representing an occlusion relationship between objects in a real scene; the scene recognition plane information may be information representing the position and size of a plane in a real scene; the scene anchor information may be information representing key points of the real scene, and the anchor information may be generally a geometric center of the real scene.
Specifically, aiming at the problems that the existing video editing methods are poor in universality and incapable of conveniently adding virtual three-dimensional elements when video editing is performed on a two-dimensional video scene, in the embodiment of the disclosure, original video data acquired by a video acquisition device and three-dimensional scene information representing a real scene collected in the process of acquiring the original video data are provided to reconstruct a three-dimensional video scene, so that a user can conveniently perform video editing processing of a whole scene based on the three-dimensional video scene.
In specific implementation, the original video data and the three-dimensional scene information may be acquired by a video acquisition device corresponding to the video editing device, and the video acquisition device and the video editing device may be in the same terminal device or may be in different terminal devices, where no special limitation is made herein; under the condition that the two types of devices are located in different terminal equipment, original video data and original three-dimensional scene information acquired by the video acquisition device can be provided to the video editing device in the forms of video files, coded data packets, communication data frames and the like. First, how the video acquisition device acquires the original video data and the three-dimensional scene information will be described below.
In one embodiment, the video capture device may capture and obtain the raw video data and the three-dimensional scene information based on the following steps: receiving input operation of a user; in response to the input operation, original video data and original three-dimensional scene information are collected based on a preset augmented reality technology.
Please refer to fig. 2, which is a schematic diagram of triggering data acquisition according to an embodiment of the disclosure. As shown in fig. 2, the video capture device may be a terminal device, for example, a camera device in a mobile phone, and specifically, after a user opens the camera device and enters a camera initial interface, the user may click an "AR" option shown by 201 in fig. 2, and then click a recording component shown by 202, so as to trigger the camera device to capture the original video data based on a preset augmented reality technology and simultaneously collect the original three-dimensional scene information.
It should be noted that, in the embodiment of the present disclosure, the preset augmented reality technology may be determined according to a technology type corresponding to the video capture device. For example, in the case that the technology type is a first preset type, the original video data and the original three-dimensional scene information may be collected based on an arkit (augmented Reality toolkit) technology; under the condition that the technology type is a second preset type, acquiring original video data and original three-dimensional scene information based on an ARCore (augmented Reality core) technology; still alternatively, in a case where the technology type is a third preset type, the original video data And the original three-dimensional scene information may be acquired based on other technologies besides the ARkit technology And the ARCore technology, for example, a Simultaneous Localization And Mapping (SLAM) technology.
After the video acquisition device acquires the original video data and the three-dimensional scene information, the two types of information can be provided to the video editing device, so that the video editing device reconstructs the three-dimensional video scene based on the two types of information, and a user can edit and process the video.
In particular implementation, the video capture device may provide the original video data and the original three-dimensional scene information to the video editing device based on any one of: the first item is used for packaging the original video data and the original three-dimensional scene information based on a first preset video file packaging format to obtain a first original video file, and sending the first original video file to the video editing device; a second item, encoding the original video data and the original three-dimensional scene information based on a preset video encoding protocol, obtaining a target encoding packet, and sending the target encoding packet to the video editing device; the third item is used for packaging the original video data based on a second preset video file packaging format to obtain a second original video file, storing the three-dimensional scene information into an original three-dimensional information file, and sending the second original video file and the original three-dimensional information file to the video editing device; and fourthly, coding the original video data and the original three-dimensional scene information based on a preset communication protocol to obtain a target data frame, and sending the target data frame to the video editing device.
Specifically, after the video capture device captures raw video data and raw three-dimensional scene information, the raw video data may be packaged based on a video file, such as an MP4 file packaging format.
For example, the audio/video data in the original video data may be encoded according to any one of video encoding standards such as H264 and H265 in VVC, and then the encoded code stream may be encapsulated according to a standard box protocol in an MP4 encapsulation format; for example, each frame of camera pose data in the three-dimensional scene information can be separately packaged in one box, and the three-dimensional scene data in the three-dimensional scene information can be separately packaged in one or more boxes; then, packaging each box obtained as above into an MP4 file, and providing the MP4 file to the video editing device; the video editing apparatus can obtain the original video data and the original three-dimensional scene information by parsing the MP4 file.
Of course, in specific implementation, the original video data and the original three-dimensional information file may also be provided to the video capture device in a manner of directly transferring the target encoding packet. For example, the original video data may be encoded based on the H264 encoding standard to obtain corresponding code stream data, and at the same time, each frame of camera pose data is filled in the supplemental enhancement information field, i.e., the SEI field, and the target encoding packet is obtained by encapsulating the code stream data in a manner that metadata, i.e., metadata, carries three-dimensional scene data, and is provided to the video acquisition device.
In addition, the original video data and the original three-dimensional scene information may also be provided to the video editing apparatus by using the method of storing the original three-dimensional scene information in an external file as described in the above a3, wherein the second preset video file package format may be a video file package format based on a standard protocol, for example, a standard MP4 file package format.
The above has described in detail how the video acquisition device acquires the original video data and the original three-dimensional scene information, and how to provide the two types of information to the video editing device; after the video editing apparatus obtains the two types of information, three-dimensional video scene reconstruction processing can be performed based on the two types of information to provide a video editing scene for a user.
That is, after step S1100, step S1200 is performed to reconstruct a three-dimensional video scene from the original video data and the original three-dimensional scene information.
Please refer to fig. 3, which is a flowchart illustrating a three-dimensional scene reconstruction process according to an embodiment of the disclosure. As shown in fig. 3, the reconstructing a three-dimensional video scene according to original video data and original three-dimensional scene information includes: step S1210, acquiring each frame of camera attitude data and three-dimensional scene data corresponding to original video data according to original three-dimensional scene information; step S1220, performing three-dimensional scene reconstruction processing according to the original video data, the camera pose data of each frame, and the three-dimensional scene data, to obtain the three-dimensional video scene.
Step S1300, receiving an editing operation of the user for the three-dimensional video scene, and generating target video data in response to the editing operation.
After reconstructing the three-dimensional video scene in step S1200, the user can perform three-dimensional video editing processing based on the three-dimensional video scene.
In one embodiment, the editing operation comprises an operation of adding a virtual three-dimensional element; the generating target video data in response to the editing operation includes: and re-performing video data rendering processing based on the added virtual three-dimensional element to generate target video data containing the virtual three-dimensional element.
In this embodiment, the virtual three-dimensional element may be a virtual element such as an advertisement tile, a three-dimensional virtual article, a three-dimensional animation, a special effect, and the like.
To sum up, in the method provided by the embodiment of the present disclosure, the video editing apparatus obtains original video data and original three-dimensional scene information representing a real scene in the process of acquiring the original video data; reconstructing a three-dimensional video scene based on the original video data and the original three-dimensional scene information; then, the device can provide the three-dimensional video scene which can conveniently and directly carry out editing operation on the video data in the three-dimensional scene for the user, and based on the three-dimensional video scene, the device can generate the target video data by receiving the editing operation of the user and responding to the editing operation.
< method example two >
Corresponding to the first method embodiment, an embodiment of the present disclosure further provides a video capturing method, please refer to fig. 4, which is a schematic flow chart of the video capturing method provided in the embodiment of the present disclosure, and the method may be implemented by a video capturing device, where the video capturing device may specifically be a terminal device, for example, a mobile phone, a tablet computer, a personal computer, and the like, and is not limited specifically here.
As shown in FIG. 4, the method of the present embodiment may include the following steps S4100-S4200, which will be described in detail below.
In step S4100, an input operation by the user is received.
Step S4200, in response to the input operation, acquiring original video data and original three-dimensional scene information based on a preset augmented reality technology, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data.
In one embodiment, after obtaining the original video data and the original three-dimensional scene information, the method provides the original video data and the original three-dimensional scene information to a video editing device by any one of: the first item is used for packaging the original video data and the original three-dimensional scene information based on a first preset video file packaging format to obtain a first original video file, and sending the first original video file to the video editing device; a second item, encoding the original video data and the original three-dimensional scene information based on a preset video encoding protocol, obtaining a target encoding packet, and sending the target encoding packet to the video editing device; the third item is used for packaging the original video data based on a second preset video file packaging format to obtain a second original video file, storing the three-dimensional scene information into an original three-dimensional information file, and sending the second original video file and the original three-dimensional information file to the video editing device; and fourthly, coding the original video data and the original three-dimensional scene information based on a preset communication protocol to obtain a target data frame, and sending the target data frame to the video editing device.
< first embodiment of the apparatus >
Corresponding to the first method embodiment, in this embodiment, a video editing apparatus is further provided, and as shown in fig. 5, the apparatus 500 may include a data raw data obtaining module 510, a three-dimensional video scene reconstructing module 520, and a target video data generating module 530.
The original data obtaining module 510 is configured to obtain original video data and original three-dimensional scene information, where the original three-dimensional scene information is information representing a real scene in a process of acquiring the original video data; the three-dimensional video scene reconstruction module 520 is configured to reconstruct a three-dimensional video scene according to the original video data and the original three-dimensional scene information; the target video data generating module 530 is configured to receive an editing operation of a user on the three-dimensional video scene, and generate target video data in response to the editing operation.
< example II of the apparatus >
Corresponding to the above-mentioned method embodiment, in this embodiment, a video capture apparatus is further provided, and as shown in fig. 6, the apparatus 600 may include a receiving module 610 and a capturing module 620.
The receiving module 610 is configured to receive an input operation of a user; the acquiring module 620 is configured to, in response to the input operation, acquire original video data and original three-dimensional scene information based on a preset augmented reality technology, where the original three-dimensional scene information is information representing a real scene in a process of acquiring the original video data.
< apparatus embodiment >
In this embodiment, there is also provided an electronic device, as shown in fig. 7, the electronic device 700 may include a processor 720 and a memory 710, the memory 710 being configured to store executable instructions; the processor 720 is configured to operate the electronic device according to the control of the instructions to perform a method according to any embodiment of the present disclosure.
< computer-readable storage Medium embodiment >
The present embodiments provide a computer-readable storage medium having stored therein an executable command, which when executed by a processor, performs the method described in any of the method embodiments of the present specification.
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present description. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions. It is well known to those skilled in the art that implementation by hardware, implementation by software, and implementation by a combination of software and hardware are equivalent.
The foregoing description of the embodiments of the present specification has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the embodiments disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein is chosen in order to best explain the principles of the embodiments, the practical application, or improvements made to the technology in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein. The scope of the application is defined by the appended claims.

Claims (10)

1. A video editing method, comprising:
acquiring original video data and original three-dimensional scene information, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data;
reconstructing a three-dimensional video scene according to the original video data and the original three-dimensional scene information;
and receiving the editing operation of the user for the three-dimensional video scene, and responding to the editing operation to generate target video data.
2. The method of claim 1, wherein reconstructing a three-dimensional video scene from the original video data and the original three-dimensional scene information comprises:
acquiring each frame of camera attitude data and three-dimensional scene data corresponding to the original video data according to the original three-dimensional scene information, wherein the three-dimensional scene data comprises at least one of scene model data, object shielding information, scene identification plane information and scene anchor point information;
and performing three-dimensional scene reconstruction processing according to the original video data, the camera attitude data of each frame and the three-dimensional scene data to obtain the three-dimensional video scene.
3. The method of claim 2, wherein the editing operation comprises an operation of adding a virtual three-dimensional element;
the generating target video data in response to the editing operation includes:
and re-performing video data rendering processing based on the added virtual three-dimensional element to generate target video data containing the virtual three-dimensional element.
4. The method of claim 1, wherein the obtaining raw video data and raw three-dimensional scene information comprises at least one of:
receiving a first original video file obtained by a video acquisition device after encapsulating the original video data and the original three-dimensional scene information based on a first preset video file encapsulation format, and obtaining the original video data and the original three-dimensional scene information by analyzing the first original video file;
receiving a target coding packet obtained by the video acquisition device after encoding the original video data and the original three-dimensional scene information based on a preset video coding protocol, and obtaining the original video data and the original three-dimensional scene information by analyzing the target coding packet;
receiving a second original video file and an original three-dimensional information file provided by the video acquisition device, and analyzing the second original video file and the original three-dimensional information file to obtain original video data and original three-dimensional scene information, wherein the second original video file is obtained by the acquisition terminal by packaging the original video data based on a second preset video file packaging format, and the original three-dimensional information file is used for storing the three-dimensional scene information;
and acquiring a target data frame which is provided by the video acquisition device and obtained after the original video data and the original three-dimensional scene information are coded based on a preset communication protocol, and acquiring the original video data and the original three-dimensional scene information by analyzing the target data frame.
5. A video capture method, comprising:
receiving input operation of a user;
and responding to the input operation, acquiring original video data and original three-dimensional scene information based on a preset augmented reality technology, wherein the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data.
6. The method of claim 5, wherein the method is applied to a video capture device, and after obtaining the original video data and the original three-dimensional scene information, the method provides the original video data and the original three-dimensional scene information to a video editing device by any one of:
packaging the original video data and the original three-dimensional scene information based on a first preset video file packaging format to obtain a first original video file, and sending the first original video file to the video editing device;
coding the original video data and the original three-dimensional scene information based on a preset video coding protocol to obtain a target coding packet, and sending the target coding packet to the video editing device;
packaging the original video data based on a second preset video file packaging format to obtain a second original video file, storing the three-dimensional scene information into an original three-dimensional information file, and sending the second original video file and the original three-dimensional information file to the video editing device;
and coding the original video data and the original three-dimensional scene information based on a preset communication protocol to obtain a target data frame, and sending the target data frame to the video editing device.
7. A video editing apparatus, comprising:
the system comprises an original data acquisition module, a data acquisition module and a data acquisition module, wherein the original data acquisition module is used for acquiring original video data and original three-dimensional scene information, and the original three-dimensional scene information is information representing a real scene in the process of acquiring the original video data;
the three-dimensional video scene reconstruction module is used for reconstructing a three-dimensional video scene according to the original video data and the original three-dimensional scene information;
and the target video data generation module is used for receiving the editing operation of the user on the three-dimensional video scene and generating target video data in response to the editing operation.
8. A video capture device, comprising:
the receiving module is used for receiving input operation of a user;
and the acquisition module is used for responding to the input operation and acquiring original video data and original three-dimensional scene information based on a preset augmented reality technology, wherein the original three-dimensional scene information is used for representing a real scene in the process of acquiring the original video data.
9. An electronic device, comprising:
a memory for storing executable instructions;
a processor configured to execute the electronic device to perform the method according to any one of claims 1 to 6 under the control of the instructions.
10. A computer-readable storage medium, in which a computer program is stored which is readable and executable by a computer, and which, when read and executed by the computer, is adapted to perform the method according to any one of claims 1 to 6.
CN202111629147.3A 2021-12-28 2021-12-28 Video editing method and device, electronic equipment and readable storage medium Pending CN114401451A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111629147.3A CN114401451A (en) 2021-12-28 2021-12-28 Video editing method and device, electronic equipment and readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111629147.3A CN114401451A (en) 2021-12-28 2021-12-28 Video editing method and device, electronic equipment and readable storage medium

Publications (1)

Publication Number Publication Date
CN114401451A true CN114401451A (en) 2022-04-26

Family

ID=81229771

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111629147.3A Pending CN114401451A (en) 2021-12-28 2021-12-28 Video editing method and device, electronic equipment and readable storage medium

Country Status (1)

Country Link
CN (1) CN114401451A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114581611A (en) * 2022-04-28 2022-06-03 阿里巴巴(中国)有限公司 Virtual scene construction method and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070268406A1 (en) * 2006-05-22 2007-11-22 Broadcom Corporation, A California Corporation Video processing system that generates sub-frame metadata
CN109803094A (en) * 2018-12-18 2019-05-24 北京美吉克科技发展有限公司 A kind of virtual three-dimensional scene editing system, method and device
CN110049266A (en) * 2019-04-10 2019-07-23 北京字节跳动网络技术有限公司 Video data issues method, apparatus, electronic equipment and storage medium
CN111583348A (en) * 2020-05-09 2020-08-25 维沃移动通信有限公司 Image data encoding method and device, display method and device, and electronic device
CN112070901A (en) * 2020-07-21 2020-12-11 马小淞 AR scene construction method and device for garden, storage medium and terminal
CN112135091A (en) * 2020-08-27 2020-12-25 杭州张量科技有限公司 Monitoring scene marking method and device, computer equipment and storage medium
CN113253842A (en) * 2021-05-20 2021-08-13 深圳市商汤科技有限公司 Scene editing method and related device and equipment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070268406A1 (en) * 2006-05-22 2007-11-22 Broadcom Corporation, A California Corporation Video processing system that generates sub-frame metadata
CN109803094A (en) * 2018-12-18 2019-05-24 北京美吉克科技发展有限公司 A kind of virtual three-dimensional scene editing system, method and device
CN110049266A (en) * 2019-04-10 2019-07-23 北京字节跳动网络技术有限公司 Video data issues method, apparatus, electronic equipment and storage medium
CN111583348A (en) * 2020-05-09 2020-08-25 维沃移动通信有限公司 Image data encoding method and device, display method and device, and electronic device
CN112070901A (en) * 2020-07-21 2020-12-11 马小淞 AR scene construction method and device for garden, storage medium and terminal
CN112135091A (en) * 2020-08-27 2020-12-25 杭州张量科技有限公司 Monitoring scene marking method and device, computer equipment and storage medium
CN113253842A (en) * 2021-05-20 2021-08-13 深圳市商汤科技有限公司 Scene editing method and related device and equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114581611A (en) * 2022-04-28 2022-06-03 阿里巴巴(中国)有限公司 Virtual scene construction method and device
CN114581611B (en) * 2022-04-28 2022-09-20 阿里巴巴(中国)有限公司 Virtual scene construction method and device

Similar Documents

Publication Publication Date Title
EP3712854A1 (en) Method and apparatus for storage and signaling of static point cloud data
KR102450781B1 (en) Method and apparatus for encoding media data comprising generated content
CN108140263B (en) AR display system and method applied to image or video
WO2020012073A1 (en) Method and apparatus for storage and signaling of compressed point clouds
CN110430441B (en) Cloud mobile phone video acquisition method, system, device and storage medium
CN111402399A (en) Face driving and live broadcasting method and device, electronic equipment and storage medium
KR20210016530A (en) Method, device, and computer program for media content transfer
CN110213458B (en) Image data processing method and device and storage medium
CN106485653B (en) User terminal and panoramic picture dynamic thumbnail generation method
US11985381B2 (en) Mapping architecture of immersive technologies media format (ITMF) specification with rendering engines
WO2023207119A1 (en) Immersive media processing method and apparatus, device, and storage medium
JP2024504421A (en) Volumetric media processing method and device, storage medium, and electronic device
CN114401451A (en) Video editing method and device, electronic equipment and readable storage medium
CN111583348A (en) Image data encoding method and device, display method and device, and electronic device
WO2020062700A1 (en) Method for processing media data, and client, and server
US8606087B2 (en) Screen recording system and method
US20150281309A1 (en) Method, apparatus and system of video and audio sharing among communication devices
TWI802204B (en) Methods and systems for derived immersive tracks
CN112188213B (en) Encoding method, apparatus, computer device, and storage medium
US20240201784A1 (en) Methods for signaling random access in haptics interchange file format
WO2024004449A1 (en) Information processing device, information processing method, and computer program
CN113784150A (en) Video data distribution method and device, electronic equipment and storage medium
KR20170028593A (en) Method and program for making moving image
KR20220071924A (en) Encoding and decoding method and apparatus for volumemetric video
CN117666919A (en) Method, device, computer equipment and storage medium for generating multimedia file

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination