CN117135416A - Video generation system based on AIGC - Google Patents
Video generation system based on AIGC Download PDFInfo
- Publication number
- CN117135416A CN117135416A CN202311393909.3A CN202311393909A CN117135416A CN 117135416 A CN117135416 A CN 117135416A CN 202311393909 A CN202311393909 A CN 202311393909A CN 117135416 A CN117135416 A CN 117135416A
- Authority
- CN
- China
- Prior art keywords
- model
- unit
- information
- keywords
- video
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000004458 analytical method Methods 0.000 claims abstract description 21
- 238000012545 processing Methods 0.000 claims abstract description 20
- 238000000034 method Methods 0.000 claims abstract description 10
- 230000008569 process Effects 0.000 claims abstract description 10
- 230000009471 action Effects 0.000 claims description 45
- 230000000875 corresponding effect Effects 0.000 claims description 23
- 238000004364 calculation method Methods 0.000 claims description 20
- 238000006243 chemical reaction Methods 0.000 claims description 11
- 238000012549 training Methods 0.000 claims description 10
- 230000001502 supplementing effect Effects 0.000 claims description 8
- 230000008859 change Effects 0.000 claims description 4
- 238000004519 manufacturing process Methods 0.000 abstract description 7
- 238000005516 engineering process Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 238000013507 mapping Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 230000002452 interceptive effect Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000009877 rendering Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/284—Lexical analysis, e.g. tokenisation or collocates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention provides an AIGC-based video generation system, which comprises an information input module, a text analysis module, a model selection module and a video generation module, wherein the information input module is used for inputting requirement information of a user, the text analysis module is used for carrying out content analysis on the requirement information, the model selection module is used for selecting a basic model of a video, and the video generation module is used for processing the basic model to obtain a complete video; the system can intelligently identify text information input by a user, select a proper model from the model library, virtually shoot the processing process of the model to obtain a required video, greatly reduce the difficulty of video production and improve the efficiency of video production.
Description
Technical Field
The invention relates to the field of electric digital data processing, in particular to an AIGC-based video generation system.
Background
AIGC, i.e., artificial intelligence generation of content, is a field in which artificial intelligence technology is applied to create content. With the continuous development of deep learning and generating models, AIGC has made remarkable progress in various fields, from image generation to music creation to text writing and video generation, and has shown great potential, in the technology of AIGC technology, the difficulty of many works can be greatly reduced, and video production is one of the works, and a system is needed to improve the efficiency of video production on the basis of combining with the AIGC technology.
The foregoing discussion of the background art is intended to facilitate an understanding of the present invention only. This discussion is not an admission or admission that any of the material referred to was common general knowledge.
Many video production systems have been developed and found to have a system as disclosed in publication number CN113784167B, which generally includes: editing a behavior tree based on performance contents and interaction contents required by all the existing objects in the 3D scene by using a behavior tree editor, outputting corresponding behavior tree data, and enabling the interaction contents to be conditional nodes of the behavior tree; inputting behavior tree data in a pre-rendering host, and disassembling each segment of complete behavior tree data under a condition node to perform 3D real-time rendering and video recording to obtain a plurality of video files; simplifying the behavior tree data converted into the video file into a behavior of playing video, and obtaining simplified behavior tree data; and inputting the simplified behavior tree data and the video file into an interactive video player. However, the system still uses the traditional mode to make 3D video, and is low in efficiency, high in difficulty and has technical requirements for producers.
Disclosure of Invention
The invention aims to overcome the defects, and provides an AIGC-based video generation system.
The invention adopts the following technical scheme:
the AIGC-based video generation system comprises an information input module, a text analysis module, a model selection module and a video generation module;
the information input module is used for inputting the requirement information of a user, the text analysis module is used for carrying out content analysis on the requirement information, the model selection module is used for selecting a basic model of the video, and the video generation module is used for processing the basic model to obtain a complete video;
the text analysis module comprises a classification disassembly unit, a combination unit and a conversion unit, wherein the classification disassembly unit is used for disassembling text information into a plurality of keywords and classifying the keywords, the combination unit is used for combining the classified keywords based on the positions of descriptive keywords to generate three word packages, and the conversion unit is used for converting the keywords in the word packages into corresponding codes to form code packages;
the model selection module comprises a model library storage unit and a model matching unit, wherein the model library storage unit is used for storing a background model and an active subject model, and the model matching unit obtains the corresponding background model and active subject model based on the coding information in the coding packet;
the video generation module comprises a model placement unit, an action unit and a shooting unit, wherein the model placement unit is used for placing a background model and a movable main body model, the action unit enables the movable main body model to complete corresponding actions according to event coding packet information, and the shooting unit is used for shooting picture information of the movable main body model when the action is completed in the background model to obtain video data;
further, the classification disassembly unit comprises a keyword training processor, a keyword recognition processor and a keyword recording processor, wherein the keyword training processor is used for training a large number of texts to obtain recognition parameters of keywords, the recognition parameters are sent to the keyword recognition processor, the keyword recognition processor recognizes the keywords in the text information based on the recognition parameters, and each time a keyword is recognized, the keyword recognition processor sends the category and position information of the keyword and the keyword to the keyword recording processor, and the keyword recording processor is used for recording the received keyword information;
further, the model matching unit comprises an encoding processor, an annotation retrieval processor and a matching calculation processor, wherein the encoding processor is used for processing the received encoding packet, the annotation retrieval processor reduces the retrieval range of annotation information according to the processing result, and the matching calculation processor is used for carrying out matching calculation on the annotation information and the encoding packet;
further, the matching calculation processor calculates a matching degree Pm between the retrieved model annotation information and the encoded packet according to the following formula:
;
wherein Ic (i) represents an indication number of the ith body code contained in the annotation information, the indication number when contained is 1, the indication number when not contained is 0,description code number for the ith body,/-code number>The annotation information contains the number of the description codes corresponding to the ith main body code;
the matching calculation processor selects model annotation information with the largest matching degree, acquires a model from a model register according to a corresponding storage address and sends the model to the model placement unit or the action unit;
further, the process of generating the video by the video generation module comprises the following steps:
s31, the model placement unit performs initialization placement on a background model and a movable main body model;
s32, the action unit disassembles the event code packet to obtain a plurality of action instructions;
s33, the model placement unit performs action processing on the movable main body model based on action instructions to obtain end positions of the movable main body model under each action instruction, and node positions are formed by the initial positions and the end positions in sequence;
s34, the model placement unit calculates a time frame number according to the positions of two adjacent nodes;
s35, the model placement unit calculates the same number of frame supplementing positions according to the time length frame number;
s36, the model placement unit changes the movable main body model according to the sequence of the node positions and the plurality of frame supplementing positions, signals are sent to the shooting unit at each change sequence position, and the shooting unit shoots a frame of picture data after receiving the signals;
s37, the shooting unit combines all picture data to form video data.
The beneficial effects obtained by the invention are as follows:
the system carries out intelligent analysis on the demand information, divides an analysis result into a model result and an action result, screens out a proper model in a model library based on the model result, presents a model and a model action in a virtual space based on the action result, shoots the whole action process to realize the generation of a video, and a user only needs to input text information in the whole process without complex operation, thereby reducing the difficulty of video production and improving the efficiency of video production.
For a further understanding of the nature and the technical aspects of the present invention, reference should be made to the following detailed description of the invention and the accompanying drawings, which are provided for purposes of reference only and are not intended to limit the invention.
Drawings
FIG. 1 is a schematic diagram of the overall structural framework of the present invention;
FIG. 2 is a schematic diagram of a text parsing module according to the present invention;
FIG. 3 is a schematic diagram of a video generation module according to the present invention;
FIG. 4 is a schematic diagram of a model matching unit of the present invention;
fig. 5 is a schematic diagram of the model placement unit of the present invention.
Detailed Description
The following embodiments of the present invention are described in terms of specific examples, and those skilled in the art will appreciate the advantages and effects of the present invention from the disclosure herein. The invention is capable of other and different embodiments and its several details are capable of modification and variation in various respects, all without departing from the spirit of the present invention. The drawings of the present invention are merely schematic illustrations, and are not intended to be drawn to actual dimensions. The following embodiments will further illustrate the related art content of the present invention in detail, but the disclosure is not intended to limit the scope of the present invention.
Embodiment one: the embodiment provides a video generation system based on AIGC, which comprises an information input module, a text analysis module, a model selection module and a video generation module, and is combined with FIG. 1;
the information input module is used for inputting the requirement information of a user, the text analysis module is used for carrying out content analysis on the requirement information, the model selection module is used for selecting a basic model of the video, and the video generation module is used for processing the basic model to obtain a complete video;
the text analysis module comprises a classification disassembly unit, a combination unit and a conversion unit, wherein the classification disassembly unit is used for disassembling text information into a plurality of keywords and classifying the keywords, the combination unit is used for combining the classified keywords based on the positions of descriptive keywords to generate three word packages, and the conversion unit is used for converting the keywords in the word packages into corresponding codes to form code packages;
the model selection module comprises a model library storage unit and a model matching unit, wherein the model library storage unit is used for storing a background model and an active subject model, and the model matching unit obtains the corresponding background model and active subject model based on the coding information in the coding packet;
the video generation module comprises a model placement unit, an action unit and a shooting unit, wherein the model placement unit is used for placing a background model and a movable main body model, the action unit enables the movable main body model to complete corresponding actions according to event coding packet information, and the shooting unit is used for shooting picture information of the movable main body model when the action is completed in the background model to obtain video data;
the classifying and disassembling unit comprises a keyword training processor, a keyword recognition processor and a keyword recording processor, wherein the keyword training processor is used for training a large number of texts to obtain recognition parameters of keywords, the recognition parameters are sent to the keyword recognition processor, the keyword recognition processor recognizes the keywords in the text information based on the recognition parameters, and each time a keyword is recognized, the keyword recognition processor sends the category and position information of the keyword and the keyword to the keyword recording processor, and the keyword recording processor is used for recording the received keyword information;
the model matching unit comprises an encoding processor, an annotation retrieval processor and a matching calculation processor, wherein the encoding processor is used for processing a received encoding packet, the annotation retrieval processor reduces the retrieval range of annotation information according to a processing result, and the matching calculation processor is used for carrying out matching calculation on the annotation information and the encoding packet;
the matching calculation processor calculates the matching degree Pm between the retrieved model annotation information and the coding packet according to the following formula:
;
wherein Ic (i) represents an indication number of the ith body code contained in the annotation information, the indication number when contained is 1, the indication number when not contained is 0,description code number for the ith body,/-code number>The annotation information contains the number of the description codes corresponding to the ith main body code;
the matching calculation processor selects model annotation information with the largest matching degree, acquires a model from a model register according to a corresponding storage address and sends the model to the model placement unit or the action unit;
the process of generating the video by the video generation module comprises the following steps:
s31, the model placement unit performs initialization placement on a background model and a movable main body model;
s32, the action unit disassembles the event code packet to obtain a plurality of action instructions;
s33, the model placement unit performs action processing on the movable main body model based on action instructions to obtain end positions of the movable main body model under each action instruction, and node positions are formed by the initial positions and the end positions in sequence;
s34, the model placement unit calculates a time frame number according to the positions of two adjacent nodes;
s35, the model placement unit calculates the same number of frame supplementing positions according to the time length frame number;
s36, the model placement unit changes the movable main body model according to the sequence of the node positions and the plurality of frame supplementing positions, signals are sent to the shooting unit at each change sequence position, and the shooting unit shoots a frame of picture data after receiving the signals;
s37, the shooting unit combines all picture data to form video data.
Embodiment two: the embodiment includes the whole content in the first embodiment, and provides a video generation system based on AIGC, which comprises an information input module, a text analysis module, a model selection module and a video generation module;
the information input module is used for inputting the requirement information of a user, the text analysis module is used for carrying out content analysis on the requirement information, the model selection module is used for selecting a basic model of the video, and the video generation module is used for processing the basic model to obtain a complete video;
the information input module comprises an interactive display unit and an information acquisition unit, wherein the interactive display unit is used for displaying window information and inputting text information in a window through input equipment, and the information acquisition unit is used for acquiring the text information in the window and sending the text information to the text analysis module;
referring to fig. 2, the text parsing module includes a classification disassembly unit, a combination unit and a conversion unit, where the classification disassembly unit is configured to disassemble text information into a plurality of keywords and divide the keywords into descriptive keywords, static main keywords, dynamic main keywords and event keywords, the combination unit combines descriptive keyword codes with the static main keywords, the dynamic main keywords and the event keywords respectively based on the positions of the descriptive keywords to generate three word packages, and the conversion unit is configured to convert the keywords in the word packages into corresponding codes to form code packages, which are respectively called a background code package, an activity code package and an event code package, and the background code package and the activity code package are sent to the model selection module, and the event code package is sent to the video generation module;
the model selection module comprises a model library storage unit and a model matching unit, wherein the model library storage unit is used for storing a background model and an active subject model, the model matching unit obtains a corresponding background model and an active subject model based on the coding information in the coding packet, and model data obtained by matching is sent to the video generation module;
referring to fig. 3, the video generating module includes a model placement unit, an action unit and a shooting unit, where the model placement unit is used to place a background model and a movable main body model, the action unit makes the movable main body model complete corresponding actions according to event coding packet information, and the shooting unit is used to shoot picture information of the movable main body model when the action is completed in the background model to obtain video data;
the classifying and disassembling unit comprises a keyword training processor, a keyword recognition processor and a keyword recording processor, wherein the keyword training processor trains a large number of texts based on an AIGC technology to obtain recognition parameters of keywords, the recognition parameters are sent to the keyword recognition processor, the keyword recognition processor recognizes the keywords in the text information based on the recognition parameters, and each time a keyword is recognized, the keyword recognition processor sends the category and position information of the keyword and the keyword to the keyword recording processor, and the keyword recording processor is used for recording the received keyword information;
the process of combining the keywords by the combining unit comprises the following steps:
s1, selecting descriptive keywords in the keyword recording processor as first target keywords, using the rest keywords as second target keywords, and establishing a description package for each second target keyword;
s2, selecting a first target keyword, and calculating a combination index Cb with the rest second target keywords respectively:
;
wherein Np1 is the position number of the first target keyword, np2 is the position number of the second target keyword,the description association coefficients of the first target keyword and the second target keyword are obtained;
s3, adding the first target keywords into a description packet of a second target keyword with the smallest combination index;
s4, repeating the step S2 and the step S3 until all the first target keywords are added into the description package;
s5, integrating all the static main body keywords and the description packages into a word package, integrating all the dynamic main body keywords and the description packages into a word package, and integrating all the event keywords and the description packages into a word package;
the conversion unit comprises a code mapping register and a conversion processor, wherein the code mapping register is used for storing the mapping relation between keywords and codes, one code can map a plurality of keywords, and the conversion processor converts the keywords in the word packet into corresponding codes according to the mapping relation;
the model storage unit comprises a model register and a model annotation register, the model register divides a model into a background model and an active subject model, the model is stored in a single model unit, and the model annotation register is used for recording annotation information of each model, the type of the corresponding model and a storage address in the model register;
referring to fig. 4, the model matching unit includes an encoding processor, an annotation retrieval processor, and a matching calculation processor, where the encoding processor is configured to process a received encoded packet, the annotation retrieval processor reduces a retrieval range of annotation information according to a processing result, and the matching calculation processor is configured to perform matching calculation on the annotation information and the encoded packet;
the codes in the code package are divided into a main code and a description code, wherein the main code refers to the code converted by the keyword in the non-description package, the description code refers to the code converted by the keyword in the description package, and the main code sequentially represents the classification fineness from the first position to the last position;
the process of processing the main body code in the code packet by the code processor comprises the following steps:
s21, acquiring a main body code in a code packet;
s22, selecting two main body codes with the smallest difference value, and respectively marking the two main body codes as Cd (1) and Cd (2), wherein the average value of the two main body codes is Cd';
s23, sorting the rest main body codes according to the absolute value of the difference value with Cd' from small to large, respectively marking the main body codes as Cd (3), cd (4), and Cd (m), wherein m is the number of the main body codes in the code packet;
s24, determining a search code set according to Cd (1) and Cd (2), wherein the search code set comprises search codes which have the same bit number as the main body code and are a plurality of continuous 0S; the search code Sc satisfies the following relation:
;
wherein n is the number of continuous 0 s of the last bit of the search code, and Ys is a search threshold;
s25, calculating a retrieval index Ps of each retrieval code according to the following formula:
;
wherein,is a range control parameter, and is set by a user;
s26, sending the retrieval code with the minimum retrieval index to the annotation retrieval processor;
the matching calculation processor calculates the matching degree Pm between the retrieved model annotation information and the coding packet according to the following formula:
;
wherein Ic (i) represents an indication number of the ith body code contained in the annotation information, the indication number when contained is 1, the indication number when not contained is 0,description code number for the ith body,/-code number>The annotation information contains the number of the description codes corresponding to the ith main body code;
the matching calculation processor selects model annotation information with the largest matching degree, acquires a model from a model register according to a corresponding storage address and sends the model to the model placement unit or the action unit;
the process of generating the video by the video generation module comprises the following steps:
s31, the model placement unit performs initialization placement on a background model and a movable main body model;
s32, the action unit disassembles the event code packet to obtain a plurality of action instructions;
s33, the model placement unit performs action processing on the movable main body model based on action instructions to obtain end positions of the movable main body model under each action instruction, and node positions are formed by the initial positions and the end positions in sequence;
s34, the model placement unit calculates a time frame number according to the positions of two adjacent nodes;
s35, the model placement unit calculates the same number of frame supplementing positions according to the time length frame number;
s36, the model placement unit changes the movable main body model according to the sequence of the node positions and the plurality of frame supplementing positions, signals are sent to the shooting unit at each change sequence position, and the shooting unit shoots a frame of picture data after receiving the signals;
s37, the shooting unit combines all picture data to form video data;
referring to fig. 5, the model placement unit includes a model display processor, an action processor, and a position processor, where the model display processor is configured to provide a virtual space for placing a model, the action processor is configured to execute an action instruction to make the model generate an action, and the position processor is configured to record position information of the model and perform calculation processing according to the position information;
the position processor calculates a time frame number T according to the following formula:
;
wherein L is the maximum moving distance of the model at two node positions, and v is the single frame moving speed of the action instruction;
the position processor calculates the frame compensating position according to the following formula:
;
;
;
Wherein,for the previous node position, +.>For the latter node position, j is the complement frame number.
The foregoing disclosure is only a preferred embodiment of the present invention and is not intended to limit the scope of the invention, so that all equivalent technical changes made by applying the description of the present invention and the accompanying drawings are included in the scope of the present invention, and in addition, elements in the present invention can be updated as the technology develops.
Claims (5)
1. The AIGC-based video generation system is characterized by comprising an information input module, a text analysis module, a model selection module and a video generation module;
the information input module is used for inputting the requirement information of a user, the text analysis module is used for carrying out content analysis on the requirement information, the model selection module is used for selecting a basic model of the video, and the video generation module is used for processing the basic model to obtain a complete video;
the text analysis module comprises a classification disassembly unit, a combination unit and a conversion unit, wherein the classification disassembly unit is used for disassembling text information into a plurality of keywords and classifying the keywords, the combination unit is used for combining the classified keywords based on the positions of descriptive keywords to generate three word packages, and the conversion unit is used for converting the keywords in the word packages into corresponding codes to form code packages;
the model selection module comprises a model library storage unit and a model matching unit, wherein the model library storage unit is used for storing a background model and an active subject model, and the model matching unit obtains the corresponding background model and active subject model based on the coding information in the coding packet;
the video generation module comprises a model placement unit, an action unit and a shooting unit, wherein the model placement unit is used for placing a background model and an active main body model, the action unit enables the active main body model to complete corresponding actions according to event coding packet information, and the shooting unit is used for shooting picture information of the active main body model when the active main body model completes actions in the background model to obtain video data.
2. The AIGC-based video generation system of claim 1, wherein the classification disassembly unit includes a keyword training processor for training a plurality of texts to obtain recognition parameters of keywords, a keyword recognition processor for recognizing keywords in text information based on the recognition parameters, and a keyword recording processor for recording received keyword information by transmitting the category and position information of the keywords and the keywords themselves to each time one keyword is recognized.
3. The AIGC-based video generation system of claim 2, wherein the model matching unit includes an encoding processor for processing the received encoded packets, an annotation retrieval processor for narrowing retrieval range of annotation information according to the processing result, and a matching calculation processor for performing matching calculation of the annotation information with the encoded packets.
4. The AIGC-based video generation system of claim 3, wherein the matching calculation processor calculates the matching degree Pm between the retrieved model annotation information and the encoded package according to the following formula:
;
wherein Ic (i) represents an indication number of the ith body code contained in the annotation information, the indication number when contained is 1, the indication number when not contained is 0,description code number for the ith body,/-code number>The annotation information contains the number of the description codes corresponding to the ith main body code;
and the matching calculation processor selects model annotation information with the maximum matching degree, acquires a model from a model register according to the corresponding storage address and sends the model to the model placement unit or the action unit.
5. The AIGC-based video generation system of claim 4, wherein the process of generating video by the video generation module comprises the steps of:
s31, the model placement unit performs initialization placement on a background model and a movable main body model;
s32, the action unit disassembles the event code packet to obtain a plurality of action instructions;
s33, the model placement unit performs action processing on the movable main body model based on action instructions to obtain end positions of the movable main body model under each action instruction, and node positions are formed by the initial positions and the end positions in sequence;
s34, the model placement unit calculates a time frame number according to the positions of two adjacent nodes;
s35, the model placement unit calculates the same number of frame supplementing positions according to the time length frame number;
s36, the model placement unit changes the movable main body model according to the sequence of the node positions and the plurality of frame supplementing positions, signals are sent to the shooting unit at each change sequence position, and the shooting unit shoots a frame of picture data after receiving the signals;
s37, the shooting unit combines all picture data to form video data.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311393909.3A CN117135416B (en) | 2023-10-26 | 2023-10-26 | Video generation system based on AIGC |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311393909.3A CN117135416B (en) | 2023-10-26 | 2023-10-26 | Video generation system based on AIGC |
Publications (2)
Publication Number | Publication Date |
---|---|
CN117135416A true CN117135416A (en) | 2023-11-28 |
CN117135416B CN117135416B (en) | 2023-12-22 |
Family
ID=88858486
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311393909.3A Active CN117135416B (en) | 2023-10-26 | 2023-10-26 | Video generation system based on AIGC |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117135416B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108597025A (en) * | 2018-04-03 | 2018-09-28 | 中国传媒大学 | Accelerated model construction method and device based on artificial intelligence Virtual reality |
US20220108726A1 (en) * | 2020-10-01 | 2022-04-07 | Loop Now Technologies, Inc. | Machine learned video template usage |
US20230089566A1 (en) * | 2020-05-30 | 2023-03-23 | Huawei Technologies Co., Ltd. | Video generation method and related apparatus |
US20230118572A1 (en) * | 2021-10-20 | 2023-04-20 | Snap Inc. | Generating ground truths for machine learning |
US20230237772A1 (en) * | 2022-01-21 | 2023-07-27 | Salesforce, Inc. | Systems and methods for unified vision-language understanding and generation |
KR20230133059A (en) * | 2022-03-10 | 2023-09-19 | 주식회사 엘지유플러스 | Ai-based digital contents automated production method, apparatus and system |
CN116881426A (en) * | 2023-08-30 | 2023-10-13 | 环球数科集团有限公司 | AIGC-based self-explanatory question-answering system |
-
2023
- 2023-10-26 CN CN202311393909.3A patent/CN117135416B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108597025A (en) * | 2018-04-03 | 2018-09-28 | 中国传媒大学 | Accelerated model construction method and device based on artificial intelligence Virtual reality |
US20230089566A1 (en) * | 2020-05-30 | 2023-03-23 | Huawei Technologies Co., Ltd. | Video generation method and related apparatus |
US20220108726A1 (en) * | 2020-10-01 | 2022-04-07 | Loop Now Technologies, Inc. | Machine learned video template usage |
US20230118572A1 (en) * | 2021-10-20 | 2023-04-20 | Snap Inc. | Generating ground truths for machine learning |
US20230237772A1 (en) * | 2022-01-21 | 2023-07-27 | Salesforce, Inc. | Systems and methods for unified vision-language understanding and generation |
KR20230133059A (en) * | 2022-03-10 | 2023-09-19 | 주식회사 엘지유플러스 | Ai-based digital contents automated production method, apparatus and system |
CN116881426A (en) * | 2023-08-30 | 2023-10-13 | 环球数科集团有限公司 | AIGC-based self-explanatory question-answering system |
Also Published As
Publication number | Publication date |
---|---|
CN117135416B (en) | 2023-12-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Ferreira et al. | Learning to dance: A graph convolutional adversarial network to generate realistic dance motions from audio | |
CN111581437A (en) | Video retrieval method and device | |
Yuan et al. | Spatiotemporal modeling for video summarization using convolutional recurrent neural network | |
US20230185848A1 (en) | Systems and methods for generating improved content based on matching mappings | |
CN111626049B (en) | Title correction method and device for multimedia information, electronic equipment and storage medium | |
CN111429885A (en) | Method for mapping audio clip to human face-mouth type key point | |
JP2009503732A (en) | Associative matrix method, system and computer program product using bit-plane representation of selected segments | |
WO2021238081A1 (en) | Voice packet recommendation method, apparatus and device, and storage medium | |
CN112489676A (en) | Model training method, device, equipment and storage medium | |
US20240177697A1 (en) | Audio data processing method and apparatus, computer device, and storage medium | |
CN114390217A (en) | Video synthesis method and device, computer equipment and storage medium | |
CN108805036A (en) | A kind of new non-supervisory video semanteme extracting method | |
CN116051688A (en) | Transition animation generation method and device, computer readable storage medium and terminal | |
CN117376502A (en) | Video production system based on AI technology | |
US11615132B2 (en) | Feature amount generation method, feature amount generation device, and feature amount generation program | |
CN113569068B (en) | Descriptive content generation method, visual content encoding and decoding method and device | |
CN117135416B (en) | Video generation system based on AIGC | |
CN109670071B (en) | Serialized multi-feature guided cross-media Hash retrieval method and system | |
Li et al. | Rhythm-aware sequence-to-sequence learning for labanotation generation with gesture-sensitive graph convolutional encoding | |
CN113761843A (en) | Voice editing method, electronic device and computer readable storage medium | |
JP2011221845A (en) | Text processing apparatus, text processing method and text processing program | |
CN116309965A (en) | Animation generation method and device, computer readable storage medium and terminal | |
JPH09128401A (en) | Moving picture retrieval device and video-on-demand device | |
KR20220051575A (en) | Deep learning-based movie scene creation method | |
Hammad et al. | Characterizing the impact of using features extracted from pre-trained models on the quality of video captioning sequence-to-sequence models |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |