CN116320521A - Three-dimensional animation live broadcast method and device based on artificial intelligence - Google Patents
Three-dimensional animation live broadcast method and device based on artificial intelligence Download PDFInfo
- Publication number
- CN116320521A CN116320521A CN202310294269.4A CN202310294269A CN116320521A CN 116320521 A CN116320521 A CN 116320521A CN 202310294269 A CN202310294269 A CN 202310294269A CN 116320521 A CN116320521 A CN 116320521A
- Authority
- CN
- China
- Prior art keywords
- audio
- video data
- dimensional animation
- live broadcast
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000013473 artificial intelligence Methods 0.000 title claims abstract description 31
- 238000000034 method Methods 0.000 title claims abstract description 29
- 238000012216 screening Methods 0.000 claims abstract description 29
- 238000012545 processing Methods 0.000 claims abstract description 17
- 238000007781 pre-processing Methods 0.000 claims abstract description 12
- 230000003993 interaction Effects 0.000 claims description 48
- 230000009471 action Effects 0.000 claims description 15
- 230000010365 information processing Effects 0.000 claims description 15
- 238000012549 training Methods 0.000 claims description 11
- 230000002452 interceptive effect Effects 0.000 claims description 9
- 238000004364 calculation method Methods 0.000 claims description 7
- 230000009467 reduction Effects 0.000 claims description 4
- 238000013528 artificial neural network Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 4
- 238000001914 filtration Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 2
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/2187—Live feed
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/441—Acquiring end-user identification, e.g. using personal code sent by the remote control or by inserting a card
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4662—Learning process for intelligent management, e.g. learning user preferences for recommending movies characterized by learning algorithms
- H04N21/4666—Learning process for intelligent management, e.g. learning user preferences for recommending movies characterized by learning algorithms using neural networks, e.g. processing the feedback provided by the user
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D30/00—Reducing energy consumption in communication networks
- Y02D30/70—Reducing energy consumption in communication networks in wireless communication networks
Landscapes
- Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Processing Or Creating Images (AREA)
Abstract
The invention discloses a three-dimensional animation live broadcast method and device based on artificial intelligence, and particularly relates to the technical field of live broadcast, comprising the following steps: acquiring audio and video data through an acquisition module, and preprocessing the audio and video data; screening the preprocessed audio and video data through a data screening module to remove repeated data; the method comprises the steps that an information identification module is used for identifying video images in the audio and video data after duplication removal, and the video images are input into a real-time motion capture model to capture motion information; the method and the device can identify, compare and screen the acquired audio and video data through the data screening module, and delete repeated redundant audio and video data, so that the aim of removing duplication of the audio and video data is fulfilled, the audio and video data is more simplified, the identification and the rapid processing of the audio and video data are facilitated, the redundancy of the data is further avoided, the smoothness of live broadcasting of a cloud live broadcasting management platform is improved, and the live broadcasting effect is improved.
Description
Technical Field
The invention relates to the technical field of live broadcasting, in particular to a three-dimensional animation live broadcasting method and device based on artificial intelligence.
Background
At present, new media live broadcast is continuously used as a commodity selling mode of merchants. With the development of computer technology, the artificial intelligence is gradually applied to new media live broadcast, so that the existing method for generating three-dimensional animation for live broadcast through simulation action gradually becomes one of live broadcast modes, so that the new media live broadcast is gradually intelligent, the labor intensity of the artificial live broadcast can be reduced, and the live broadcast efficiency is improved.
However, when the existing three-dimensional animation live broadcast method based on artificial intelligence is actually used, audio and video data are acquired, the data cannot be screened and de-duplicated, the data are long, and the main technical problem that the live broadcast effect is affected due to the fact that a cloud live broadcast management platform is blocked is solved.
Disclosure of Invention
The invention aims to provide a three-dimensional animation live broadcast method and device based on artificial intelligence, which solve the defects in the technology.
In order to achieve the above object, the present invention provides the following technical solutions:
a three-dimensional animation live broadcast method based on artificial intelligence comprises the following steps:
s1, acquiring audio and video data through an acquisition module, and preprocessing the audio and video data;
s2, screening the preprocessed audio and video data through a data screening module to remove repeated data;
s3, identifying video images in the audio and video data subjected to duplication removal through an information identification module, and inputting the video images into a real-time motion capture model to capture motion information so as to obtain motion information;
s4, inputting the action information into a three-dimensional animation binding logic module through a data processing module, and calculating to obtain three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module;
s5, acquiring a three-dimensional animation initial model through an animation generation module, and driving the three-dimensional animation initial model according to the three-dimensional animation control parameters to generate a three-dimensional animation;
s6, identifying audio data in the audio-video data through an audio pairing module, and pairing the audio with actions in the three-dimensional animation so as to realize dubbing of the three-dimensional animation;
s7, transmitting the dubbed three-dimensional animation to a cloud live broadcast management platform for live broadcast through the data processing module.
In one possible implementation manner, the acquisition module in S1 includes an audio information acquisition unit and a video image data acquisition unit.
In a possible implementation manner, the preprocessing of the audio and video data in S1 mainly includes performing noise reduction processing on the audio information acquired by the audio information acquisition unit.
In a possible implementation manner, the step of filtering the preprocessed audio/video data through the data filtering module in the step S2, and removing the duplicate data specifically includes the following steps:
s21, acquiring the preprocessed audio and video data through a collecting unit;
s22, identifying the acquired audio and video data through an identification unit;
s23, comparing and checking the audio and video data identified by the data identification unit through a comparison unit;
s24, deleting redundant audio and video data through a screening unit, so that each audio and video data only remains one, and repeating is completed;
s25, storing and backing up the audio and video data subjected to duplication removal through a storage unit, and preventing the audio and video data from being lost.
In a possible implementation manner, the cloud live broadcast management platform in S7 further includes a live broadcast interaction management unit, where the live broadcast interaction management unit is configured to implement interaction between the user side and the viewing fan.
In one possible implementation manner, the specific steps of the live interaction management unit for implementing the interaction between the user side and the viewing fan are as follows:
(1) Acquiring interaction information sent by a user side through the live broadcast interaction management unit;
(2) Obtaining an interactive information processing model through artificial intelligence learning training;
(3) The live broadcast interaction management unit inputs the acquired interaction information into the interaction information processing model and outputs the interaction information processed by the interaction information processing model;
(4) And the live broadcast interaction management unit feeds back corresponding audio and video data or three-dimensional animation or custom reply according to the processed interaction information.
In one possible implementation manner, the live interaction management unit obtains interaction information sent by the user side, including but not limited to text comments and text popups.
In one possible implementation manner, the interactive information processing model obtained through artificial intelligence learning training mainly uses a semantic system training sample and a semantic information rule sample to learn and train an artificial neural network model so as to obtain an interactive information processing model for identifying semantics.
A three-dimensional animated live broadcast device based on artificial intelligence, comprising:
the acquisition module is used for acquiring audio and video data and preprocessing the audio and video data;
the data screening module is used for screening and de-duplicating the preprocessed audio and video data;
the information identification module is used for identifying video images in the audio and video data after the duplication removal, inputting the video images into the real-time motion capture model for capturing motion information and obtaining the motion information;
the animation generation module is used for acquiring a three-dimensional animation initial model, driving the three-dimensional animation initial model according to the three-dimensional animation control parameters and generating a three-dimensional animation;
the audio pairing module is used for identifying the audio data in the audio-video data and pairing the audio with the action in the three-dimensional animation so as to realize the dubbing of the three-dimensional animation;
the data processing module is used for inputting the action information into the three-dimensional animation binding logic module, calculating to obtain three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module, and transmitting the dubbed three-dimensional animation to the cloud live broadcast management platform for live broadcast.
In one possible implementation, the data filtering module includes:
the receiving unit is used for acquiring the preprocessed audio and video data;
the identification unit is used for identifying the acquired audio and video data;
the comparison unit is used for comparing and checking the audio and video data identified by the data identification unit;
the screening unit is used for deleting redundant audio and video data, so that each audio and video data only remains one, and the duplicate removal is completed;
and the storage unit is used for storing and backing up the audio and video data after the duplication removal and preventing the audio and video data from being lost.
In the technical scheme, the invention has the technical effects and advantages that:
the method and the device can identify, compare and screen the acquired audio and video data through the data screening module, and delete repeated redundant audio and video data, so that the aim of removing duplication of the audio and video data is fulfilled, the audio and video data is more simplified, the identification and the rapid processing of the audio and video data are facilitated, the redundancy of the data is further avoided, the smoothness of live broadcasting of a cloud live broadcasting management platform is improved, and the live broadcasting effect is improved.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings that are needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the present invention, and other drawings may be obtained according to these drawings for a person having ordinary skill in the art.
FIG. 1 is a flow chart of a three-dimensional animation live broadcast method based on artificial intelligence;
FIG. 2 is a flowchart of S2 in an artificial intelligence based three-dimensional animation live broadcast method provided by the invention;
FIG. 3 is a system block diagram of an artificial intelligence based three-dimensional animation live broadcast device provided by the invention;
fig. 4 is a block diagram of a data screening module in the three-dimensional animation live broadcast device based on artificial intelligence.
Detailed Description
In order to make the technical scheme of the present invention better understood by those skilled in the art, the present invention will be further described in detail with reference to the accompanying drawings.
The technical scheme in the embodiment of the application aims to solve the problems, and the overall thought is as follows:
example 1
As shown in fig. 1 and fig. 2, the specific structure of the embodiment is that a three-dimensional animation live broadcast method based on artificial intelligence includes the following steps:
s1, acquiring audio and video data through an acquisition module, and preprocessing the audio and video data;
s2, screening the preprocessed audio and video data through a data screening module to remove repeated data;
s3, identifying video images in the audio and video data subjected to duplication removal through an information identification module, and inputting the video images into a real-time motion capture model to capture motion information so as to obtain motion information;
s4, inputting action information into a three-dimensional animation binding logic module through a data processing module, and calculating to obtain three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module;
s5, acquiring a three-dimensional animation initial model through an animation generation module, and driving the three-dimensional animation initial model according to three-dimensional animation control parameters to generate a three-dimensional animation;
s6, identifying audio data in the audio-video data through an audio pairing module, and pairing the audio with actions in the three-dimensional animation so as to realize dubbing of the three-dimensional animation;
s7, transmitting the dubbed three-dimensional animation to a cloud live broadcast management platform through a data processing module for live broadcast.
In some examples, the acquisition module in S1 includes an audio information acquisition unit and a video image data acquisition unit.
In some examples, the preprocessing of the audio and video data in S1 mainly includes performing noise reduction processing on the audio information acquired by the audio information acquisition unit.
In some examples, in S2, the filtering the preprocessed audio/video data by the data filtering module, and removing the duplicate data specifically includes the following steps:
s21, acquiring the preprocessed audio and video data through a collecting unit;
s22, identifying the acquired audio and video data through an identification unit;
s23, comparing and checking the audio and video data identified by the data identification unit through the comparison unit;
s24, deleting redundant audio and video data through a screening unit, so that each audio and video data only remains one, and repeating is completed;
s25, storing and backing up the audio and video data subjected to duplication removal through a storage unit, and preventing the audio and video data from being lost.
Example 2
As shown in fig. 1 and fig. 2, the specific structure of the embodiment is that a three-dimensional animation live broadcast method based on artificial intelligence includes the following steps:
s1, acquiring audio and video data through an acquisition module, and preprocessing the audio and video data;
s2, screening the preprocessed audio and video data through a data screening module to remove repeated data;
s3, identifying video images in the audio and video data subjected to duplication removal through an information identification module, and inputting the video images into a real-time motion capture model to capture motion information so as to obtain motion information;
s4, inputting action information into a three-dimensional animation binding logic module through a data processing module, and calculating to obtain three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module;
s5, acquiring a three-dimensional animation initial model through an animation generation module, and driving the three-dimensional animation initial model according to three-dimensional animation control parameters to generate a three-dimensional animation;
s6, identifying audio data in the audio-video data through an audio pairing module, and pairing the audio with actions in the three-dimensional animation so as to realize dubbing of the three-dimensional animation;
s7, transmitting the dubbed three-dimensional animation to a cloud live broadcast management platform through a data processing module for live broadcast.
In some examples, the acquisition module in S1 includes an audio information acquisition unit and a video image data acquisition unit.
In some examples, the preprocessing of the audio and video data in S1 mainly includes performing noise reduction processing on the audio information acquired by the audio information acquisition unit.
In some examples, in S2, the filtering the preprocessed audio/video data by the data filtering module, and removing the duplicate data specifically includes the following steps:
s21, acquiring the preprocessed audio and video data through a collecting unit;
s22, identifying the acquired audio and video data through an identification unit;
s23, comparing and checking the audio and video data identified by the data identification unit through the comparison unit;
s24, deleting redundant audio and video data through a screening unit, so that each audio and video data only remains one, and repeating is completed;
s25, storing and backing up the audio and video data subjected to duplication removal through a storage unit, and preventing the audio and video data from being lost.
In some examples, the cloud live broadcast management platform in S7 further includes a live broadcast interaction management unit, where the live broadcast interaction management unit is configured to implement interaction between the user side and viewing the fan.
In some examples, the specific steps of the live interaction management unit for implementing interaction between the user side and viewing the fan are as follows:
(1) Acquiring interaction information sent by a user terminal through a live broadcast interaction management unit;
(2) Obtaining an interactive information processing model through artificial intelligence learning training;
(3) The live broadcast interaction management unit inputs the acquired interaction information into an interaction information processing model and outputs the interaction information processed by the interaction information processing model;
(4) And the live interaction management unit feeds back corresponding audio and video data or three-dimensional animation or custom reply according to the processed interaction information.
In some examples, the live interaction management unit obtains interaction information sent by the user side, including but not limited to text comments and text bullet screens.
In some examples, the interactive information processing model is obtained through artificial intelligence learning training, and mainly learning training is performed on the artificial neural network model by using a semantic system training sample and a semantic information rule sample to obtain the interactive information processing model for identifying semantics.
In addition, according to the two embodiments, the invention further provides a three-dimensional animation live broadcast device based on artificial intelligence, as shown in fig. 3 and fig. 4, which comprises:
the acquisition module is used for acquiring audio and video data and preprocessing the audio and video data;
the data screening module is used for screening and de-duplicating the preprocessed audio and video data;
the information identification module is used for identifying video images in the audio and video data after the duplication removal, inputting the video images into the real-time motion capture model for capturing motion information and obtaining the motion information;
the animation generation module is used for acquiring the three-dimensional animation initial model, driving the three-dimensional animation initial model according to the three-dimensional animation control parameters and generating a three-dimensional animation;
the audio pairing module is used for identifying audio data in the audio-video data and pairing the audio with actions in the three-dimensional animation so as to realize dubbing of the three-dimensional animation;
the data processing module is used for inputting the action information into the three-dimensional animation binding logic module, calculating three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module, and transmitting the dubbed three-dimensional animation to the cloud live broadcast management platform for live broadcast.
In some examples, the data screening module includes:
the receiving unit is used for acquiring the preprocessed audio and video data;
the identification unit is used for identifying the acquired audio and video data;
the comparison unit is used for comparing the audio and video data identified by the data identification unit and checking the audio and video data;
the screening unit is used for deleting redundant audio and video data, so that each audio and video data only remains one, and the duplicate removal is completed;
and the storage unit is used for storing and backing up the audio and video data after the duplication removal and preventing the audio and video data from being lost.
While certain exemplary embodiments of the present invention have been described above by way of illustration only, it will be apparent to those of ordinary skill in the art that modifications may be made to the described embodiments in various different ways without departing from the spirit and scope of the invention. Accordingly, the drawings and description are to be regarded as illustrative in nature and not as restrictive of the scope of the invention, which is defined by the appended claims.
Claims (10)
1. The three-dimensional animation live broadcast method based on artificial intelligence is characterized by comprising the following steps of:
s1, acquiring audio and video data through an acquisition module, and preprocessing the audio and video data;
s2, screening the preprocessed audio and video data through a data screening module to remove repeated data;
s3, identifying video images in the audio and video data subjected to duplication removal through an information identification module, and inputting the video images into a real-time motion capture model to capture motion information so as to obtain motion information;
s4, inputting the action information into a three-dimensional animation binding logic module through a data processing module, and calculating to obtain three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module;
s5, acquiring a three-dimensional animation initial model through an animation generation module, and driving the three-dimensional animation initial model according to the three-dimensional animation control parameters to generate a three-dimensional animation;
s6, identifying audio data in the audio-video data through an audio pairing module, and pairing the audio with actions in the three-dimensional animation so as to realize dubbing of the three-dimensional animation;
s7, transmitting the dubbed three-dimensional animation to a cloud live broadcast management platform for live broadcast through the data processing module.
2. The artificial intelligence based three-dimensional animation live broadcast method according to claim 1, wherein the method comprises the following steps: the acquisition module in the S1 comprises an audio information acquisition unit and a video image data acquisition unit.
3. The artificial intelligence based three-dimensional animation live broadcast method according to claim 2, wherein: the preprocessing of the audio and video data in the S1 mainly comprises the step of carrying out noise reduction processing on the audio information acquired by the audio information acquisition unit.
4. The artificial intelligence based three-dimensional animation live broadcast method according to claim 1, wherein the method comprises the following steps: in the step S2, the preprocessed audio and video data is screened by a data screening module, and the step of removing the repeated data specifically comprises the following steps:
s21, acquiring the preprocessed audio and video data through a collecting unit;
s22, identifying the acquired audio and video data through an identification unit;
s23, comparing and checking the audio and video data identified by the data identification unit through a comparison unit;
s24, deleting redundant audio and video data through a screening unit, so that each audio and video data only remains one, and repeating is completed;
s25, storing and backing up the audio and video data subjected to duplication removal through a storage unit, and preventing the audio and video data from being lost.
5. The artificial intelligence based three-dimensional animation live broadcast method according to claim 1, wherein the method comprises the following steps: the cloud live broadcast management platform in the S7 further comprises a live broadcast interaction management unit, and the live broadcast interaction management unit is used for realizing interaction between the user side and the watching fan.
6. The artificial intelligence based three-dimensional animation live broadcast method according to claim 5, wherein the method comprises the following steps: the live broadcast interaction management unit realizes the interaction between the user side and the watching fan as follows:
(1) Acquiring interaction information sent by a user side through the live broadcast interaction management unit;
(2) Obtaining an interactive information processing model through artificial intelligence learning training;
(3) The live broadcast interaction management unit inputs the acquired interaction information into the interaction information processing model and outputs the interaction information processed by the interaction information processing model;
(4) And the live broadcast interaction management unit feeds back corresponding audio and video data or three-dimensional animation or custom reply according to the processed interaction information.
7. The artificial intelligence based three-dimensional animation live broadcast method according to claim 6, wherein the method comprises the following steps: the live broadcast interaction management unit obtains interaction information sent by a user side, including but not limited to text comments and text barrages.
8. The artificial intelligence based three-dimensional animation live broadcast method according to claim 6, wherein the method comprises the following steps: the interactive information processing model is obtained through artificial intelligent learning training, and mainly comprises learning and training an artificial neural network model by utilizing a semantic system training sample and a semantic information rule sample so as to obtain the interactive information processing model for identifying semantics.
9. A three-dimensional animation live broadcast device based on artificial intelligence, which is characterized by comprising:
the acquisition module is used for acquiring audio and video data and preprocessing the audio and video data;
the data screening module is used for screening and de-duplicating the preprocessed audio and video data;
the information identification module is used for identifying video images in the audio and video data after the duplication removal, inputting the video images into the real-time motion capture model for capturing motion information and obtaining the motion information;
the animation generation module is used for acquiring a three-dimensional animation initial model, driving the three-dimensional animation initial model according to the three-dimensional animation control parameters and generating a three-dimensional animation;
the audio pairing module is used for identifying the audio data in the audio-video data and pairing the audio with the action in the three-dimensional animation so as to realize the dubbing of the three-dimensional animation;
the data processing module is used for inputting the action information into the three-dimensional animation binding logic module, calculating to obtain three-dimensional animation control parameters through a three-dimensional animation binding logic information calculation unit in the three-dimensional animation binding logic module, and transmitting the dubbed three-dimensional animation to the cloud live broadcast management platform for live broadcast.
10. The artificial intelligence based three-dimensional animation live broadcast device of claim 9, wherein: the data screening module comprises:
the receiving unit is used for acquiring the preprocessed audio and video data;
the identification unit is used for identifying the acquired audio and video data;
the comparison unit is used for comparing and checking the audio and video data identified by the data identification unit;
the screening unit is used for deleting redundant audio and video data, so that each audio and video data only remains one, and the duplicate removal is completed;
and the storage unit is used for storing and backing up the audio and video data after the duplication removal and preventing the audio and video data from being lost.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310294269.4A CN116320521A (en) | 2023-03-24 | 2023-03-24 | Three-dimensional animation live broadcast method and device based on artificial intelligence |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310294269.4A CN116320521A (en) | 2023-03-24 | 2023-03-24 | Three-dimensional animation live broadcast method and device based on artificial intelligence |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116320521A true CN116320521A (en) | 2023-06-23 |
Family
ID=86830349
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310294269.4A Pending CN116320521A (en) | 2023-03-24 | 2023-03-24 | Three-dimensional animation live broadcast method and device based on artificial intelligence |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116320521A (en) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108200446A (en) * | 2018-01-12 | 2018-06-22 | 北京蜜枝科技有限公司 | Multimedia interactive system and method on the line of virtual image |
WO2020017668A1 (en) * | 2018-07-16 | 2020-01-23 | 주식회사 이누씨 | Method and apparatus for generating avatar by using multi-view image matching |
CN111147873A (en) * | 2019-12-19 | 2020-05-12 | 武汉西山艺创文化有限公司 | Virtual image live broadcasting method and system based on 5G communication |
CN111340917A (en) * | 2020-02-11 | 2020-06-26 | 腾讯科技(深圳)有限公司 | Three-dimensional animation generation method and device, storage medium and computer equipment |
CN111355977A (en) * | 2020-04-16 | 2020-06-30 | 广东小天才科技有限公司 | Optimized storage method and device for live webcast video |
CN113132741A (en) * | 2021-03-03 | 2021-07-16 | 广州鑫泓设备设计有限公司 | Virtual live broadcast system and method |
CN113766268A (en) * | 2021-11-08 | 2021-12-07 | 阿里巴巴达摩院(杭州)科技有限公司 | Video processing method and device, electronic equipment and readable medium |
CN114373041A (en) * | 2021-12-15 | 2022-04-19 | 聚好看科技股份有限公司 | Three-dimensional reconstruction method and equipment |
-
2023
- 2023-03-24 CN CN202310294269.4A patent/CN116320521A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108200446A (en) * | 2018-01-12 | 2018-06-22 | 北京蜜枝科技有限公司 | Multimedia interactive system and method on the line of virtual image |
WO2020017668A1 (en) * | 2018-07-16 | 2020-01-23 | 주식회사 이누씨 | Method and apparatus for generating avatar by using multi-view image matching |
CN111147873A (en) * | 2019-12-19 | 2020-05-12 | 武汉西山艺创文化有限公司 | Virtual image live broadcasting method and system based on 5G communication |
CN111340917A (en) * | 2020-02-11 | 2020-06-26 | 腾讯科技(深圳)有限公司 | Three-dimensional animation generation method and device, storage medium and computer equipment |
CN111355977A (en) * | 2020-04-16 | 2020-06-30 | 广东小天才科技有限公司 | Optimized storage method and device for live webcast video |
CN113132741A (en) * | 2021-03-03 | 2021-07-16 | 广州鑫泓设备设计有限公司 | Virtual live broadcast system and method |
CN113766268A (en) * | 2021-11-08 | 2021-12-07 | 阿里巴巴达摩院(杭州)科技有限公司 | Video processing method and device, electronic equipment and readable medium |
CN114373041A (en) * | 2021-12-15 | 2022-04-19 | 聚好看科技股份有限公司 | Three-dimensional reconstruction method and equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107231566B (en) | Video transcoding method, device and system | |
US20120170805A1 (en) | Object detection in crowded scenes | |
EP3583777A1 (en) | A method and technical equipment for video processing | |
CN110852961A (en) | Real-time video denoising method and system based on convolutional neural network | |
CN114936605A (en) | Knowledge distillation-based neural network training method, device and storage medium | |
CN111968095A (en) | Product surface defect detection method, system, device and medium | |
WO2023082453A1 (en) | Image processing method and device | |
CN113570689B (en) | Portrait cartoon method, device, medium and computing equipment | |
Balchandani et al. | A deep learning framework for smart street cleaning | |
CN108985221A (en) | Video clip detection method, device, equipment and storage medium | |
CN109086737B (en) | Convolutional neural network-based shipping cargo monitoring video identification method and system | |
CN111339357A (en) | Recommendation method and device based on live user behaviors | |
CN105141974B (en) | A kind of video clipping method and device | |
CN116320521A (en) | Three-dimensional animation live broadcast method and device based on artificial intelligence | |
CN116992396A (en) | Redundancy self-adaptive multi-mode robust fusion learning method and system | |
US20230186608A1 (en) | Method, device, and computer program product for video processing | |
CN111212322A (en) | Video compression method based on multi-video de-duplication splicing | |
EP4311202A1 (en) | End-edge-cloud coordination system and method based on digital retina, and device | |
CN113628121B (en) | Method and device for processing and training multimedia data | |
CN114299089A (en) | Image processing method, image processing device, electronic equipment and storage medium | |
CN114913516A (en) | Tobacco retail license identification method and system | |
CN111524077B (en) | Method and system for repairing image data | |
CN112949344B (en) | Characteristic autoregression method for anomaly detection | |
CN112200721A (en) | Image processing method, system, device and medium | |
CN112954261B (en) | Video conference network flow control method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |