CN113490027A - Short video production generation processing method and equipment and computer storage medium - Google Patents

Short video production generation processing method and equipment and computer storage medium Download PDF

Info

Publication number
CN113490027A
CN113490027A CN202110766814.6A CN202110766814A CN113490027A CN 113490027 A CN113490027 A CN 113490027A CN 202110766814 A CN202110766814 A CN 202110766814A CN 113490027 A CN113490027 A CN 113490027A
Authority
CN
China
Prior art keywords
video
sub
segment
short
short video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110766814.6A
Other languages
Chinese (zh)
Inventor
孔祥兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Yirong Xinke Technology Co ltd
Original Assignee
Wuhan Yirong Xinke Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Yirong Xinke Technology Co ltd filed Critical Wuhan Yirong Xinke Technology Co ltd
Priority to CN202110766814.6A priority Critical patent/CN113490027A/en
Publication of CN113490027A publication Critical patent/CN113490027A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • H04N21/43072Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen of multiple content streams on the same device
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/2224Studio circuitry; Studio devices; Studio equipment related to virtual studio applications
    • H04N5/2226Determination of depth image, e.g. for foreground/background separation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects
    • H04N5/278Subtitling

Abstract

The invention discloses a method, a device and a computer storage medium for producing and processing short video, which are characterized in that by acquiring foreground and background images of each video frame image in each segment of sub-video in a short video material, detecting the contrast ratio of the foreground and the background images of each video frame image in each segment of sub-video in the short video material, analyzing the comprehensive image contrast ratio of each segment of sub-video in the short video material, calculating the contrast ratio of subtitles in each segment of sub-video in the short video material, simultaneously identifying and checking text information in each segment of sub-video in the short video material, acquiring the audio appearance time and audio disappearance time in each segment of sub-video in the short video material, calculating the appearance speed of subtitles in each segment of sub-video in the short video material, and carrying out corresponding subtitle contrast setting and corresponding subtitle appearance speed adjustment processing on the text information in each segment of sub-video in the short video material, therefore, the overall watching quality of the short video is improved, and the diversified watching requirements of people are met.

Description

Short video production generation processing method and equipment and computer storage medium
Technical Field
The invention relates to the technical field of short video production and generation, in particular to a short video production and generation processing method, equipment and a computer storage medium.
Background
With the continuous popularization of mobile terminals and the acceleration of network access speed, short videos gradually get the favor of users due to the characteristics of short videos, fast videos, large videos and the like. People like to generate and edit the short video in the short video production process, so that the content of the short video can be greatly enriched.
At present, the existing short video production generation processing technology has the following defects:
1. most of the existing short video production, generation and processing methods add subtitles to the short video manually, so that the intelligent degree of the short video production, generation and processing is reduced, and the problem that the played audio of the short video is not matched with the subtitles exists, so that the normal watching effect of the short video is influenced, and the watching experience and interest of people on the short video are reduced;
2. the existing short video production generation processing method sets the subtitle contrast according to the manual production experience, cannot perform intelligent setting according to the image contrast of the short video, has the problem that the contrast difference between the short video subtitle and the image is not obvious, causes the overall viewing quality of the short video to be reduced, and cannot meet the diversified viewing requirements of people;
in order to solve the above problems, a short video production generation processing method, a device, and a computer storage medium are now designed.
Disclosure of Invention
The invention aims to provide a short video production generation processing method, equipment and a computer storage medium, wherein a short video material is divided into sub-videos of all segments, foreground images and background images of all video frame images in all segments of sub-videos in the short video material are obtained and are subjected to image processing, the contrast of the foreground images and the contrast of the background images of all video frame images in all segments of sub-videos in the short video material are detected at the same time, the comprehensive image contrast of all segments of sub-videos in the short video material is analyzed, the contrast of subtitles in all segments of sub-videos in the short video material is calculated, text information in all segments of sub-videos in the short video material is identified and checked at the same time, the audio appearance time and audio disappearance time in all segments of sub-videos in the short video material are obtained, the subtitle appearance speed in all segments of sub-videos in the short video material is calculated, and the text information in all segments of sub-videos in the short video material is subjected to corresponding subtitle contrast setting and corresponding to the subtitle appearance time segments in the short video material The existing speed adjustment processing solves the problems existing in the background technology.
The purpose of the invention can be realized by the following technical scheme:
a short video production generation processing method comprises the following steps:
s1, short video material division: dividing the imported short video material into sub-videos of each segment according to a set division rule, and numbering the sub-videos in sequence;
s2, video frame image segmentation: respectively segmenting each video frame image in each segment of sub-video in the short video material by acquiring each video frame image in each segment of sub-video in the short video material to obtain a foreground image and a background image of each video frame image in each segment of sub-video in the short video material;
s3, video frame image processing: respectively processing foreground images and background images of all video frame images in all sections of sub-videos in a short video material by adopting an image processing technology to obtain foreground processed images and background processed images of all video frame images in all sections of sub-videos in the short video material;
s4, detecting image contrast: analyzing the comprehensive image contrast of each segment of sub-video in the short video material by respectively detecting the foreground processing image contrast and the background processing image contrast of each video frame image in each segment of sub-video in the short video material;
s5, subtitle contrast analysis: calculating the contrast ratio of the subtitles in each segment of sub-video in the short video material by extracting the standard contrast ratio of the video image and the subtitles in the short video stored in the storage database;
s6, identifying video and audio information: respectively identifying audio information in each segment of sub-video in the short video material through a voice identification technology, and converting to obtain text information in each segment of sub-video in the short video material;
s7, text information auditing: performing semantic association audit and structure association audit on text information in each segment of sub-video in the short video material respectively, and if the semantic association or the structure association of the text information in a certain segment of sub-video in the short video material is not accordant, performing appropriate correction on the text information in the segment of sub-video manually;
s8, acquiring audio appearance and disappearance time: calculating the time period of text information in each segment of sub-video in the short video material by acquiring the audio appearing time and audio disappearing time in each segment of sub-video in the short video material, and calculating the subtitle appearing speed in each segment of sub-video in the short video material;
s9, video subtitle setting processing: the text information in each segment of sub-video in the short video material is set according to the corresponding subtitle contrast, and the text information in each segment of sub-video in the short video material is subjected to corresponding subtitle occurrence speed adjustment processing within the corresponding occurrence time period.
Further, in the step S1, the short video material is divided into sub-videos according to the video voice pause sequence, and the sub-videos are numbered sequentially according to the video playing sequence, where the number of each sub-video in the short video material is 1,2,.
Further, the step S2 includes counting foreground images of each video frame image in each segment of sub-video in the short video material, and forming a foreground image set P of each video frame image in each segment of sub-video in the short video materialiA(pia1,pia2,...,piaj,...,piam),piajRepresenting the foreground image as the jth video frame image in the ith segment of sub-video in the short video material; meanwhile, the background images of all video frame images in all segments of sub-videos in the short video material are counted to form a background scene image set P of all video frame images in all segments of sub-videos in the short video materialiB(pib1,pib2,...,pibj,...,pibm),pibjAnd representing the background image of the jth video frame image in the ith sub-video in the short video material.
Further, the image processing technique adopted in step S3 includes performing geometric normalization processing on the foreground image and the background image of each video frame image in each segment of sub-video in the short video material, respectively, transforming into the foreground image and the background image of each video frame image in each segment of sub-video in a fixed standard form, performing optimization enhancement processing on the foreground image of each video frame image in each segment of sub-video after transformation, and performing blurring processing on the background image of each video frame image in each segment of sub-video after transformation.
Further, the step S4 includes the following steps:
s41, detecting the foreground processing image contrast of each video frame image in each segment of sub-video in the short video material to form a foreground processing image contrast set k of each video frame image in each segment of sub-video in the short video materialiA(kia1,kia2,...,kiaj,...,kiam),kiajRepresenting the contrast of a foreground processing image of a jth video frame image in an ith segment of sub-video in the short video material;
s42, detecting the contrast of the background processing image of each video frame image in each segment of sub-video in the short video material, and forming a background processing image contrast set k of each video frame image in each segment of sub-video in the short video materialiB(kib1,kib2,...,kibj,...,kibm),kibjRepresenting the contrast of a background processing image of a jth video frame image in an ith segment of sub-video in the short video material;
s43, calculating the comprehensive image contrast of each segment of sub-video in the short video material
Figure BDA0003152038330000041
The method comprises the steps of expressing the comprehensive image contrast of the ith segment of sub-video in a short video material, expressing m as the number of video frame images in the ith segment of sub-video in the short video material, and expressing alpha and beta as weighting proportion coefficients corresponding to foreground images and background images in the video frame images respectively, wherein alpha + beta is 1.
Further, the calculation formula of the subtitle contrast in each segment of sub-video in the short video material is as follows
Figure BDA0003152038330000042
k′iExpressed as the subtitle contrast, lambda, in the ith sub-video within the short video materialSign boardExpressed as the standard contrast ratio of video images to subtitles in short video.
Further, the speech recognition technique adopted in step S6 includes the following steps:
s61, filtering and framing preprocessing are carried out on the audio information in each segment of sub-video in the short video material, and redundant information is removed;
s62, extracting key information which influences voice recognition and characteristic information expressing voice meaning in the audio information in each section of sub-video;
s63, recognizing words by using the minimum unit according to the characteristic information in the audio information in each segment of sub-video, and sequentially recognizing the words according to the grammars corresponding to the audio information in each segment of sub-video;
and S64, connecting the recognized words in each segment of sub-video according to semantic analysis, and adjusting sentence composition according to the meaning of a sentence to obtain text information in each segment of sub-video in the short video material.
Further, the step S8 includes the following steps:
s81, obtaining the audio frequency appearance time in each segment of sub-video in the short video material, and forming the audio frequency appearance time set t (t) in each segment of sub-video in the short video material1,t2,...,ti,...,tn),tiRepresenting the audio occurrence time in the ith sub-video in the short video material;
s82, obtaining the audio disappearance time in each segment of sub-video in the short video material, and forming an audio disappearance time set t ' (t ' in each segment of sub-video in the short video material '1,t′2,...,t′i,...,t′n),t′iRepresenting the audio disappearance time in the ith sub-video in the short video material;
s83, extracting the word number of the text information in each segment of the sub-video in the short video material, and calculating the subtitle occurrence speed in each segment of the sub-video in the short video material
Figure BDA0003152038330000051
viExpressed as the speed of occurrence, x, of subtitles in the ith sub-video within the short video materialiRepresented as text messages in the ith sub-video within the short video materialThe number of words of the message.
An apparatus, comprising: the system comprises a processor, a memory and a network interface, wherein the memory and the network interface are connected with the processor; the network interface is connected with a nonvolatile memory in the server; when the processor runs, the processor calls the computer program from the nonvolatile memory through the network interface, and runs the computer program through the memory, so as to execute the short video production and generation processing method.
A computer storage medium is burned with a computer program, and when the computer program runs in a memory of a server, the short video production generation processing method is realized.
Has the advantages that:
(1) the invention provides a short video production generation processing method, equipment and computer storage medium, which divide a short video material into each segment of sub-video, acquire foreground images and background images of each video frame image in each segment of sub-video in the short video material, and perform image processing, thereby reducing the time and task amount required by image analysis, simultaneously detect the foreground and background processing image contrast of each video frame image in each segment of sub-video in the short video material, analyze the comprehensive image contrast of each segment of sub-video in the short video material, calculate the subtitle contrast in each segment of sub-video in the short video material, and perform corresponding subtitle contrast setting on text information in each segment of sub-video in the short video material, thereby realizing the intelligent setting function of the subtitle contrast of the short video, and ensuring that the difference between the short video subtitle and the image contrast is obvious, the overall watching quality of the short video is improved, and further the diversified watching requirements of people are met.
(2) According to the method, the accuracy and the reliability of text information identification in the short video are improved by identifying and checking the text information in each segment of the sub-video in the short video material, the audio appearing time and the audio disappearing time in each segment of the sub-video in the short video material are obtained, the subtitle appearing speed in each segment of the sub-video in the short video material is calculated, and the text information in each segment of the sub-video in the short video material is adjusted and processed according to the corresponding subtitle appearing speed in the corresponding appearing time period, so that the matching of the playing audio and the subtitle of the short video is ensured, the normal watching effect of the short video is prevented from being influenced, the intelligentization degree of the short video production and generation processing is improved, and the watching experience and interest of people on the short video are increased.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic diagram of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Referring to fig. 1, a method for generating and processing a short video includes the following steps:
s1, short video material division: the imported short video materials are divided into segments of sub-videos according to set division rules and are numbered in sequence.
In this embodiment, in step S1, the short video material is divided into segments of sub-videos according to the video voice pause sequence, and the segments of sub-videos in the short video material are numbered sequentially according to the video playing sequence, where the numbers of the segments of sub-videos in the short video material are 1,2, a.
S2, video frame image segmentation: the method comprises the steps of obtaining each video frame image in each segment of sub-video in the short video material, and respectively segmenting each video frame image in each segment of sub-video in the short video material to obtain a foreground image and a background image of each video frame image in each segment of sub-video in the short video material.
In this embodiment, the step S2 includes counting foreground images of each video frame image in each segment of sub-video in the short video material, and forming a foreground image set P of each video frame image in each segment of sub-video in the short video materialiA(pia1,pia2,...,piaj,...,piam),piajRepresenting the foreground image as the jth video frame image in the ith segment of sub-video in the short video material; meanwhile, the background images of all video frame images in all segments of sub-videos in the short video material are counted to form a background scene image set P of all video frame images in all segments of sub-videos in the short video materialiB(pib1,pib2,...,pibj,...,pibm),pibjAnd representing the background image of the jth video frame image in the ith sub-video in the short video material.
S3, video frame image processing: the foreground image and the background image of each video frame image in each segment of sub-video in the short video material are respectively processed by adopting an image processing technology, so that the foreground processing image and the background processing image of each video frame image in each segment of sub-video in the short video material are obtained.
In this embodiment, the image processing technique adopted in step S3 includes performing geometric normalization processing on the foreground image and the background image of each video frame image in each segment of sub-video in the short video material, respectively, transforming into the foreground image and the background image of each video frame image in each segment of sub-video in a fixed standard form, performing optimization enhancement processing on the foreground image of each video frame image in each segment of sub-video after transformation, and performing blurring processing on the background image of each video frame image in each segment of sub-video after transformation.
Specifically, the short video material is divided into the sub-videos, the foreground image and the background image of each video frame image in each sub-video in the short video material are obtained, and image processing is performed, so that the time and the task amount required by image analysis are reduced.
S4, detecting image contrast: and analyzing the comprehensive image contrast of each segment of sub-video in the short video material by respectively detecting the foreground processing image contrast and the background processing image contrast of each video frame image in each segment of sub-video in the short video material.
In this embodiment, the step S4 includes the following steps:
s41, detecting the foreground processing image contrast of each video frame image in each segment of sub-video in the short video material to form a foreground processing image contrast set k of each video frame image in each segment of sub-video in the short video materialiA(kia1,kia2,...,kiaj,...,kiam),kiajRepresenting the contrast of a foreground processing image of a jth video frame image in an ith segment of sub-video in the short video material;
s42, detecting the contrast of the background processing image of each video frame image in each segment of sub-video in the short video material, and forming a background processing image contrast set k of each video frame image in each segment of sub-video in the short video materialiB(kib1,kib2,...,kibj,...,kibm),kibjRepresenting the contrast of a background processing image of a jth video frame image in an ith segment of sub-video in the short video material;
s43, calculating the comprehensive image contrast of each segment of sub-video in the short video material
Figure BDA0003152038330000091
The method comprises the steps of expressing the comprehensive image contrast of the ith segment of sub-video in a short video material, expressing m as the number of video frame images in the ith segment of sub-video in the short video material, and expressing alpha and beta as weighting proportion coefficients corresponding to foreground images and background images in the video frame images respectively, wherein alpha + beta is 1.
S5, subtitle contrast analysis: and calculating the contrast ratio of the subtitles in each segment of sub-video in the short video material by extracting the standard contrast ratio of the video image and the subtitles in the short video stored in the storage database.
In this embodiment, the calculation formula of the subtitle contrast in each segment of sub-video in the short video material is
Figure BDA0003152038330000092
k′iExpressed as the subtitle contrast, lambda, in the ith sub-video within the short video materialSign boardExpressed as the standard contrast ratio of video images to subtitles in short video.
Specifically, the method processes the image contrast by detecting the foreground and the background of each video frame image in each segment of the sub-video in the short video material, analyzes the comprehensive image contrast of each segment of the sub-video in the short video material, and calculates the subtitle contrast in each segment of the sub-video in the short video material, so that the difference between the short video subtitle and the image contrast is obvious enough, the overall watching quality of the short video is improved, the diversified watching requirements of people are further met, and a reliable reference basis is provided for the subtitle contrast in each segment of the sub-video in the later period.
S6, identifying video and audio information: and respectively identifying the audio information in each segment of sub-video in the short video material through a voice identification technology, and converting to obtain the text information in each segment of sub-video in the short video material.
In this embodiment, the speech recognition technique adopted in step S6 includes the following steps:
s61, filtering and framing preprocessing are carried out on the audio information in each segment of sub-video in the short video material, and redundant information is removed;
s62, extracting key information which influences voice recognition and characteristic information expressing voice meaning in the audio information in each section of sub-video;
s63, recognizing words by using the minimum unit according to the characteristic information in the audio information in each segment of sub-video, and sequentially recognizing the words according to the grammars corresponding to the audio information in each segment of sub-video;
and S64, connecting the recognized words in each segment of sub-video according to semantic analysis, and adjusting sentence composition according to the meaning of a sentence to obtain text information in each segment of sub-video in the short video material.
S7, text information auditing: by respectively performing semantic association audit and structure association audit on the text information in each segment of sub-video in the short video material, if the semantic association or structure association of the text information in a certain segment of sub-video in the short video material is not met, the text information in the segment of sub-video is appropriately corrected.
In this embodiment, if the semantic association and the structural association of the text information in a certain segment of sub-video in the short video material are both satisfied, it is indicated that the review of the text information in the segment of sub-video is passed, and if the semantic association or the structural association of the text information in a certain segment of sub-video in the short video material are not satisfied, it is indicated that the review of the text information in the segment of sub-video is not passed, the text information in the segment of sub-video is appropriately corrected manually.
Specifically, the accuracy and reliability of text information identification in the short video are improved by identifying and auditing the text information in each segment of sub-video in the short video material.
S8, acquiring audio appearance and disappearance time: the method comprises the steps of obtaining the audio appearing time and audio disappearing time in each section of sub-video in the short video material, counting the time period of the text information appearing in each section of sub-video in the short video material, and calculating the subtitle appearing speed in each section of sub-video in the short video material.
In this embodiment, the step S8 includes the following steps:
s81, obtaining the audio frequency appearance time in each segment of sub-video in the short video material, and forming the audio frequency appearance time set t (t) in each segment of sub-video in the short video material1,t2,...,ti,...,tn),tiRepresenting the audio occurrence time in the ith sub-video in the short video material;
s82, obtaining the audio disappearance time in each segment of sub-video in the short video material, and forming an audio disappearance time set t ' (t ' in each segment of sub-video in the short video material '1,t′2,...,t′i,...,t′n),t′iRepresenting the audio disappearance time in the ith sub-video in the short video material;
s83, extracting the word number of the text information in each segment of the sub-video in the short video material, and calculating the subtitle occurrence speed in each segment of the sub-video in the short video material
Figure BDA0003152038330000111
viExpressed as the speed of occurrence, x, of subtitles in the ith sub-video within the short video materialiRepresented as the number of words of text information in the ith sub-video within the short video material.
Specifically, the method and the device calculate the caption appearance speed in each segment of sub-video in the short video material by acquiring the audio appearance time and audio disappearance time in each segment of sub-video in the short video material, thereby ensuring that the playing audio of the short video can be matched with the caption, avoiding influencing the normal watching effect of the short video, and providing reliable reference basis for the later regulation and processing of the caption in each sub-video.
S9, video subtitle setting processing: the text information in each segment of sub-video in the short video material is set according to the corresponding subtitle contrast, and the text information in each segment of sub-video in the short video material is subjected to corresponding subtitle occurrence speed adjustment processing within the corresponding occurrence time period.
Specifically, the method and the device realize the intelligent setting function of the short video subtitle contrast by setting the corresponding subtitle contrast of the text information in each segment of the sub-video in the short video material, and adjust the appearance speed of the corresponding subtitle in the corresponding appearance time period of the text information in each segment of the sub-video in the short video material, thereby improving the intelligent degree of the short video production and generation processing and increasing the watching experience and interest of people on the short video.
An apparatus, comprising: the system comprises a processor, a memory and a network interface, wherein the memory and the network interface are connected with the processor; the network interface is connected with a nonvolatile memory in the server; when the processor runs, the processor calls the computer program from the nonvolatile memory through the network interface, and runs the computer program through the memory, so as to execute the short video production and generation processing method.
A computer storage medium is burned with a computer program, and when the computer program runs in a memory of a server, the short video production generation processing method is realized.
The foregoing is merely exemplary and illustrative of the principles of the present invention and various modifications, additions and substitutions of the specific embodiments described herein may be made by those skilled in the art without departing from the principles of the present invention or exceeding the scope of the claims set forth herein.

Claims (10)

1. A short video production generation processing method is characterized by comprising the following steps: the method comprises the following steps:
s1, short video material division: dividing the imported short video material into sub-videos of each segment according to a set division rule, and numbering the sub-videos in sequence;
s2, video frame image segmentation: respectively segmenting each video frame image in each segment of sub-video in the short video material by acquiring each video frame image in each segment of sub-video in the short video material to obtain a foreground image and a background image of each video frame image in each segment of sub-video in the short video material;
s3, video frame image processing: respectively processing foreground images and background images of all video frame images in all sections of sub-videos in a short video material by adopting an image processing technology to obtain foreground processed images and background processed images of all video frame images in all sections of sub-videos in the short video material;
s4, detecting image contrast: analyzing the comprehensive image contrast of each segment of sub-video in the short video material by respectively detecting the foreground processing image contrast and the background processing image contrast of each video frame image in each segment of sub-video in the short video material;
s5, subtitle contrast analysis: calculating the contrast ratio of the subtitles in each segment of sub-video in the short video material by extracting the standard contrast ratio of the video image and the subtitles in the short video stored in the storage database;
s6, identifying video and audio information: respectively identifying audio information in each segment of sub-video in the short video material through a voice identification technology, and converting to obtain text information in each segment of sub-video in the short video material;
s7, text information auditing: performing semantic association audit and structure association audit on text information in each segment of sub-video in the short video material respectively, and if the semantic association or the structure association of the text information in a certain segment of sub-video in the short video material is not accordant, performing appropriate correction on the text information in the segment of sub-video manually;
s8, acquiring audio appearance and disappearance time: calculating the time period of text information in each segment of sub-video in the short video material by acquiring the audio appearing time and audio disappearing time in each segment of sub-video in the short video material, and calculating the subtitle appearing speed in each segment of sub-video in the short video material;
s9, video subtitle setting processing: the text information in each segment of sub-video in the short video material is set according to the corresponding subtitle contrast, and the text information in each segment of sub-video in the short video material is subjected to corresponding subtitle occurrence speed adjustment processing within the corresponding occurrence time period.
2. The short video production generation processing method according to claim 1, wherein: in the step S1, the short video material is divided into sub-videos according to the video voice pause sequence, and the sub-videos are sequentially numbered according to the video playing sequence, where the number of each sub-video in the short video material is 1,2,.
3. The short video production generation processing method according to claim 1, wherein: the step S2 includes counting foreground images of each video frame image in each segment of sub-video in the short video material, and forming a foreground image set P of each video frame image in each segment of sub-video in the short video materialiA(pia1,pia2,...,piaj,...,piam),piajRepresenting the foreground image as the jth video frame image in the ith segment of sub-video in the short video material; meanwhile, the background images of all video frame images in all segments of sub-videos in the short video material are counted to form a background scene image set P of all video frame images in all segments of sub-videos in the short video materialiB(pib1,pib2,...,pibj,...,pibm),pibjAnd representing the background image of the jth video frame image in the ith sub-video in the short video material.
4. The short video production generation processing method according to claim 1, wherein: the image processing technique adopted in step S3 includes performing geometric normalization processing on the foreground image and the background image of each video frame image in each segment of sub-video in the short video material, respectively, to transform the foreground image and the background image into the foreground image and the background image of each video frame image in each segment of sub-video in a fixed standard form, performing optimization enhancement processing on the foreground image of each video frame image in each segment of sub-video after transformation, and performing blurring processing on the background image of each video frame image in each segment of sub-video after transformation.
5. The short video production generation processing method according to claim 1, wherein: the step S4 includes the following steps:
s41, detecting the foreground processing image contrast of each video frame image in each segment of sub-video in the short video material to form a foreground processing image contrast set k of each video frame image in each segment of sub-video in the short video materialiA(kia1,kia2,...,kiaj,...,kiam),kiajRepresenting the contrast of a foreground processing image of a jth video frame image in an ith segment of sub-video in the short video material;
s42, and comparing each of the short video materialsDetecting the contrast of the background processing image of each video frame image in the segment of the sub-video to form a background processing image contrast set k of each video frame image in each segment of the sub-video in the short video materialiB(kib1,kib2,...,kibj,...,kibm),kibjRepresenting the contrast of a background processing image of a jth video frame image in an ith segment of sub-video in the short video material;
s43, calculating the comprehensive image contrast of each segment of sub-video in the short video material
Figure FDA0003152038320000031
Figure FDA0003152038320000032
The method comprises the steps of expressing the comprehensive image contrast of the ith segment of sub-video in a short video material, expressing m as the number of video frame images in the ith segment of sub-video in the short video material, and expressing alpha and beta as weighting proportion coefficients corresponding to foreground images and background images in the video frame images respectively, wherein alpha + beta is 1.
6. The short video production generation processing method according to claim 1, wherein: the calculation formula of the subtitle contrast in each segment of sub-video in the short video material is
Figure FDA0003152038320000033
k′iExpressed as the subtitle contrast, lambda, in the ith sub-video within the short video materialSign boardExpressed as the standard contrast ratio of video images to subtitles in short video.
7. The short video production generation processing method according to claim 1, wherein: the speech recognition technique employed in step S6 includes the steps of:
s61, filtering and framing preprocessing are carried out on the audio information in each segment of sub-video in the short video material, and redundant information is removed;
s62, extracting key information which influences voice recognition and characteristic information expressing voice meaning in the audio information in each section of sub-video;
s63, recognizing words by using the minimum unit according to the characteristic information in the audio information in each segment of sub-video, and sequentially recognizing the words according to the grammars corresponding to the audio information in each segment of sub-video;
and S64, connecting the recognized words in each segment of sub-video according to semantic analysis, and adjusting sentence composition according to the meaning of a sentence to obtain text information in each segment of sub-video in the short video material.
8. The short video production generation processing method according to claim 1, wherein: the step S8 includes the following steps:
s81, obtaining the audio frequency appearance time in each segment of sub-video in the short video material, and forming the audio frequency appearance time set t (t) in each segment of sub-video in the short video material1,t2,...,ti,...,tn),tiRepresenting the audio occurrence time in the ith sub-video in the short video material;
s82, obtaining the audio disappearance time in each segment of sub-video in the short video material, and forming an audio disappearance time set t ' (t ' in each segment of sub-video in the short video material '1,t′2,...,t′i,...,t′n),t′iRepresenting the audio disappearance time in the ith sub-video in the short video material;
s83, extracting the word number of the text information in each segment of the sub-video in the short video material, and calculating the subtitle occurrence speed in each segment of the sub-video in the short video material
Figure FDA0003152038320000041
viExpressed as the speed of occurrence, x, of subtitles in the ith sub-video within the short video materialiRepresented as the number of words of text information in the ith sub-video within the short video material.
9. An apparatus, characterized by: the method comprises the following steps: the system comprises a processor, a memory and a network interface, wherein the memory and the network interface are connected with the processor; the network interface is connected with a nonvolatile memory in the server; the processor retrieves a computer program from the non-volatile memory through the network interface when running, and runs the computer program through the memory to execute a short video production generation processing method according to any one of claims 1 to 8.
10. A computer storage medium, characterized in that: the computer storage medium is burned with a computer program, and the computer program realizes a short video production generation processing method according to any one of claims 1 to 8 when running in a memory of a server.
CN202110766814.6A 2021-07-07 2021-07-07 Short video production generation processing method and equipment and computer storage medium Pending CN113490027A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110766814.6A CN113490027A (en) 2021-07-07 2021-07-07 Short video production generation processing method and equipment and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110766814.6A CN113490027A (en) 2021-07-07 2021-07-07 Short video production generation processing method and equipment and computer storage medium

Publications (1)

Publication Number Publication Date
CN113490027A true CN113490027A (en) 2021-10-08

Family

ID=77941740

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110766814.6A Pending CN113490027A (en) 2021-07-07 2021-07-07 Short video production generation processing method and equipment and computer storage medium

Country Status (1)

Country Link
CN (1) CN113490027A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114466137A (en) * 2022-01-11 2022-05-10 隗建华 Remote real-time synchronous monitoring system for short video shooting

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1879403A (en) * 2003-11-10 2006-12-13 皇家飞利浦电子股份有限公司 Adaptation of close-captioned text based on surrounding video content
CN101345820A (en) * 2008-08-01 2009-01-14 中兴通讯股份有限公司 Image brightness reinforcing method
CN102999901A (en) * 2012-10-17 2013-03-27 中国科学院计算技术研究所 Method and system for processing split online video on the basis of depth sensor
US20130129206A1 (en) * 2011-05-31 2013-05-23 John W. Worthington Methods and Apparatus for Improved Display of Foreground Elements
US20140050355A1 (en) * 2012-08-20 2014-02-20 Behavioral Recognition Systems, Inc. Method and system for detecting sea-surface oil
CN107124561A (en) * 2016-12-27 2017-09-01 福建新大陆自动识别技术有限公司 A kind of bar code image exposure adjustment system and method based on CMOS
CN111986656A (en) * 2020-08-31 2020-11-24 上海松鼠课堂人工智能科技有限公司 Teaching video automatic caption processing method and system

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1879403A (en) * 2003-11-10 2006-12-13 皇家飞利浦电子股份有限公司 Adaptation of close-captioned text based on surrounding video content
CN101345820A (en) * 2008-08-01 2009-01-14 中兴通讯股份有限公司 Image brightness reinforcing method
US20130129206A1 (en) * 2011-05-31 2013-05-23 John W. Worthington Methods and Apparatus for Improved Display of Foreground Elements
US20140050355A1 (en) * 2012-08-20 2014-02-20 Behavioral Recognition Systems, Inc. Method and system for detecting sea-surface oil
CN102999901A (en) * 2012-10-17 2013-03-27 中国科学院计算技术研究所 Method and system for processing split online video on the basis of depth sensor
CN107124561A (en) * 2016-12-27 2017-09-01 福建新大陆自动识别技术有限公司 A kind of bar code image exposure adjustment system and method based on CMOS
CN111986656A (en) * 2020-08-31 2020-11-24 上海松鼠课堂人工智能科技有限公司 Teaching video automatic caption processing method and system

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114466137A (en) * 2022-01-11 2022-05-10 隗建华 Remote real-time synchronous monitoring system for short video shooting

Similar Documents

Publication Publication Date Title
CN108766440B (en) Speaker separation model training method, two-speaker separation method and related equipment
Chung et al. Spot the conversation: speaker diarisation in the wild
Makino et al. Recurrent neural network transducer for audio-visual speech recognition
Afouras et al. LRS3-TED: a large-scale dataset for visual speech recognition
US10497382B2 (en) Associating faces with voices for speaker diarization within videos
CN108648746A (en) A kind of open field video natural language description generation method based on multi-modal Fusion Features
Hong et al. Dynamic captioning: video accessibility enhancement for hearing impairment
CN109257547B (en) Chinese online audio/video subtitle generating method
CN110222719B (en) Figure identification method and system based on multi-frame audio and video fusion network
CN111488487B (en) Advertisement detection method and detection system for all-media data
CN108615532B (en) Classification method and device applied to sound scene
Chuang et al. Improved lite audio-visual speech enhancement
CN113242361B (en) Video processing method and device and computer readable storage medium
CN113490027A (en) Short video production generation processing method and equipment and computer storage medium
CN112466306B (en) Conference summary generation method, device, computer equipment and storage medium
CN115439614B (en) Virtual image generation method and device, electronic equipment and storage medium
CN114495946A (en) Voiceprint clustering method, electronic device and storage medium
CN111128190B (en) Expression matching method and system
CN114283493A (en) Artificial intelligence-based identification system
JP7137825B2 (en) Video information provision system
CN113329190B (en) Animation design video production analysis management method, equipment, system and computer storage medium
Chaloupka A prototype of audio-visual broadcast transcription system
Bendris et al. Talking faces indexing in TV-content
Hukkeri et al. Erratic navigation in lecture videos using hybrid text based index point generation
CN116708055B (en) Intelligent multimedia audiovisual image processing method, system and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20211008