CN111355977A - Optimized storage method and device for live webcast video - Google Patents

Optimized storage method and device for live webcast video Download PDF

Info

Publication number
CN111355977A
CN111355977A CN202010301586.0A CN202010301586A CN111355977A CN 111355977 A CN111355977 A CN 111355977A CN 202010301586 A CN202010301586 A CN 202010301586A CN 111355977 A CN111355977 A CN 111355977A
Authority
CN
China
Prior art keywords
video
audio
file
repeated
frame file
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010301586.0A
Other languages
Chinese (zh)
Other versions
CN111355977B (en
Inventor
曾乙峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TCL China Star Optoelectronics Technology Co Ltd
Original Assignee
Shenzhen China Star Optoelectronics Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen China Star Optoelectronics Technology Co Ltd filed Critical Shenzhen China Star Optoelectronics Technology Co Ltd
Priority to CN202010301586.0A priority Critical patent/CN111355977B/en
Publication of CN111355977A publication Critical patent/CN111355977A/en
Application granted granted Critical
Publication of CN111355977B publication Critical patent/CN111355977B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/21Server components or server architectures
    • H04N21/218Source of audio or video content, e.g. local disk arrays
    • H04N21/2187Live feed
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/4302Content synchronisation processes, e.g. decoder synchronisation
    • H04N21/4307Synchronising the rendering of multiple content streams or additional data on devices, e.g. synchronisation of audio on a mobile phone with the video output on the TV screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8547Content authoring involving timestamps for synchronizing content

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The embodiment of the application discloses an optimized storage method and device for a live network video. According to the technical scheme, the video source file of the live broadcast is obtained, the corresponding video picture frame file and the corresponding audio frame file are analyzed from the video source file, the repeated picture frames in the video picture frame file are determined one by one according to the set time interval based on the timestamp information, the repeated picture frames are compressed, the repeated position audio frames corresponding to the repeated picture frames in the audio frame file are determined, and the reset audio frames are compressed or deleted. And merging and storing the processed video picture frame file and the audio frame file based on the timestamp information. By adopting the technical means, a large number of repeated pictures in the live network can be compressed, audio and video synchronization is further realized by compressing or deleting corresponding audio frames, the live network video is optimized and stored, unnecessary video data amount is reduced, and the video watching experience of a user is optimized.

Description

Optimized storage method and device for live webcast video
Technical Field
The embodiment of the application relates to the technical field of video processing, in particular to an optimized storage method and device for a live network video.
Background
Live webcasting refers to a media propagation mode in which independent signal acquisition equipment (audio and video) is installed and guided into a broadcast guide terminal (broadcast guide equipment or platform), and then uploaded to a server through a network and released to a live webcasting platform for people to watch. At present, as the network live broadcast is gradually popularized and popular, the network live broadcast video is also visible everywhere. In order to facilitate viewing at any time, some high quality live video needs to be saved.
However, most webcast videos are unprocessed during storage, and the amount of video data is large due to the fact that a large number of repeated pictures may exist in the live broadcasting process. Moreover, certain trouble is brought when watching recorded and broadcasted videos.
Disclosure of Invention
The embodiment of the application provides an optimized storage method and device for a live network video, an electronic device and a storage medium, which can optimize the storage of the live network video, reduce the video data volume and optimize the video watching experience of a user.
In a first aspect, an embodiment of the present application provides an optimized saving method for a live webcast video, including:
acquiring a video source file of network live broadcast, and analyzing a corresponding video picture frame file and an audio frame file from the video source file, wherein the video picture frame file and the audio frame file contain corresponding timestamp information;
determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the repeated position audio frames;
and merging and storing the processed video picture frame file and the processed audio frame file based on the timestamp information.
Further, determining repeated picture frames in the video picture frame file one by one according to a set time period based on the timestamp information, including:
extracting image characteristic information of each video picture frame under corresponding timestamp information based on a set time period;
and determining the video picture frame with the similarity reaching a set threshold value as a repeated picture frame based on the image characteristic information comparison.
Further, the set period is 1S.
Further, the compressing or deleting process of the reset audio frame includes:
and determining whether the audio frame at the repeated position is effective audio based on a preset audio waveform analysis model, if so, compressing the audio frame at the repeated position, and if not, deleting the audio frame at the repeated position.
Further, before determining whether the repeated position audio frame is valid audio based on a preset audio waveform analysis model, the method further includes:
and inputting the voice waveform characteristics as training samples into an audio waveform analysis model to perform model training to construct the audio waveform analysis model.
Further, determining repeated picture frames in the video picture frame file one by one according to a set time period based on the timestamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the reset audio frames, including:
processing the video picture frame file and the audio frame file based on FFmpeg;
correspondingly, the merging and storing the processed video picture frame file and the audio frame file based on the timestamp information includes:
and merging and storing the video picture frame file and the audio frame file based on FFmpeg.
Further, acquiring a video source file of the webcast, and analyzing a corresponding video picture frame file and an audio frame file from the video source file, further includes:
and respectively carrying out noise reduction treatment on the video picture frame file and the audio frame file, and eliminating picture background noise of the video picture frame file and environmental background noise of the audio frame file.
In a second aspect, an embodiment of the present application provides an optimized saving apparatus for webcast video, including:
the system comprises an acquisition module, a time stamp module and a processing module, wherein the acquisition module is used for acquiring a video source file of network live broadcast and analyzing a corresponding video picture frame file and an audio frame file from the video source file, and the video picture frame file and the audio frame file contain corresponding time stamp information;
the processing module is used for determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the reset audio frames;
and the merging module is used for merging and storing the processed video picture frame file and the audio frame file based on the timestamp information.
In a third aspect, an embodiment of the present application provides an electronic device, including:
a memory and one or more processors;
the memory for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement the optimized save method for webcast video according to the first aspect.
In a fourth aspect, embodiments of the present application provide a storage medium containing computer-executable instructions, which when executed by a computer processor, are configured to perform the method for optimized preservation of webcast video according to the first aspect.
According to the embodiment of the application, a video source file of live webcasting is obtained, a corresponding video picture frame file and an audio frame file are analyzed from the video source file, repeated picture frames in the video picture frame file are determined one by one according to a set time period based on timestamp information, the repeated picture frames are compressed, repeated position audio frames corresponding to the repeated picture frames in the audio frame file are determined, and the repeated position audio frames are compressed or deleted. And merging and storing the processed video picture frame file and the audio frame file based on the timestamp information. By adopting the technical means, a large number of repeated pictures in the live network can be compressed, audio and video synchronization is further realized by compressing or deleting corresponding audio frames, the live network video is optimized and stored, unnecessary video data amount is reduced, and the video watching experience of a user is optimized.
Drawings
Fig. 1 is a flowchart of an optimized saving method for webcast video according to an embodiment of the present application;
fig. 2 is a flowchart of determining a repeated picture frame according to a first embodiment of the present application;
fig. 3 is a schematic structural diagram of an optimized saving device for webcast video according to a second embodiment of the present application;
fig. 4 is a schematic structural diagram of an electronic device according to a third embodiment of the present application.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, specific embodiments of the present application will be described in detail with reference to the accompanying drawings. It is to be understood that the specific embodiments described herein are merely illustrative of the application and are not limiting of the application. It should be further noted that, for the convenience of description, only some but not all of the relevant portions of the present application are shown in the drawings. Before discussing exemplary embodiments in more detail, it should be noted that some exemplary embodiments are described as processes or methods depicted as flowcharts. Although a flowchart may describe the operations (or steps) as a sequential process, many of the operations can be performed in parallel, concurrently or simultaneously. In addition, the order of the operations may be re-arranged. The process may be terminated when its operations are completed, but may have additional steps not included in the figure. The processes may correspond to methods, functions, procedures, subroutines, and the like.
The application provides an optimized storage method of a live webcast video, which aims to compress highly repeated video picture frames in the live webcast video and delete partial repeated video picture frames. And further compressing or deleting the audio frames at the repeated positions to realize audio synchronization of the live network video. Therefore, when a user watches the video of the live network video, the user can avoid watching repeated and meaningless video pictures, and the video watching experience of the user is optimized. The traditional network live video storage method is used for storing the network live video. The video source file is not optimized well, but is directly saved. Therefore, when a user watches the live network video, the user needs to spend some time on watching repeated video pictures, the whole video watching process is long, and the watching experience of the user is relatively poor. Based on this, the optimized storage method of the live webcast video is provided to solve the technical problems that the existing live webcast video is long in video recording and has a large number of repeated video pictures.
The first embodiment is as follows:
fig. 1 is a flowchart of an optimized saving method for webcast video provided in an embodiment of the present application, where the optimized saving method for webcast video provided in this embodiment may be executed by an optimized saving device for webcast video, the optimized saving device for webcast video may be implemented in a software and/or hardware manner, and the optimized saving device for webcast video may be formed by two or more physical entities or may be formed by one physical entity. Generally speaking, the optimized storage device for the webcast video can be an intelligent terminal device such as a computer, a mobile phone, a tablet and the like.
The following description will be given by taking an optimized saving device of a webcast video as an example of a main body for executing an optimized saving method of a webcast video. Referring to fig. 1, the optimized saving method for the webcast video specifically includes:
s110, acquiring a video source file of live webcasting, and analyzing a corresponding video picture frame file and an audio frame file from the video source file, wherein the video picture frame file and the audio frame file contain corresponding timestamp information.
Specifically, when the webcast video processing is performed, the corresponding video source file including the corresponding video picture data and audio data is obtained, and the corresponding video picture frame file and audio frame file are obtained by analyzing the video source file. It will be appreciated that each frame of video picture contains corresponding time stamp information, and likewise, each frame of audio picture also contains corresponding time stamp information. And the time stamps of the video picture frames and the audio frames are in one-to-one correspondence, and the corresponding video frame rate and the audio frame rate can be set according to the requirement for realization.
Further, a video picture frame file and an audio frame file are obtained based on the analysis, and the video picture frame file and the audio frame file are subjected to noise reduction processing respectively to eliminate picture background noise of the video picture frame file and environmental background noise of the audio frame file. The method comprises the steps of obtaining a brightness difference histogram of a current video picture frame by utilizing a denoising result of a previous video picture frame and a gradient module value histogram of the current video picture frame corresponding to denoising processing of a video picture frame file, and estimating the noise level of the current video picture frame according to the brightness difference histogram; calculating the airspace distance of any two pixel points in the current video picture frame, acquiring the airspace similarity of the any two pixel points, and denoising the current video picture frame according to the airspace similarity; calculating the pixel time domain distance between any pixel point in the current video picture frame and a pixel point at the corresponding position of the previous frame of the de-noised video picture frame, calculating the corresponding time domain similarity, and performing three-dimensional recursive de-noising on the video picture frame according to the obtained time domain similarity, the de-noising result of the space domain similarity and the de-noising result of the previous frame. By using the spatial and temporal correlation of the pixels to carry out three-dimensional recursive denoising, stronger complex noise can be removed, the details of video picture frames can be kept, and the stability of the denoising effect can be ensured.
Furthermore, when the noise reduction processing of the video picture frame is carried out, for some use scenes with low image background importance, the image background noise can be eliminated. For example, in a scene of live webcasting teaching, image backgrounds other than people are usually relatively minor, so that object recognition may be performed to determine irrelevant image backgrounds in video picture frames, eliminate the irrelevant image backgrounds in the video picture frames, and reduce the storage amount of video picture frame data by displaying the backgrounds in a blurring manner. For example, for a teaching video obtained by live online teaching, it is necessary to store the image contents of teachers and blackboards, and other background data, and in order to reduce the storage amount of data, the background information can be eliminated through blurring processing. In addition, the image background data of the first frame of video picture frame can be identified and detected, and in the subsequent video picture frame, the image background data identified by the first frame is mapped to all the subsequent video picture frames for rendering and displaying, so that all the subsequent image background data can be deleted, and only the image background data of the first frame of video picture frame needs to be stored.
For the noise reduction processing of the audio frame file, whether the audio parameter information is matched with the target noise reduction parameter information is judged by acquiring the audio parameter information of the audio frame and judging whether the audio parameter information is effective; if not, resampling the audio frame until the audio parameter information of the audio data is matched with the target noise reduction parameter information; and partitioning the audio frame according to a preset time interval, and outputting the partitioned audio frame after performing noise reduction treatment, so that the environmental noise in audio and video live broadcast can be effectively removed. And performing noise reduction processing according to the audio parameter information of the effective audio, which is stored in advance. It should be noted that, in the prior art, there are many noise reduction processing manners for video picture frames and audio frames, and a specific audio/video noise reduction processing means is not fixedly limited in the embodiments of the present application, which is not described herein repeatedly.
S120, determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the repeated audio frames.
And (4) processing the video picture frame according to the video picture frame and the audio frame obtained by analysis, and eliminating the video picture frame which is repeatedly relapsed. The video picture frames with highly similar image characteristics are determined according to the image characteristic information of each video picture frame, and are defined as repeated picture frames. Referring to fig. 2, the process of determining the repeated picture frame includes:
s1201, extracting image characteristic information of each video picture frame under corresponding timestamp information based on a set time period;
s1202, determining that the video picture frame with the similarity reaching the set threshold is a repeated picture frame based on the image characteristic information comparison.
Specifically, grouping processing of video picture frames is performed corresponding to timestamp information in a video picture frame file, wherein grouping is performed according to a set time period, and the set time period in the embodiment of the present application is 1S. Namely, according to the timestamp information, every 1S video picture frame is divided into a group to compare the image characteristic information. And when the similarity of the image characteristic information reaches a set similarity threshold (such as 90 percent) through comparison, the two video picture frames are regarded as repeated picture frames. Further, when comparing the image characteristic information, the gray information of the pixel points corresponding to each video picture frame may be compared, and if the similarity of the gray information reaches a set threshold, the two are repeated picture frames. In practical application, according to different image comparison algorithms, the repeated picture frame can be determined in an image characteristic information comparison mode such as image contour comparison. There are many technical means for determining similar images based on image feature information comparison, and details are not repeated herein. And finally determining a plurality of video picture frames with the similarity reaching a set threshold value in a mode of cross comparison of all the video picture frames during comparison. After the determination of the repeated picture frames of one group is completed, further determining whether the repeated picture frames exist in the video picture frames of the next group, and so on until the determination of all the repeated picture frames of the whole video picture frame file is completed.
Further, after determining the repeated picture frames, one or more video picture frames are selected from each grouped repeated picture frame, and the rest repeated picture frames are deleted. Moreover, when selecting the reserved video picture frames, a random selection mode can be adopted, and the video picture frames ranked in front or behind can be selected according to the timestamp information. If several video picture frames are reserved, a mode of frame extraction and reservation at intervals can be adopted. In addition, in practical application, in order to ensure that after repeated video picture frames are deleted, the video pictures still maintain high continuity and stability. Then, the video picture frames with the high similarity to the front and rear video picture frames in picture brightness and contrast need to be selected and retained from the repeated picture frames, so as to ensure smooth transition between the video picture frames and ensure the stability of video playing. After the partial repeated video picture frames are deleted, the compression processing of the repeated video pictures can be completed.
Furthermore, after the repeated picture frames are compressed, the audio frames corresponding to the repeated picture frames need to be processed, so that the audio and video synchronization is ensured, and the condition that the audio and video are not synchronized is avoided. And determining the audio frame at the repetition position corresponding to the repeated picture frame according to the time stamp information corresponding to the audio frame of each group by referring to the mode of processing the video picture frame by the group. It should be noted that the repeated position audio frame may be valid audio content containing a human voice or a target sound, or may be invalid audio content not containing a human voice or a target sound. It is understood that if the repeated positional audio frame contains human voice or target sound data, it is necessary to retain the part of the sound data. Otherwise, the content of the audio frame can be deleted.
Specifically, whether the audio frame at the repetition position is an effective audio is determined based on a preset audio waveform analysis model, if so, the audio frame at the repetition position is compressed, and if not, the audio frame at the repetition position is deleted.
Taking the human voice audio frame processing as an example, an audio waveform analysis model corresponding to human voice is first constructed to detect and identify the human voice. When the human voice recognition model is constructed, the human voice waveform characteristics are used as training samples, and the audio waveform analysis model is input to perform model training to construct the audio waveform analysis model. The voice recognition model can be constructed based on a convolutional neural network, and the voice frequency waveform is input into the preset voice recognition model by converting the voice frequency frame into the corresponding voice frequency waveform. And the voice recognition model carries out recognition analysis based on the voice frequency waveform, and if the voice frequency waveform is judged to be highly similar to the voice frequency waveform (if the similarity reaches 90%), the voice frequency waveform is judged to be the voice frequency waveform, and the corresponding voice frequency frame is effective voice frequency. Otherwise, the corresponding audio frame is an invalid audio. Further, after the audio frame at the repeated position is determined to be the effective audio, since the partial repeated video picture frame is deleted before, the audio frame needs to be compressed correspondingly to ensure the audio and video synchronization. After the corresponding audio frame at the repeated position is determined to be the wireless audio, the audio frame at the repeated position is directly deleted without storing the audio frame.
And S130, merging and storing the processed video picture frame file and the processed audio frame file based on the timestamp information.
And after the video picture frame and the audio frame are processed, combining and storing the video picture frame and the audio frame according to the corresponding timestamp information. And adopting FFmpeg to merge video picture frames and audio frames to generate a video source file which is correspondingly compressed. In the embodiment of the application, the pulling and the analysis of the video file, the noise reduction and the compression processing of the video picture frame file and the audio frame file, and the combination and the storage of the final video picture frame and the audio frame are all processed based on an FFmpeg program. In practice, other open source computer programs related to video processing may also be used.
In one embodiment, the video picture frames with emphasis in the video are also labeled according to the labeling operation of the user. Namely, during optimization processing of the live webcast video, according to the video picture frame marked by the user and corresponding to the timestamp, emphasis identification prompting is carried out on the corresponding video picture frame, so that the timestamp corresponding to the emphasis position in the video can be confirmed when the subsequent live webcast video is watched. And due to the marking operation of the user, the video picture frame under the timestamp does not participate in the compression processing of the repeated picture frame. Furthermore, the video picture frames in the corresponding time period can be selected according to the marking operation of the user, and the video picture frames in the time period do not participate in the compression processing of the repeated picture frames. Therefore, the error compression and the error deletion of the important video picture frames are avoided.
In one embodiment, corresponding to one video source file, only the video picture frame and the audio frame corresponding to the time period may be compressed, merged and stored according to the time period in the video source file selected by the user. And intercepting the video source file in the time period according to the watching requirement by a user so as to preprocess and delete part of irrelevant video file content. It can be understood that, corresponding to a certain network live video source file, the user selects the time period required to be saved according to the actual requirement, and the video file content in some irrelevant time periods is not selected correspondingly. Therefore, the storage of irrelevant video data is reduced through pre-selection, the concentration time of a subsequent user on an invalid video picture when watching a video is further reduced, and the concentration degree of the user is improved.
The method comprises the steps of acquiring a video source file of the network live broadcast, analyzing a corresponding video picture frame file and an audio frame file from the video source file, determining repeated picture frames in the video picture frame file one by one according to a set time interval based on timestamp information, compressing the repeated picture frames, determining audio frames at repeated positions corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the reset audio frames. And merging and storing the processed video picture frame file and the audio frame file based on the timestamp information. By adopting the technical means, a large number of repeated pictures in the live network can be compressed, audio and video synchronization is further realized by compressing or deleting corresponding audio frames, the live network video is optimized and stored, unnecessary video data amount is reduced, and the video watching experience of a user is optimized.
Example two:
on the basis of the foregoing embodiment, fig. 3 is a schematic structural diagram of an optimized storage device for live webcast video according to a second embodiment of the present application. Referring to fig. 3, the optimized saving device for webcast video provided in this embodiment specifically includes: an acquisition module 21, a processing module 22 and a merging module 23.
The acquiring module 21 is configured to acquire a video source file of a live webcast, and parse a corresponding video picture frame file and an audio frame file from the video source file, where the video picture frame file and the audio frame file include corresponding timestamp information;
the processing module 22 is configured to determine repeated picture frames in the video picture frame file one by one according to the time stamp information, compress the repeated picture frames, determine an audio frame at a repeated position in the audio frame file corresponding to the repeated picture frame, and compress or delete the reset audio frame;
and the merging module 23 is configured to merge and store the processed video picture frame file and the audio frame file based on the timestamp information.
The method comprises the steps of acquiring a video source file of the network live broadcast, analyzing a corresponding video picture frame file and an audio frame file from the video source file, determining repeated picture frames in the video picture frame file one by one according to a set time interval based on timestamp information, compressing the repeated picture frames, determining audio frames at repeated positions corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the reset audio frames. And merging and storing the processed video picture frame file and the audio frame file based on the timestamp information. By adopting the technical means, a large number of repeated pictures in the live network can be compressed, audio and video synchronization is further realized by compressing or deleting corresponding audio frames, the live network video is optimized and stored, unnecessary video data amount is reduced, and the video watching experience of a user is optimized.
Specifically, the processing module 22 includes:
the extraction unit is used for extracting image characteristic information of each video picture frame under corresponding timestamp information based on a set time period;
and the comparison unit is used for determining that the video picture frame with the similarity reaching the set threshold is a repeated picture frame based on the image characteristic information comparison.
Specifically, the obtaining module 21 includes:
and the noise reduction unit is used for respectively carrying out noise reduction treatment on the video picture frame file and the audio frame file and eliminating the picture background noise of the video picture frame file and the environment background noise of the audio frame file.
The optimized saving device for the live webcast video provided by the second embodiment of the present application can be used to execute the optimized saving method for the live webcast video provided by the first embodiment of the present application, and has corresponding functions and beneficial effects.
Example three:
an embodiment of the present application provides an electronic device, and with reference to fig. 4, the electronic device includes: a processor 31, a memory 32, a communication module 33, an input device 34, and an output device 35. The number of processors in the electronic device may be one or more, and the number of memories in the electronic device may be one or more. The processor, memory, communication module, input device, and output device of the electronic device may be connected by a bus or other means.
The memory 32 is a computer-readable storage medium, and can be used to store software programs, computer-executable programs, and modules, such as program instructions/modules corresponding to the optimized saving method for webcast video according to any embodiment of the present application (for example, an obtaining module, a processing module, and a merging module in the optimized saving apparatus for webcast video). The memory can mainly comprise a program storage area and a data storage area, wherein the program storage area can store an operating system and an application program required by at least one function; the storage data area may store data created according to use of the device, and the like. Further, the memory may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid state storage device. In some examples, the memory may further include memory located remotely from the processor, and these remote memories may be connected to the device over a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The communication module 33 is used for data transmission.
The processor 31 executes various functional applications and data processing of the device by running software programs, instructions and modules stored in the memory, that is, the above-mentioned optimized saving method for webcast video is realized.
The input device 34 may be used to receive entered numeric or character information and to generate key signal inputs relating to user settings and function controls of the apparatus. The output device 35 may include a display device such as a display screen.
The electronic device provided by the embodiment can be used for executing the optimized storage method of the live webcast video provided by the embodiment, and has corresponding functions and beneficial effects.
Example four:
the embodiment of the present application further provides a storage medium containing computer executable instructions, where the computer executable instructions are executed by a computer processor to perform a method for optimally storing a webcast video, where the method for optimally storing a webcast video includes: acquiring a video source file of network live broadcast, and analyzing a corresponding video picture frame file and an audio frame file from the video source file, wherein the video picture frame file and the audio frame file contain corresponding timestamp information; determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the repeated position audio frames; and merging and storing the processed video picture frame file and the processed audio frame file based on the timestamp information.
Storage medium-any of various types of memory devices or storage devices. The term "storage medium" is intended to include: mounting media such as CD-ROM, floppy disk, or tape devices; computer system memory or random access memory such as DRAM, DDR RAM, SRAM, EDO RAM, Lanbas (Rambus) RAM, etc.; non-volatile memory such as flash memory, magnetic media (e.g., hard disk or optical storage); registers or other similar types of memory elements, etc. The storage medium may also include other types of memory or combinations thereof. In addition, the storage medium may be located in a first computer system in which the program is executed, or may be located in a different second computer system connected to the first computer system through a network (such as the internet). The second computer system may provide program instructions to the first computer for execution. The term "storage medium" may include two or more storage media residing in different locations, e.g., in different computer systems connected by a network. The storage medium may store program instructions (e.g., embodied as a computer program) that are executable by one or more processors.
Of course, the storage medium containing the computer-executable instructions provided in the embodiments of the present application is not limited to the above-described optimized saving method for webcast video, and may also perform related operations in the optimized saving method for webcast video provided in any embodiments of the present application.
The optimized saving device, the storage medium, and the electronic device for webcast video provided in the foregoing embodiments may execute the optimized saving method for webcast video provided in any embodiment of the present application, and reference may be made to the optimized saving method for webcast video provided in any embodiment of the present application without detailed technical details described in the foregoing embodiments.
The foregoing is considered as illustrative of the preferred embodiments of the invention and the technical principles employed. The present application is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the invention. Therefore, although the present application has been described in more detail with reference to the above embodiments, the present application is not limited to the above embodiments, and may include other equivalent embodiments without departing from the spirit of the present application, and the scope of the present application is determined by the scope of the claims.

Claims (10)

1. An optimized storage method for network live video is characterized by comprising the following steps:
acquiring a video source file of network live broadcast, and analyzing a corresponding video picture frame file and an audio frame file from the video source file, wherein the video picture frame file and the audio frame file contain corresponding timestamp information;
determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the repeated position audio frames;
and merging and storing the processed video picture frame file and the processed audio frame file based on the timestamp information.
2. The optimized saving method for webcast video according to claim 1, wherein determining the repeated picture frames in the video picture frame file one by one according to the time stamp information and the set time period comprises:
extracting image characteristic information of each video picture frame under corresponding timestamp information based on a set time period;
and determining the video picture frame with the similarity reaching a set threshold value as a repeated picture frame based on the image characteristic information comparison.
3. The optimized saving method for webcast video according to claim 2, wherein the set time period is 1S.
4. The optimized saving method for webcast video according to claim 1, wherein the compressing or deleting the repeated position audio frame includes:
and determining whether the audio frame at the repeated position is effective audio based on a preset audio waveform analysis model, if so, compressing the audio frame at the repeated position, and if not, deleting the audio frame at the repeated position.
5. The optimized saving method for webcast video according to claim 4, wherein before determining whether the audio frame at the repetition position is valid audio based on a preset audio waveform analysis model, the method further comprises:
and inputting the voice waveform characteristics as training samples into an audio waveform analysis model to perform model training to construct the audio waveform analysis model.
6. The method for optimally saving webcast video according to claim 1, wherein the steps of determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining a repeated audio frame corresponding to the repeated picture frame in the audio frame file, and compressing or deleting the repeated audio frame comprise:
processing the video picture frame file and the audio frame file based on FFmpeg;
correspondingly, the merging and storing the processed video picture frame file and the audio frame file based on the timestamp information includes:
and merging and storing the video picture frame file and the audio frame file based on FFmpeg.
7. The optimized saving method for webcast video according to claim 1, wherein a webcast video source file is obtained, and a corresponding video picture frame file and an audio frame file are parsed from the video source file, further comprising:
and respectively carrying out noise reduction treatment on the video picture frame file and the audio frame file, and eliminating picture background noise of the video picture frame file and environmental background noise of the audio frame file.
8. An optimized save apparatus for webcast video, comprising:
the system comprises an acquisition module, a time stamp module and a processing module, wherein the acquisition module is used for acquiring a video source file of network live broadcast and analyzing a corresponding video picture frame file and an audio frame file from the video source file, and the video picture frame file and the audio frame file contain corresponding time stamp information;
the processing module is used for determining repeated picture frames in the video picture frame file one by one according to the time stamp information, compressing the repeated picture frames, determining repeated position audio frames corresponding to the repeated picture frames in the audio frame file, and compressing or deleting the reset audio frames;
and the merging module is used for merging and storing the processed video picture frame file and the audio frame file based on the timestamp information.
9. An electronic device, comprising:
a memory and one or more processors;
the memory for storing one or more programs;
when executed by the one or more processors, cause the one or more processors to implement a method for optimized preservation of webcast video as recited in any of claims 1-7.
10. A storage medium containing computer-executable instructions for performing the method for optimized preservation of webcast video of any of claims 1-7 when executed by a computer processor.
CN202010301586.0A 2020-04-16 2020-04-16 Optimized storage method and device for live webcast video Active CN111355977B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010301586.0A CN111355977B (en) 2020-04-16 2020-04-16 Optimized storage method and device for live webcast video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010301586.0A CN111355977B (en) 2020-04-16 2020-04-16 Optimized storage method and device for live webcast video

Publications (2)

Publication Number Publication Date
CN111355977A true CN111355977A (en) 2020-06-30
CN111355977B CN111355977B (en) 2022-05-24

Family

ID=71197769

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010301586.0A Active CN111355977B (en) 2020-04-16 2020-04-16 Optimized storage method and device for live webcast video

Country Status (1)

Country Link
CN (1) CN111355977B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112862101A (en) * 2021-01-29 2021-05-28 网易有道信息技术(北京)有限公司 Method and apparatus for optimizing neural network model inference
CN113786605A (en) * 2021-08-23 2021-12-14 咪咕文化科技有限公司 Video processing method, apparatus and computer readable storage medium
CN113873295A (en) * 2021-10-26 2021-12-31 北京金山云网络技术有限公司 Multimedia information processing method, device, equipment and storage medium
CN116166627A (en) * 2023-04-19 2023-05-26 启迪数字科技(深圳)有限公司 Experiment operation track recording method, device, equipment and medium
CN116320521A (en) * 2023-03-24 2023-06-23 吉林动画学院 Three-dimensional animation live broadcast method and device based on artificial intelligence

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5995153A (en) * 1995-11-02 1999-11-30 Prime Image, Inc. Video processing system with real time program duration compression and expansion
US6014491A (en) * 1997-03-04 2000-01-11 Parsec Sight/Sound, Inc. Method and system for manipulation of audio or video signals
CN102761733A (en) * 2011-04-29 2012-10-31 富泰华工业(深圳)有限公司 Monitoring system and video data storage method
CN103747270A (en) * 2013-12-19 2014-04-23 中山大学 Redundancy elimination method and system for capsule endoscope video image
US20140195643A1 (en) * 2012-03-16 2014-07-10 Tencent Technology (Shenzhen) Company Limited Offline download method and system
CN107547817A (en) * 2016-09-12 2018-01-05 郑州蓝视科技有限公司 Video stores and its backup method
CN109151558A (en) * 2018-07-27 2019-01-04 努比亚技术有限公司 A kind of method for processing video frequency, terminal and computer readable storage medium
CN109587489A (en) * 2019-01-11 2019-04-05 杭州富阳优信科技有限公司 A kind of method of video compression
CN109921801A (en) * 2019-03-01 2019-06-21 迅雷计算机(深圳)有限公司 A kind of data compression and decompression method, device, system and data processing system
CN110719425A (en) * 2018-07-11 2020-01-21 视联动力信息技术股份有限公司 Video data playing method and device

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5995153A (en) * 1995-11-02 1999-11-30 Prime Image, Inc. Video processing system with real time program duration compression and expansion
US6014491A (en) * 1997-03-04 2000-01-11 Parsec Sight/Sound, Inc. Method and system for manipulation of audio or video signals
CN102761733A (en) * 2011-04-29 2012-10-31 富泰华工业(深圳)有限公司 Monitoring system and video data storage method
US20140195643A1 (en) * 2012-03-16 2014-07-10 Tencent Technology (Shenzhen) Company Limited Offline download method and system
CN103747270A (en) * 2013-12-19 2014-04-23 中山大学 Redundancy elimination method and system for capsule endoscope video image
CN107547817A (en) * 2016-09-12 2018-01-05 郑州蓝视科技有限公司 Video stores and its backup method
CN110719425A (en) * 2018-07-11 2020-01-21 视联动力信息技术股份有限公司 Video data playing method and device
CN109151558A (en) * 2018-07-27 2019-01-04 努比亚技术有限公司 A kind of method for processing video frequency, terminal and computer readable storage medium
CN109587489A (en) * 2019-01-11 2019-04-05 杭州富阳优信科技有限公司 A kind of method of video compression
CN109921801A (en) * 2019-03-01 2019-06-21 迅雷计算机(深圳)有限公司 A kind of data compression and decompression method, device, system and data processing system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
陈玉芸: "船舶黑匣子中对音视频信号压缩存储处理研究", 《舰船科学技术》 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112862101A (en) * 2021-01-29 2021-05-28 网易有道信息技术(北京)有限公司 Method and apparatus for optimizing neural network model inference
CN112862101B (en) * 2021-01-29 2022-02-08 网易有道信息技术(北京)有限公司 Method and apparatus for optimizing neural network model inference
CN113786605A (en) * 2021-08-23 2021-12-14 咪咕文化科技有限公司 Video processing method, apparatus and computer readable storage medium
CN113786605B (en) * 2021-08-23 2024-03-22 咪咕文化科技有限公司 Video processing method, apparatus and computer readable storage medium
CN113873295A (en) * 2021-10-26 2021-12-31 北京金山云网络技术有限公司 Multimedia information processing method, device, equipment and storage medium
CN116320521A (en) * 2023-03-24 2023-06-23 吉林动画学院 Three-dimensional animation live broadcast method and device based on artificial intelligence
CN116166627A (en) * 2023-04-19 2023-05-26 启迪数字科技(深圳)有限公司 Experiment operation track recording method, device, equipment and medium

Also Published As

Publication number Publication date
CN111355977B (en) 2022-05-24

Similar Documents

Publication Publication Date Title
CN111355977B (en) Optimized storage method and device for live webcast video
CN109242802B (en) Image processing method, image processing device, electronic equipment and computer readable medium
US9576202B1 (en) Systems and methods for identifying a scene-change/non-scene-change transition between frames
EP2960905A1 (en) Method and device of displaying a neutral facial expression in a paused video
WO2022017083A1 (en) Data processing method and apparatus, device, and readable storage medium
CN102655585B (en) Video conference system and time delay testing method, device and system thereof
EP3876546A1 (en) Video processing method and apparatus, and electronic device and storage medium
CN112423075B (en) Audio and video timestamp processing method and device, electronic equipment and storage medium
CN110691281B (en) Video playing processing method, terminal device, server and storage medium
CN111556332A (en) Live broadcast method, electronic device and readable storage medium
CN111757137A (en) Multi-channel close-up playing method and device based on single-shot live video
CN111294600A (en) Compression method, decompression method and device for video sequence frame
CN110335590B (en) Voice recognition test method, device and system
CN113033677A (en) Video classification method and device, electronic equipment and storage medium
CN111757138A (en) Close-up display method and device based on single-shot live video
CN112019827A (en) Method, device, equipment and storage medium for enhancing video image color
CN109286760B (en) Entertainment video production method and terminal thereof
CN104093069B (en) A kind of video broadcasting method and player device
CN111787357A (en) Video processing method and electronic equipment
EP3855754A1 (en) Video processing method and apparatus, electronic device, and storage medium
CN108769525B (en) Image adjusting method, device, equipment and storage medium
CN115439660A (en) Detection method, detection device, electronic equipment and medium
CN115604497A (en) Over-sharpening identification device for live broadcast object
CN110619362B (en) Video content comparison method and device based on perception and aberration
CN114387440A (en) Video clipping method and device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant