WO2022116668A1 - Video filtering method and system based on identical content, and device - Google Patents

Video filtering method and system based on identical content, and device Download PDF

Info

Publication number
WO2022116668A1
WO2022116668A1 PCT/CN2021/121872 CN2021121872W WO2022116668A1 WO 2022116668 A1 WO2022116668 A1 WO 2022116668A1 CN 2021121872 W CN2021121872 W CN 2021121872W WO 2022116668 A1 WO2022116668 A1 WO 2022116668A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
videos
target video
similar
information content
Prior art date
Application number
PCT/CN2021/121872
Other languages
French (fr)
Chinese (zh)
Inventor
李美影
Original Assignee
苏州浪潮智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 苏州浪潮智能科技有限公司 filed Critical 苏州浪潮智能科技有限公司
Publication of WO2022116668A1 publication Critical patent/WO2022116668A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream

Definitions

  • the present application relates to the field of video technology, and in particular, to a content-based video filtering method, system, and device.
  • the purpose of the present application is to propose a video filtering method, system and device based on the same content, so as to solve the problem that the video with the same content cannot be accurately filtered out in the prior art.
  • the present application provides a content-based video filtering method, comprising the following steps:
  • sample extraction processing is performed on the matched video segments, respectively, to obtain n frames of images
  • the image information content is calculated for n frames of images respectively, and the video information content of the corresponding target video and similar videos is obtained based on the calculated image information content, and the videos with small video information content are eliminated.
  • performing timing extraction processing on the target video and the similar video respectively to obtain the respective feature timing diagrams includes: extracting the gray value of the middle pixel point of the image on a frame-by-frame basis from the target video and the similar video respectively, and establishing As a function of it and time, the respective characteristic timing diagrams are obtained.
  • using the dynamic time warping algorithm to match the feature sequence diagram of the target video and the feature sequence diagram of the similar video into the same video segment includes: selecting the feature sequence diagram of the target video and the feature sequence diagram of the similar video by the dynamic time warping algorithm. There are different keyframes in the feature sequence diagram, and based on the keyframes, the same video clips that are matched are obtained.
  • sample extraction processing is performed on the matched video clips based on the duration of the target video, respectively, and the n frames of images obtained respectively include:
  • the matched video segments are extracted frame by frame, and n frames of images are obtained respectively.
  • establishing the extraction frequency function based on the duration of the target video includes: dividing the duration of the target video into four temporal gradients, which are: 1s to 60s, 1min to 60min, 1h to 10h, and 10h or more.
  • the time gradient of establishes different times of decimation.
  • the image information content calculation includes: calculating the image two-dimensional entropy as the image information content.
  • respectively obtaining the video information content of the corresponding target video and similar videos based on the calculated image information content, and removing the videos with small video information content includes:
  • the first average value is obtained to the n-frame image information content of the target video, and the product of the first average value and the target video duration is used as its video information content;
  • the second average value is calculated for the n-frame image information content of the similar video, and the product of the second average value and the similar video duration is used as its video information content;
  • the method further includes: in response to the video with little video information being culled, updating the same-content video library and the regular video library, wherein the regular video library is used to store videos with separate content.
  • Another aspect of the present application also provides a video filtering system based on the same content, including:
  • a video detection module configured to detect and obtain a corresponding similar video in a video library with the same content based on the target video
  • the timing extraction module is configured to perform timing extraction processing on the target video and similar videos respectively to obtain their respective feature sequence diagrams
  • the video segment matching module is configured to match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through the dynamic time warping algorithm;
  • a sample extraction module configured to perform sample extraction processing on the matched video segments based on the duration of the target video, respectively, to obtain n frames of images
  • the video culling module is configured to calculate the image information amount of the n-frame images respectively, and obtain the video information amount of the corresponding target video and similar videos based on the calculated image information amount, and eliminate the video with a small amount of video information. .
  • a computer device including a memory and a processor, where a computer program is stored in the memory, and when the computer program is executed by the processor, any one of the above methods is executed.
  • FIG. 1 is a schematic diagram of an embodiment of a content-based video filtering method provided according to the present application.
  • FIG. 2 is a feature sequence diagram matched by a target video and a corresponding similar video through a dynamic time warping algorithm according to an embodiment of the present application;
  • FIG. 3 is a schematic diagram of an embodiment of a video filtering system based on the same content provided according to the present application;
  • FIG. 4 is a schematic diagram of a hardware structure of an embodiment of a computer device for performing the same content-based video filtering method provided by the present application.
  • FIG. 1 shows a schematic diagram of an embodiment of the same content-based video filtering method provided by the present application.
  • the embodiment of the present application includes the following steps:
  • Step S10 based on the target video, detect and obtain the corresponding similar video in the video library with the same content;
  • Step S20 performing timing extraction processing on the target video and the similar video respectively, to obtain respective feature sequence diagrams
  • Step S30 match the feature sequence diagram of the target video and the feature sequence diagram of the similar video to the same video segment by the dynamic time warping algorithm
  • Step S40 performing sample extraction processing on the matched video clips based on the duration of the target video, respectively obtaining n frames of images;
  • Step S50 Calculate the amount of image information for the n frames of images respectively, and obtain the video information amount of the corresponding target video and similar videos based on the calculated amount of image information, and remove the videos with small video information amount.
  • the target video and similar videos are respectively subjected to timing extraction processing and video clip matching to efficiently screen out video clips with consistent content; through sample extraction processing, image information content calculation, and video information content calculation, information is selected. More valuable videos; by eliminating videos with small amount of video information, the disk usage is reduced, network resources are saved, and at the same time, it helps to improve the efficiency of video retrieval, and further brings greater benefits to the video platform. Economic Value.
  • performing timing extraction processing on the target video and the similar video respectively to obtain the respective feature timing diagrams includes: extracting the gray value of the middle pixel point of the image on a frame-by-frame basis from the target video and the similar video respectively, and establishing As a function of it and time, the respective characteristic timing diagrams are obtained.
  • Figure 2 shows the feature sequence diagram of the target video and similar videos, where video A represents the target video, video B represents the similar video, the abscissa indicates that video A and video B are extracted regularly, and the ordinate indicates that The gray value of the middle pixel of the image of the corresponding time frame, the gray value and time form a time series function, and the time series is used as the feature value of the video.
  • using the dynamic time warping algorithm to match the feature sequence diagram of the target video and the feature sequence diagram of the similar video into the same video segment includes: selecting the feature sequence diagram of the target video and the feature sequence diagram of the similar video by the dynamic time warping algorithm. There are different keyframes in the feature sequence diagram, and based on the keyframes, the same video clips that are matched are obtained.
  • Fig. 2 is the feature sequence diagram that the target video according to the embodiment of the present application and the corresponding similar video are matched by the dynamic time warping algorithm, the key frame with difference is obtained by the dynamic time warping algorithm, and the matching is shown by the dotted line segmentation in the figure of the same video clip.
  • performing sample extraction processing on the matched video clips based on the duration of the target video, respectively obtaining n frames of images includes: establishing an extraction frequency function based on the duration of the target video; Fragments are extracted frame by frame, and each obtains n frames of images.
  • establishing a decimation frequency function based on the duration of the target video includes: dividing the duration of the target video into four temporal gradients, which are: 1s to 60s, 1min to 60min, 1h to 10h, and more than 10h. The time gradient of , establishes different times of decimation. In this embodiment, the longer the video time is, the more times of extraction, because the longer the video is, the more data is required for comparison to complete the similarity comparison.
  • the extraction frequency function f(x) is shown in the following formula, where x is the video time length; when the video time is 1s to 60s, it is extracted once in 0.05x seconds; when the video time is from 1min to 60min, it is extracted once in 0.025x minutes; If the time is from 1h to 10h, it will be extracted once every 0.01x hour. When the video time is more than 10h, it will be extracted once every 0.01x hour.
  • the image information content calculation includes: calculating the image two-dimensional entropy as the image information content.
  • the two-dimensional entropy of the image is used as the amount of image information
  • the average gray value of the neighborhood of image pixels is selected as the spatial feature of gray distribution, which is combined with the pixel gray of the image to form a feature two-tuple, denoted as (i, j) , f(i,j) represents the frequency of the feature two-tuple (i,j), where i represents the gray value of the pixel, j represents the average gray value of the neighborhood;
  • M, N represent the image size; such as the following formula:
  • the two-dimensional entropy of the image is as follows:
  • obtaining the video information content of the corresponding target video and similar videos based on the calculated image information content, and removing the video with a small video information content includes: calculating the first image information content of n frames of the target video. an average value, and the product of the first average value and the target video duration is used as its video information content; a second average value is obtained for the information content of n frames of similar videos, and the product of the second average value and the similar video time duration is used as Its video information content; compare the video information content of the target video and the video information content of similar videos, and remove the videos with small video information content.
  • the amount of video information is equal to the two-dimensional entropy of the image multiplied by the duration, and H sum represents the amount of video information, L indicates the duration of the video, and the unit is frame, then the amount of video information is:
  • the duplicate videos with small amount of information will be eliminated or deleted. That is, if the amount of video information of the target video is small, the target video will be eliminated. If the amount of video information of the corresponding similar video Less, similar videos will be eliminated.
  • the method further includes: in response to the video with little video information being culled, updating the same-content video library and the regular video library, wherein the regular video library is used to store videos with separate content.
  • FIG. 3 shows a schematic diagram of an embodiment of the same content-based video filtering system provided by the present application.
  • a video filtering system based on the same content includes: a video detection module 10, configured to detect and obtain a corresponding similar video in a video library with the same content based on a target video; a timing extraction module 20, configured to detect the target video and the similar video.
  • the video segment matching module 30 is configured to match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through a dynamic time warping algorithm; sample extraction The module 40 is configured to perform sample extraction processing on the matched video clips based on the duration of the target video, respectively, to obtain n frames of images; The video information content of the corresponding target video and similar videos is obtained based on the calculated image information content, and the videos with small video information content are eliminated.
  • the video filtering system based on the same content in this embodiment efficiently filters out video clips with the same content by performing timing extraction processing and video clip matching on the target video and similar videos respectively; By calculating the amount of information, videos with more information and more valuable are selected; by eliminating the videos with less video information, the disk usage is reduced, network resources are saved, and at the same time, it helps to improve the efficiency of retrieving videos, and further improves the efficiency of video retrieval. Video platforms bring great economic value.
  • a computer device including a memory 302 and a processor 301, where a computer program is stored in the memory, and when the computer program is executed by the processor, any one of the foregoing embodiments is implemented method.
  • FIG. 4 it is a schematic diagram of a hardware structure of an embodiment of a computer device for performing the same content-based video filtering method provided by the present application.
  • the computer device includes a processor 301 and a memory 302 , and may also include an input device 303 and an output device 304 .
  • the processor 301 , the memory 302 , the input device 303 and the output device 304 may be connected by a bus or in other ways, and the connection by a bus is taken as an example in FIG. 4 .
  • the input device 303 may receive input numerical or character information, and generate key signal input related to user settings and function control of the same content-based video filtering system.
  • the output device 304 may include a display device such as a display screen.
  • the processor 301 executes various functional applications and data processing of the server by running the non-volatile software programs, instructions and modules stored in the memory 302, that is, implementing the same content-based video filtering method of the above method embodiments.
  • nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory memory.
  • Volatile memory may include random access memory (RAM), which may act as external cache memory.
  • RAM is available in various forms such as Synchronous RAM (DRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDR SDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), and Direct Rambus RAM (DRRAM).
  • DRAM Synchronous RAM
  • DRAM Dynamic RAM
  • SDRAM Synchronous DRAM
  • DDR SDRAM Double Data Rate SDRAM
  • ESDRAM Enhanced SDRAM
  • SLDRAM Synchronous Link DRAM
  • DRRAM Direct Rambus RAM
  • the storage devices of the disclosed aspects are intended to include, but not be limited to, these and other suitable types of memory.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

Provided are a video filtering method and system based on identical content, and a device. The method comprises the following steps: on the basis of a target video, detecting and acquiring corresponding similar videos from a library of videos having identical content; respectively performing timing extraction on the target video and the similar videos so as to obtain respective feature timing diagrams; matching the feature timing diagram of the target video and the feature timing diagrams of the similar videos by using a dynamic time warping algorithm, so as to obtain identical video clips; respectively performing, on the basis of the duration of the target video, sample extraction on the video clips obtained by means of matching, so as to obtain N frames of images, respectively; and respectively performing image information amount calculation on the N frames of images, respectively obtaining corresponding video information amounts of the target video and the similar videos on the basis of the calculated image information amount, and removing videos with a small video information amount. In the present application, videos with a greater information amount and having higher values are efficiently screened out, thereby reducing disk usage, saving on network resources, and facilitating the improvement of the efficiency of video retrieval.

Description

一种基于内容相同的视频过滤方法、系统及设备A kind of video filtering method, system and device based on the same content
本申请要求在2020年12月04日提交中国专利局、申请号为202011406954.4、发明名称为“一种基于内容相同的视频过滤方法、系统及设备”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed on December 04, 2020 with the application number 202011406954.4 and the title of the invention is "a video filtering method, system and device based on the same content", the entire contents of which are obtained through Reference is incorporated in this application.
技术领域technical field
本申请涉及视频技术领域,尤其涉及一种基于内容相同的视频过滤方法、系统及设备。The present application relates to the field of video technology, and in particular, to a content-based video filtering method, system, and device.
背景技术Background technique
随着信息技术的快速发展与广泛应用,网络视频越来越多,但同时也存在着大量的相似视频,它们视频内容一样,只是进行了格式转换、伸缩变形、增加水印、广告、滤镜等。相似视频内容重复且占用了大量的磁盘资源、网络资源,同时也拉低了视频检索速度,由此导致了巨大的经济价值浪费。对于视频平台来说,成本是重中之重,相似视频内容相同、只具有极低的价值甚至可以直接忽略,却占用了大量的资源,与此同时,推送相同内容视频给用户容易影响使用体验。因此迫切需要对内容相同视频进行筛选,如何筛选保留价值更高的视频,就成为了当前迫切需要解决的问题。With the rapid development and wide application of information technology, there are more and more online videos, but at the same time, there are also a large number of similar videos. Their video content is the same, but they have undergone format conversion, scaling and deformation, adding watermarks, advertisements, filters, etc. . Similar video content is repeated and occupies a lot of disk resources and network resources, and also slows down the video retrieval speed, which leads to a huge waste of economic value. For video platforms, cost is the top priority. Similar videos have the same content, have very low value or can even be ignored, but take up a lot of resources. At the same time, pushing the same video to users will easily affect the user experience. . Therefore, it is urgent to screen videos with the same content, and how to screen and retain videos with higher value has become an urgent problem to be solved at present.
现有的内容相同视频过滤技术及存在的技术问题如下:The existing video filtering technologies with the same content and the existing technical problems are as follows:
1.对比视频文件大小,默认大文件信息量更大,过滤掉小文件;在视频格式转换、视频分辨率调高、视频每秒帧数增加等情况下,此方法很可能选出信息量不大或信息量相同却更消耗资源的视频;1. Compare the size of video files. By default, large files have a larger amount of information, and small files are filtered out. In the case of video format conversion, video resolution increase, and video frames per second increase, this method is likely to select the most information content. Larger or more resource-intensive videos with the same amount of information;
2.对比视频分辨率,默认大分辨率信息量更大,过滤掉低分辨率视频; 然而一些视频网站为统一分辨率管理,将视频源文件转换为更高分辨率的视频以适配平台,该情况下此方法很可能选出信息量不大或信息量相同却更消耗资源的视频;2. Compared with the video resolution, the default large resolution has more information, and the low-resolution video is filtered out; however, some video websites manage the unified resolution and convert the video source file to a higher-resolution video to adapt to the platform. In this case, this method is likely to select videos with less information or the same amount of information but more resource-consuming videos;
3.对比视频长短,默认时长越长信息量更大,过滤掉低时长视频;视频长但模糊不清、视频短但更清晰信息量更大,该情况下此方法很可能选出信息量小的视频。3. Comparing the length of the video, the longer the default duration, the greater the amount of information, and the low-duration video is filtered out; the video is long but blurry, and the video is short but clearer, and the amount of information is greater. In this case, this method is likely to select a small amount of information. 's video.
发明内容SUMMARY OF THE INVENTION
有鉴于此,本申请的目的在于提出一种基于内容相同的视频过滤方法、系统及设备,用以解决现有技术中无法准确过滤掉内容相同的视频的问题。In view of this, the purpose of the present application is to propose a video filtering method, system and device based on the same content, so as to solve the problem that the video with the same content cannot be accurately filtered out in the prior art.
基于上述目的,本申请提供了一种基于内容相同的视频过滤方法,包括如下步骤:Based on the above purpose, the present application provides a content-based video filtering method, comprising the following steps:
基于目标视频在内容相同视频库中检测并获取对应的相似视频;Detect and obtain the corresponding similar videos in the video library with the same content based on the target video;
对目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图;Perform timing extraction processing on the target video and similar videos respectively to obtain their respective feature sequence diagrams;
通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段;Match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through the dynamic time warping algorithm;
基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像;Based on the duration of the target video, sample extraction processing is performed on the matched video segments, respectively, to obtain n frames of images;
分别对n帧图像进行图像信息量计算,且分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除。The image information content is calculated for n frames of images respectively, and the video information content of the corresponding target video and similar videos is obtained based on the calculated image information content, and the videos with small video information content are eliminated.
在一些实施例中,对目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图包括:对目标视频和相似视频分别定时按帧抽取图像最中间像素点的灰度值,并分别建立其与时间的函数,得到各自的特征时序图。In some embodiments, performing timing extraction processing on the target video and the similar video respectively to obtain the respective feature timing diagrams includes: extracting the gray value of the middle pixel point of the image on a frame-by-frame basis from the target video and the similar video respectively, and establishing As a function of it and time, the respective characteristic timing diagrams are obtained.
在一些实施例中,通过动态时间规整算法将目标视频的特征时序图和 相似视频的特征时序图匹配出相同的视频片段包括:通过动态时间规整算法选出目标视频的特征时序图和相似视频的特征时序图中有差异的关键帧,并基于关键帧得到匹配的相同的视频片段。In some embodiments, using the dynamic time warping algorithm to match the feature sequence diagram of the target video and the feature sequence diagram of the similar video into the same video segment includes: selecting the feature sequence diagram of the target video and the feature sequence diagram of the similar video by the dynamic time warping algorithm. There are different keyframes in the feature sequence diagram, and based on the keyframes, the same video clips that are matched are obtained.
在一些实施例中,基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像包括:In some embodiments, sample extraction processing is performed on the matched video clips based on the duration of the target video, respectively, and the n frames of images obtained respectively include:
建立基于目标视频的时长的抽取频率函数;Establish a decimation frequency function based on the duration of the target video;
根据抽取频率函数对匹配出的视频片段按帧抽取,各自得到n帧图像。According to the extraction frequency function, the matched video segments are extracted frame by frame, and n frames of images are obtained respectively.
在一些实施例中,建立基于目标视频的时长的抽取频率函数包括:将目标视频的时长分为四个时间梯度,分别是:1s至60s、1min至60min、1h至10h及10h以上,对不同的时间梯度建立不同的抽取次数。In some embodiments, establishing the extraction frequency function based on the duration of the target video includes: dividing the duration of the target video into four temporal gradients, which are: 1s to 60s, 1min to 60min, 1h to 10h, and 10h or more. The time gradient of , establishes different times of decimation.
在一些实施例中,图像信息量计算包括:计算图像二维熵作为图像信息量。In some embodiments, the image information content calculation includes: calculating the image two-dimensional entropy as the image information content.
在一些实施例中,分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除包括:In some embodiments, respectively obtaining the video information content of the corresponding target video and similar videos based on the calculated image information content, and removing the videos with small video information content includes:
对目标视频的n帧图像信息量求第一平均值,并以第一平均值与目标视频时长的乘积作为其视频信息量;The first average value is obtained to the n-frame image information content of the target video, and the product of the first average value and the target video duration is used as its video information content;
对相似视频的n帧图像信息量求第二平均值,并以第二平均值与相似视频时长的乘积作为其视频信息量;The second average value is calculated for the n-frame image information content of the similar video, and the product of the second average value and the similar video duration is used as its video information content;
比较目标视频的视频信息量和相似视频的视频信息量,并将其中视频信息量小的视频剔除。Compare the video information content of the target video with the video information content of similar videos, and remove the videos with less video information.
在一些实施例中,方法还包括:响应于视频信息量小的视频被剔除,更新内容相同视频库和常规视频库,其中,常规视频库用于存放具有单独内容的视频。In some embodiments, the method further includes: in response to the video with little video information being culled, updating the same-content video library and the regular video library, wherein the regular video library is used to store videos with separate content.
本申请的另一方面,还提供了一种基于内容相同的视频过滤系统,包括:Another aspect of the present application also provides a video filtering system based on the same content, including:
视频检测模块,配置用于基于目标视频在内容相同视频库中检测并获 取对应的相似视频;A video detection module, configured to detect and obtain a corresponding similar video in a video library with the same content based on the target video;
定时抽取模块,配置用于对目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图;The timing extraction module is configured to perform timing extraction processing on the target video and similar videos respectively to obtain their respective feature sequence diagrams;
视频片段匹配模块,配置用于通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段;The video segment matching module is configured to match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through the dynamic time warping algorithm;
样本抽取模块,配置用于基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像;以及a sample extraction module, configured to perform sample extraction processing on the matched video segments based on the duration of the target video, respectively, to obtain n frames of images; and
视频剔除模块,配置用于分别对n帧图像进行图像信息量计算,且分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除。The video culling module is configured to calculate the image information amount of the n-frame images respectively, and obtain the video information amount of the corresponding target video and similar videos based on the calculated image information amount, and eliminate the video with a small amount of video information. .
本申请的再一方面,还提供了一种计算机设备,包括存储器和处理器,存储器中存储有计算机程序,该计算机程序被处理器执行时执行上述任意一项方法。In another aspect of the present application, a computer device is also provided, including a memory and a processor, where a computer program is stored in the memory, and when the computer program is executed by the processor, any one of the above methods is executed.
本申请至少具有以下有益技术效果:The present application has at least the following beneficial technical effects:
本申请通过对目标视频和相似视频分别进行定时抽取处理和视频片段匹配,高效筛选出了内容一致的视频片段;通过样本抽取处理、图像信息量计算和视频信息量计算,选出了信息量更大、更具有价值的视频;通过将视频信息量小的视频剔除,降低了磁盘占用,节约了网络资源,同时有助于提高检索视频的效率,进一步为视频平台带来了较大的经济价值。In this application, by performing timing extraction processing and video clip matching on the target video and similar videos, the video clips with the same content are efficiently screened; Large and more valuable videos; by eliminating videos with small video information, the disk usage is reduced, network resources are saved, and at the same time, it helps to improve the efficiency of video retrieval, and further brings greater economic value to the video platform. .
附图说明Description of drawings
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的实施例。In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the following briefly introduces the accompanying drawings required for the description of the embodiments or the prior art. Obviously, the drawings in the following description are only These are some embodiments of the present application. For those of ordinary skill in the art, other embodiments can also be obtained according to these drawings without creative efforts.
图1为根据本申请提供的基于内容相同的视频过滤方法的实施例的示 意图;1 is a schematic diagram of an embodiment of a content-based video filtering method provided according to the present application;
图2为根据本申请实施例的目标视频和对应的相似视频经动态时间规整算法匹配的特征时序图;2 is a feature sequence diagram matched by a target video and a corresponding similar video through a dynamic time warping algorithm according to an embodiment of the present application;
图3为根据本申请提供的基于内容相同的视频过滤系统的实施例的示意图;3 is a schematic diagram of an embodiment of a video filtering system based on the same content provided according to the present application;
图4为本申请提供的执行基于内容相同的视频过滤方法的计算机设备的一个实施例的硬件结构示意图。FIG. 4 is a schematic diagram of a hardware structure of an embodiment of a computer device for performing the same content-based video filtering method provided by the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案和优点更加清楚明白,以下结合具体实施例,并参照附图,对本申请实施例进一步详细说明。In order to make the objectives, technical solutions and advantages of the present application clearer, the following describes the embodiments of the present application in detail with reference to the accompanying drawings and specific embodiments.
需要说明的是,本申请实施例中所有使用“第一”和“第二”的表述均是为了区分两个相同名称的非相同的实体或者非相同的参量,可见“第一”“第二”仅为了表述的方便,不应理解为对本申请实施例的限定。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备固有的其他步骤或单元。It should be noted that all the expressions using "first" and "second" in the embodiments of this application are to distinguish two non-identical entities or non-identical parameters with the same name. " is only for the convenience of expression, and should not be construed as a limitation on the embodiments of the present application. Furthermore, the terms "comprising" and "having", and any variations thereof, are intended to cover non-exclusive inclusion, for example, other steps or units inherent in a process, method, system, product or apparatus comprising a series of steps or units.
基于上述目的,本申请实施例的第一个方面,提出了一种基于内容相同的视频过滤方法的实施例。图1示出的是本申请提供的基于内容相同的视频过滤方法的实施例的示意图。如图1所示,本申请实施例包括如下步骤:Based on the above purpose, in the first aspect of the embodiments of the present application, an embodiment of a video filtering method based on the same content is proposed. FIG. 1 shows a schematic diagram of an embodiment of the same content-based video filtering method provided by the present application. As shown in Figure 1, the embodiment of the present application includes the following steps:
步骤S10、基于目标视频在内容相同视频库中检测并获取对应的相似视频;Step S10, based on the target video, detect and obtain the corresponding similar video in the video library with the same content;
步骤S20、对目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图;Step S20, performing timing extraction processing on the target video and the similar video respectively, to obtain respective feature sequence diagrams;
步骤S30、通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段;Step S30, match the feature sequence diagram of the target video and the feature sequence diagram of the similar video to the same video segment by the dynamic time warping algorithm;
步骤S40、基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像;Step S40, performing sample extraction processing on the matched video clips based on the duration of the target video, respectively obtaining n frames of images;
步骤S50、分别对n帧图像进行图像信息量计算,且分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除。Step S50: Calculate the amount of image information for the n frames of images respectively, and obtain the video information amount of the corresponding target video and similar videos based on the calculated amount of image information, and remove the videos with small video information amount.
本申请实施例通过对目标视频和相似视频分别进行定时抽取处理和视频片段匹配,高效筛选出了内容一致的视频片段;通过样本抽取处理、图像信息量计算和视频信息量计算,选出了信息量更大、更具有价值的视频;通过将视频信息量小的视频剔除,降低了磁盘占用,节约了网络资源,同时有助于提高检索视频的效率,进一步为视频平台带来了较大的经济价值。In the embodiment of the present application, the target video and similar videos are respectively subjected to timing extraction processing and video clip matching to efficiently screen out video clips with consistent content; through sample extraction processing, image information content calculation, and video information content calculation, information is selected. More valuable videos; by eliminating videos with small amount of video information, the disk usage is reduced, network resources are saved, and at the same time, it helps to improve the efficiency of video retrieval, and further brings greater benefits to the video platform. Economic Value.
在一些实施例中,对目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图包括:对目标视频和相似视频分别定时按帧抽取图像最中间像素点的灰度值,并分别建立其与时间的函数,得到各自的特征时序图。图2示出了目标视频和相似视频的特征时序图,其中,视频A表示目标视频,视频B表示相似视频,横坐标表示出了视频A和视频B是被定时抽取的,纵坐标表示出了相应时间帧的图像的最中间像素点的灰度值,灰度值与时间形成了时间序列函数,时间序列即作为视频的特征值。In some embodiments, performing timing extraction processing on the target video and the similar video respectively to obtain the respective feature timing diagrams includes: extracting the gray value of the middle pixel point of the image on a frame-by-frame basis from the target video and the similar video respectively, and establishing As a function of it and time, the respective characteristic timing diagrams are obtained. Figure 2 shows the feature sequence diagram of the target video and similar videos, where video A represents the target video, video B represents the similar video, the abscissa indicates that video A and video B are extracted regularly, and the ordinate indicates that The gray value of the middle pixel of the image of the corresponding time frame, the gray value and time form a time series function, and the time series is used as the feature value of the video.
在一些实施例中,通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段包括:通过动态时间规整算法选出目标视频的特征时序图和相似视频的特征时序图中有差异的关键帧,并基于关键帧得到匹配的相同的视频片段。图2为根据本申请实施例的目标视频和对应的相似视频经动态时间规整算法匹配的特征时序图,经动态时间规整算法得到了有差异的关键帧,通过图中的虚线分割示意出了匹配的相同的视频片段。In some embodiments, using the dynamic time warping algorithm to match the feature sequence diagram of the target video and the feature sequence diagram of the similar video into the same video segment includes: selecting the feature sequence diagram of the target video and the feature sequence diagram of the similar video by the dynamic time warping algorithm. There are different keyframes in the feature sequence diagram, and based on the keyframes, the same video clips that are matched are obtained. Fig. 2 is the feature sequence diagram that the target video according to the embodiment of the present application and the corresponding similar video are matched by the dynamic time warping algorithm, the key frame with difference is obtained by the dynamic time warping algorithm, and the matching is shown by the dotted line segmentation in the figure of the same video clip.
在一些实施例中,基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像包括:建立基于目标视频的时长的抽取频率函数;根据抽取频率函数对匹配出的视频片段按帧抽取,各自得到n帧图像。在一些实施例中,建立基于目标视频的时长的抽取频率函数包括: 将目标视频的时长分为四个时间梯度,分别是:1s至60s、1min至60min、1h至10h及10h以上,对不同的时间梯度建立不同的抽取次数。本实施例中,视频时间越长,抽取次数越多,因为视频越长,越需要更多的数据来进行比对来完成相似度比较。抽取频率函数f(x)如以下公式,其中x为视频时间长度;当视频时间为1s至60s,则0.05x秒抽取一次;当视频时间为1min至60min,则0.025x分钟抽取一次;当视频时间为1h至10h,则0.01x小时抽取一次,当视频时间为10h以上,则0.01x小时抽取一次。In some embodiments, performing sample extraction processing on the matched video clips based on the duration of the target video, respectively obtaining n frames of images includes: establishing an extraction frequency function based on the duration of the target video; Fragments are extracted frame by frame, and each obtains n frames of images. In some embodiments, establishing a decimation frequency function based on the duration of the target video includes: dividing the duration of the target video into four temporal gradients, which are: 1s to 60s, 1min to 60min, 1h to 10h, and more than 10h. The time gradient of , establishes different times of decimation. In this embodiment, the longer the video time is, the more times of extraction, because the longer the video is, the more data is required for comparison to complete the similarity comparison. The extraction frequency function f(x) is shown in the following formula, where x is the video time length; when the video time is 1s to 60s, it is extracted once in 0.05x seconds; when the video time is from 1min to 60min, it is extracted once in 0.025x minutes; If the time is from 1h to 10h, it will be extracted once every 0.01x hour. When the video time is more than 10h, it will be extracted once every 0.01x hour.
Figure PCTCN2021121872-appb-000001
Figure PCTCN2021121872-appb-000001
在一些实施例中,图像信息量计算包括:计算图像二维熵作为图像信息量。本实施例中,使用图像二维熵作为图像信息量,选择图像像素点邻域灰度均值作为灰度分布空间特征,与图像的像素灰度组成特征二元组,记为(i,j),f(i,j)表示特征二元组(i,j)出现的频数,其中i表示像素的灰度值,j表示邻域灰度均值;M、N表示图像尺寸;如以下公式:In some embodiments, the image information content calculation includes: calculating the image two-dimensional entropy as the image information content. In this embodiment, the two-dimensional entropy of the image is used as the amount of image information, and the average gray value of the neighborhood of image pixels is selected as the spatial feature of gray distribution, which is combined with the pixel gray of the image to form a feature two-tuple, denoted as (i, j) , f(i,j) represents the frequency of the feature two-tuple (i,j), where i represents the gray value of the pixel, j represents the average gray value of the neighborhood; M, N represent the image size; such as the following formula:
Figure PCTCN2021121872-appb-000002
Figure PCTCN2021121872-appb-000002
图像二维熵如以下公式:The two-dimensional entropy of the image is as follows:
Figure PCTCN2021121872-appb-000003
Figure PCTCN2021121872-appb-000003
在一些实施例中,分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除包括:对目标视频的n帧图像信息量求第一平均值,并以第一平均值与目标视频时长的乘积作为其视频信息量;对相似视频的n帧图像信息量求第二平均值,并以第二平均值与相似视频时长的乘积作为其视频信息量;比较目标视频的视频信息量和相似视频的视频信息量,并将其中视频信息量小的视频剔除。本实施例中,视频信息量等于图像二维熵乘以时长,以H sum表示视频信息量,L表示视频时长,单位为帧,则视频信息量为: In some embodiments, obtaining the video information content of the corresponding target video and similar videos based on the calculated image information content, and removing the video with a small video information content includes: calculating the first image information content of n frames of the target video. an average value, and the product of the first average value and the target video duration is used as its video information content; a second average value is obtained for the information content of n frames of similar videos, and the product of the second average value and the similar video time duration is used as Its video information content; compare the video information content of the target video and the video information content of similar videos, and remove the videos with small video information content. In this embodiment, the amount of video information is equal to the two-dimensional entropy of the image multiplied by the duration, and H sum represents the amount of video information, L indicates the duration of the video, and the unit is frame, then the amount of video information is:
H sum=H*L H sum = H*L
得到视频信息量,信息量越小可以被认为价值更小,剔除或删除信息量小的重复视频,即若目标视频的视频信息量小,将目标视频剔除,若对应的相似视频的视频信息量少,将相似视频剔除。Get the video information amount. The smaller the amount of information, the smaller the value. The duplicate videos with small amount of information will be eliminated or deleted. That is, if the amount of video information of the target video is small, the target video will be eliminated. If the amount of video information of the corresponding similar video Less, similar videos will be eliminated.
在一些实施例中,方法还包括:响应于视频信息量小的视频被剔除,更新内容相同视频库和常规视频库,其中,常规视频库用于存放具有单独内容的视频。通过对内容相同视频库和常规视频库进行更新,便于提高后续视频检索的效率。In some embodiments, the method further includes: in response to the video with little video information being culled, updating the same-content video library and the regular video library, wherein the regular video library is used to store videos with separate content. By updating the video library with the same content and the conventional video library, it is convenient to improve the efficiency of subsequent video retrieval.
本申请实施例的第二个方面,还提供了一种基于内容相同的视频过滤系统。图3示出的是本申请提供的基于内容相同的视频过滤系统的实施例的示意图。一种基于内容相同的视频过滤系统包括:视频检测模块10,配置用于基于目标视频在内容相同视频库中检测并获取对应的相似视频;定时抽取模块20,配置用于对目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图;视频片段匹配模块30,配置用于通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段;样本抽取模块40,配置用于基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像;以及视频剔除模块50,配置用于分别对n帧图像进行图像信息量计算,且分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除。In a second aspect of the embodiments of the present application, a video filtering system based on the same content is also provided. FIG. 3 shows a schematic diagram of an embodiment of the same content-based video filtering system provided by the present application. A video filtering system based on the same content includes: a video detection module 10, configured to detect and obtain a corresponding similar video in a video library with the same content based on a target video; a timing extraction module 20, configured to detect the target video and the similar video. Respectively perform timing extraction processing to obtain respective feature sequence diagrams; the video segment matching module 30 is configured to match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through a dynamic time warping algorithm; sample extraction The module 40 is configured to perform sample extraction processing on the matched video clips based on the duration of the target video, respectively, to obtain n frames of images; The video information content of the corresponding target video and similar videos is obtained based on the calculated image information content, and the videos with small video information content are eliminated.
本实施例的基于内容相同的视频过滤系统,通过对目标视频和相似视频分别进行定时抽取处理和视频片段匹配,高效筛选出了内容一致的视频片段;通过样本抽取处理、图像信息量计算和视频信息量计算,选出了信息量更大、更具有价值的视频;通过将视频信息量小的视频剔除,降低了磁盘占用,节约了网络资源,同时有助于提高检索视频的效率,进一步为视频平台带来了较大的经济价值。The video filtering system based on the same content in this embodiment efficiently filters out video clips with the same content by performing timing extraction processing and video clip matching on the target video and similar videos respectively; By calculating the amount of information, videos with more information and more valuable are selected; by eliminating the videos with less video information, the disk usage is reduced, network resources are saved, and at the same time, it helps to improve the efficiency of retrieving videos, and further improves the efficiency of video retrieval. Video platforms bring great economic value.
本申请实施例的第三个方面,还提供了一种计算机设备,包括存储器302和处理器301,该存储器中存储有计算机程序,该计算机程序被该处理 器执行时实现上述任意一项实施例方法。In a third aspect of the embodiments of the present application, a computer device is further provided, including a memory 302 and a processor 301, where a computer program is stored in the memory, and when the computer program is executed by the processor, any one of the foregoing embodiments is implemented method.
如图4所示,为本申请提供的执行基于内容相同的视频过滤方法的计算机设备的一个实施例的硬件结构示意图。以如图4所示的计算机设备为例,在该计算机设备中包括一个处理器301以及一个存储器302,并还可以包括:输入装置303和输出装置304。处理器301、存储器302、输入装置303和输出装置304可以通过总线或者其他方式连接,图4中以通过总线连接为例。输入装置303可接收输入的数字或字符信息,以及产生与基于内容相同的视频过滤系统的用户设置以及功能控制有关的键信号输入。输出装置304可包括显示屏等显示设备。处理器301通过运行存储在存储器302中的非易失性软件程序、指令以及模块,从而执行服务器的各种功能应用以及数据处理,即实现上述方法实施例的基于内容相同的视频过滤方法。As shown in FIG. 4 , it is a schematic diagram of a hardware structure of an embodiment of a computer device for performing the same content-based video filtering method provided by the present application. Taking the computer device shown in FIG. 4 as an example, the computer device includes a processor 301 and a memory 302 , and may also include an input device 303 and an output device 304 . The processor 301 , the memory 302 , the input device 303 and the output device 304 may be connected by a bus or in other ways, and the connection by a bus is taken as an example in FIG. 4 . The input device 303 may receive input numerical or character information, and generate key signal input related to user settings and function control of the same content-based video filtering system. The output device 304 may include a display device such as a display screen. The processor 301 executes various functional applications and data processing of the server by running the non-volatile software programs, instructions and modules stored in the memory 302, that is, implementing the same content-based video filtering method of the above method embodiments.
最后需要说明的是,本文的计算机可读存储介质(例如,存储器)可以是易失性存储器或非易失性存储器,或者可以包括易失性存储器和非易失性存储器两者。作为例子而非限制性的,非易失性存储器可以包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦写可编程ROM(EEPROM)或快闪存储器。易失性存储器可以包括随机存取存储器(RAM),该RAM可以充当外部高速缓存存储器。作为例子而非限制性的,RAM可以以多种形式获得,比如同步RAM(DRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据速率SDRAM(DDR SDRAM)、增强SDRAM(ESDRAM)、同步链路DRAM(SLDRAM)、以及直接Rambus RAM(DRRAM)。所公开的方面的存储设备意在包括但不限于这些和其它合适类型的存储器。Finally, it should be noted that computer-readable storage media (eg, memory) herein may be volatile memory or non-volatile memory, or may include both volatile and non-volatile memory. By way of example and not limitation, nonvolatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory memory. Volatile memory may include random access memory (RAM), which may act as external cache memory. By way of example and not limitation, RAM is available in various forms such as Synchronous RAM (DRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDR SDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), and Direct Rambus RAM (DRRAM). The storage devices of the disclosed aspects are intended to include, but not be limited to, these and other suitable types of memory.
本领域技术人员还将明白的是,结合这里的公开所描述的各种示例性逻辑块、模块、电路和算法步骤可以被实现为电子硬件、计算机软件或两者的组合。为了清楚地说明硬件和软件的这种可互换性,已经就各种示意性组件、方块、模块、电路和步骤的功能对其进行了一般性的描述。这种功能是被实现为软件还是被实现为硬件取决于具体应用以及施加给整个系统的设计约束。本领域技术人员可以针对每种具体应用以各种方式来实现 所述的功能,但是这种实现决定不应被解释为导致脱离本申请实施例公开的范围。Those skilled in the art will also appreciate that the various exemplary logical blocks, modules, circuits, and algorithm steps described in connection with the disclosure herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described generally in terms of their functionality. Whether such functionality is implemented as software or hardware depends on the specific application and design constraints imposed on the overall system. Those skilled in the art may implement the described functions in various ways for each specific application, but such implementation decisions should not be interpreted as causing a departure from the scope of the disclosure of the embodiments of the present application.
以上是本申请公开的示例性实施例,但是应当注意,在不背离权利要求限定的本申请实施例公开的范围的前提下,可以进行多种改变和修改。根据这里描述的公开实施例的方法权利要求的功能、步骤和/或动作不需以任何特定顺序执行。此外,尽管本申请实施例公开的元素可以以个体形式描述或要求,但除非明确限制为单数,也可以理解为多个。The above are exemplary embodiments disclosed in the present application, but it should be noted that various changes and modifications may be made without departing from the scope of the disclosure of the embodiments of the present application defined by the claims. The functions, steps and/or actions of the method claims in accordance with the disclosed embodiments described herein need not be performed in any particular order. Furthermore, although elements disclosed in the embodiments of the present application may be described or claimed in an individual form, unless explicitly limited to the singular, they may also be construed as a plurality.
应当理解的是,在本文中使用的,除非上下文清楚地支持例外情况,单数形式“一个”旨在也包括复数形式。还应当理解的是,在本文中使用的“和/或”是指包括一个或者一个以上相关联地列出的项目的任意和所有可能组合。上述本申请实施例公开实施例序号仅仅为了描述,不代表实施例的优劣。It should be understood that, as used herein, the singular form "a" is intended to include the plural form as well, unless the context clearly supports an exception. It will also be understood that "and/or" as used herein is meant to include any and all possible combinations of one or more of the associated listed items. The above-mentioned embodiments of the present application disclose the serial numbers of the embodiments only for description, and do not represent the advantages and disadvantages of the embodiments.
所属领域的普通技术人员应当理解:以上任何实施例的讨论仅为示例性的,并非旨在暗示本申请实施例公开的范围(包括权利要求)被限于这些例子;在本申请实施例的思路下,以上实施例或者不同实施例中的技术特征之间也可以进行组合,并存在如上的本申请实施例的不同方面的许多其它变化,为了简明它们没有在细节中提供。因此,凡在本申请实施例的精神和原则之内,所做的任何省略、修改、等同替换、改进等,均应包含在本申请实施例的保护范围之内。Those of ordinary skill in the art should understand that the discussion of any of the above embodiments is only exemplary, and is not intended to imply that the scope (including the claims) disclosed by the embodiments of the present application is limited to these examples; under the idea of the embodiments of the present application , the technical features in the above embodiments or different embodiments can also be combined, and there are many other changes in different aspects of the above embodiments of the present application, which are not provided in detail for the sake of brevity. Therefore, any omission, modification, equivalent replacement, improvement, etc. made within the spirit and principle of the embodiments of the present application should be included within the protection scope of the embodiments of the present application.

Claims (11)

  1. 一种基于内容相同的视频过滤方法,其特征在于,包括以下步骤:A kind of video filtering method based on the same content, is characterized in that, comprises the following steps:
    基于目标视频在内容相同视频库中检测并获取对应的相似视频;Detect and obtain the corresponding similar videos in the video library with the same content based on the target video;
    对所述目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图;The target video and the similar video are respectively subjected to timing extraction processing to obtain their respective feature sequence diagrams;
    通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段;Match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through the dynamic time warping algorithm;
    基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像;Based on the duration of the target video, sample extraction processing is performed on the matched video segments, respectively, to obtain n frames of images;
    分别对n帧图像进行图像信息量计算,且分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除。The image information content is calculated for n frames of images respectively, and the video information content of the corresponding target video and similar videos is obtained based on the calculated image information content, and the videos with small video information content are eliminated.
  2. 根据权利要求1所述的方法,其特征在于,对所述目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图包括:The method according to claim 1, characterized in that, performing timing extraction processing on the target video and similar videos respectively, and obtaining respective feature timing diagrams comprising:
    对所述目标视频和相似视频分别定时按帧抽取图像最中间像素点的灰度值,并分别建立其与时间的函数,得到各自的特征时序图。From the target video and the similar video, the gray value of the middle pixel point of the image is periodically extracted by frame, and the function of the gray value and the time is established respectively to obtain the respective feature sequence diagrams.
  3. 根据权利要求1所述的方法,其特征在于,通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段包括:The method according to claim 1, wherein matching the feature sequence diagram of the target video and the feature sequence diagram of the similar video to the same video segment by a dynamic time warping algorithm comprises:
    通过动态时间规整算法选出目标视频的特征时序图和相似视频的特征时序图中有差异的关键帧,并基于所述关键帧得到匹配的相同的视频片段。A dynamic time warping algorithm is used to select key frames with differences between the feature sequence diagram of the target video and the feature sequence diagrams of similar videos, and based on the key frames, matching identical video segments are obtained.
  4. 根据权利要求1所述的方法,其特征在于,基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像包括:The method according to claim 1, wherein the sample extraction processing is performed on the matched video clips based on the duration of the target video, and the n frames of images obtained respectively include:
    建立基于目标视频的时长的抽取频率函数;Establish a decimation frequency function based on the duration of the target video;
    根据所述抽取频率函数对匹配出的视频片段按帧抽取,各自得到n帧 图像。According to the extraction frequency function, the matched video segments are extracted frame by frame, and n frames of images are obtained respectively.
  5. 根据权利要求4所述的方法,其特征在于,建立基于目标视频的时长的抽取频率函数包括:The method according to claim 4, wherein establishing the extraction frequency function based on the duration of the target video comprises:
    将目标视频的时长分为四个时间梯度,分别是:1s至60s、1min至60min、1h至10h及10h以上,对不同的时间梯度建立不同的抽取次数。The duration of the target video is divided into four time gradients, namely: 1s to 60s, 1min to 60min, 1h to 10h, and more than 10h, and different extraction times are established for different time gradients.
  6. 根据权利要求1所述的方法,其特征在于,所述图像信息量计算包括:The method according to claim 1, wherein the calculation of the amount of image information comprises:
    计算图像二维熵作为图像信息量。Calculate the two-dimensional entropy of the image as the amount of image information.
  7. 根据权利要求1所述的方法,其特征在于,分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除包括:The method according to claim 1, wherein, respectively obtaining the video information content of the corresponding target video and the similar video based on the calculated image information content, and removing the videos with small video information content comprises:
    对目标视频的n帧图像信息量求第一平均值,并以第一平均值与目标视频时长的乘积作为其视频信息量;The first average value is obtained to the n-frame image information content of the target video, and the product of the first average value and the target video duration is used as its video information content;
    对相似视频的n帧图像信息量求第二平均值,并以第二平均值与相似视频时长的乘积作为其视频信息量;The second average value is calculated for the n-frame image information content of the similar video, and the product of the second average value and the similar video duration is used as its video information content;
    比较目标视频的视频信息量和相似视频的视频信息量,并将其中视频信息量小的视频剔除。Compare the video information content of the target video with the video information content of similar videos, and remove the videos with less video information.
  8. 根据权利要求1所述的方法,其特征在于,所述方法还包括:The method according to claim 1, wherein the method further comprises:
    响应于视频信息量小的视频被剔除,更新所述内容相同视频库和常规视频库,其中,所述常规视频库用于存放具有单独内容的视频。In response to the video having a small amount of video information being culled, the same-content video library and the regular video library are updated, wherein the regular video library is used to store videos with separate contents.
  9. 根据权利要求7所述的方法,其特征在于,所述比较目标视频的视频信息量和相似视频的视频信息量,并将其中视频信息量小的视频剔除还包括:The method according to claim 7, wherein the comparing the video information content of the target video and the video information content of similar videos, and culling the videos with small video information content further comprises:
    响应于目标视频的视频信息量小于相似视频的视频信息量,将目标视频剔除;In response to the amount of video information of the target video being less than the amount of video information of the similar videos, the target video is eliminated;
    响应于相似视频的视频信息量小于目标视频的视频信息量,将相似视 频剔除。In response to the video information content of the similar videos being less than the video information content of the target video, the similar videos are eliminated.
  10. 一种基于内容相同的视频过滤系统,其特征在于,包括:A video filtering system based on the same content, comprising:
    视频检测模块,配置用于基于目标视频在内容相同视频库中检测并获取对应的相似视频;a video detection module, configured to detect and acquire corresponding similar videos in a video library with the same content based on the target video;
    定时抽取模块,配置用于对所述目标视频和相似视频分别进行定时抽取处理,得到各自的特征时序图;a timing extraction module, configured to perform timing extraction processing on the target video and similar videos respectively to obtain respective feature timing diagrams;
    视频片段匹配模块,配置用于通过动态时间规整算法将目标视频的特征时序图和相似视频的特征时序图匹配出相同的视频片段;The video segment matching module is configured to match the feature sequence diagram of the target video and the feature sequence diagram of similar videos to the same video segment through the dynamic time warping algorithm;
    样本抽取模块,配置用于基于目标视频的时长分别对匹配出的视频片段进行样本抽取处理,各自得到n帧图像;以及a sample extraction module, configured to perform sample extraction processing on the matched video segments based on the duration of the target video, respectively, to obtain n frames of images; and
    视频剔除模块,配置用于分别对n帧图像进行图像信息量计算,且分别基于计算得到的图像信息量得到相应的目标视频和相似视频的视频信息量,并将其中视频信息量小的视频剔除。The video culling module is configured to calculate the image information amount of the n-frame images respectively, and obtain the video information amount of the corresponding target video and similar videos based on the calculated image information amount, and eliminate the video with a small amount of video information. .
  11. 一种计算机设备,包括存储器和处理器,其特征在于,所述存储器中存储有计算机程序,所述计算机程序被所述处理器执行时执行如权利要求1至9任意一项所述的方法。A computer device comprising a memory and a processor, wherein a computer program is stored in the memory, and the computer program is executed by the processor to execute the method according to any one of claims 1 to 9.
PCT/CN2021/121872 2020-12-04 2021-09-29 Video filtering method and system based on identical content, and device WO2022116668A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011406954.4 2020-12-04
CN202011406954.4A CN112653928B (en) 2020-12-04 2020-12-04 Video filtering method, system and equipment based on same content

Publications (1)

Publication Number Publication Date
WO2022116668A1 true WO2022116668A1 (en) 2022-06-09

Family

ID=75350274

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/121872 WO2022116668A1 (en) 2020-12-04 2021-09-29 Video filtering method and system based on identical content, and device

Country Status (2)

Country Link
CN (1) CN112653928B (en)
WO (1) WO2022116668A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112653928B (en) * 2020-12-04 2022-12-02 苏州浪潮智能科技有限公司 Video filtering method, system and equipment based on same content
CN113542771A (en) * 2021-07-15 2021-10-22 广东电网有限责任公司中山供电局 Video high-efficiency compression processing method based on content weight
CN115243073B (en) * 2022-07-22 2024-05-14 腾讯科技(深圳)有限公司 Video processing method, device, equipment and storage medium

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3621323B2 (en) * 2000-02-28 2005-02-16 日本電信電話株式会社 Video registration / search processing method and video search device
US20110150423A1 (en) * 2009-12-18 2011-06-23 Electronics And Telecommunications Research Institute Digital video managing and searching system
CN102779184A (en) * 2012-06-29 2012-11-14 中国科学院自动化研究所 Automatic positioning method of approximately repeated video clips
CN103678702A (en) * 2013-12-30 2014-03-26 优视科技有限公司 Video duplicate removal method and device
CN110620937A (en) * 2019-10-21 2019-12-27 电子科技大学 Dynamic self-adaptive encrypted video traffic identification method based on HTTP
CN110996123A (en) * 2019-12-18 2020-04-10 广州市百果园信息技术有限公司 Video processing method, device, equipment and medium
CN112653928A (en) * 2020-12-04 2021-04-13 苏州浪潮智能科技有限公司 Video filtering method, system and equipment based on same content

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5658430B2 (en) * 2008-08-15 2015-01-28 パナソニックIpマネジメント株式会社 Image processing device
CN104504101B (en) * 2014-12-30 2018-10-30 北京奇艺世纪科技有限公司 A kind of determination method and device of similar video
CN109189991B (en) * 2018-08-17 2021-06-08 百度在线网络技术(北京)有限公司 Duplicate video identification method, device, terminal and computer readable storage medium
CN109726765A (en) * 2019-01-02 2019-05-07 京东方科技集团股份有限公司 A kind of sample extraction method and device of visual classification problem
CN110598014B (en) * 2019-09-27 2021-12-10 腾讯科技(深圳)有限公司 Multimedia data processing method, device and storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3621323B2 (en) * 2000-02-28 2005-02-16 日本電信電話株式会社 Video registration / search processing method and video search device
US20110150423A1 (en) * 2009-12-18 2011-06-23 Electronics And Telecommunications Research Institute Digital video managing and searching system
CN102779184A (en) * 2012-06-29 2012-11-14 中国科学院自动化研究所 Automatic positioning method of approximately repeated video clips
CN103678702A (en) * 2013-12-30 2014-03-26 优视科技有限公司 Video duplicate removal method and device
CN110620937A (en) * 2019-10-21 2019-12-27 电子科技大学 Dynamic self-adaptive encrypted video traffic identification method based on HTTP
CN110996123A (en) * 2019-12-18 2020-04-10 广州市百果园信息技术有限公司 Video processing method, device, equipment and medium
CN112653928A (en) * 2020-12-04 2021-04-13 苏州浪潮智能科技有限公司 Video filtering method, system and equipment based on same content

Also Published As

Publication number Publication date
CN112653928B (en) 2022-12-02
CN112653928A (en) 2021-04-13

Similar Documents

Publication Publication Date Title
WO2022116668A1 (en) Video filtering method and system based on identical content, and device
CN109151501B (en) Video key frame extraction method and device, terminal equipment and storage medium
US11132555B2 (en) Video detection method, server and storage medium
US9036905B2 (en) Training classifiers for deblurring images
US20160196478A1 (en) Image processing method and device
CN112163120A (en) Classification method, terminal and computer storage medium
CN116363554A (en) Method, system, medium, equipment and terminal for extracting key frames of surveillance video
CN110991310A (en) Portrait detection method, portrait detection device, electronic equipment and computer readable medium
CN111222450A (en) Model training method, model training device, model live broadcast processing equipment and storage medium
CN113496208A (en) Video scene classification method and device, storage medium and terminal
US10839251B2 (en) Method and system for implementing image authentication for authenticating persons or items
WO2017070841A1 (en) Image processing method and apparatus
CN108966042B (en) Video abstract generation method and device based on shortest path
CN116431857B (en) Video processing method and system for unmanned scene
CN112001842A (en) Picture generation method and device, electronic equipment and computer readable storage medium
CN112749660B (en) Method and device for generating video content description information
Chathurika et al. A revised averaging algorithm for an effective feature extraction in component-based image retrieval system
CN102495843A (en) Salient region detection algorithm based on local features
CN113095239A (en) Key frame extraction method, terminal and computer readable storage medium
CN116033182B (en) Method and device for determining video cover map, electronic equipment and storage medium
CN114710474B (en) Data stream processing and classifying method based on Internet of things
WO2015054994A1 (en) Video extraction method and device
CN113434731B (en) Music video genre classification method, device, computer equipment and storage medium
CN111598053B (en) Image data processing method and device, medium and system thereof
WO2023056833A1 (en) Background picture generation method and apparatus, image fusion method and apparatus, and electronic device and readable medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21899702

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21899702

Country of ref document: EP

Kind code of ref document: A1