WO2014094492A1 - Method and system for screening depth fusion video - Google Patents

Method and system for screening depth fusion video Download PDF

Info

Publication number
WO2014094492A1
WO2014094492A1 PCT/CN2013/085739 CN2013085739W WO2014094492A1 WO 2014094492 A1 WO2014094492 A1 WO 2014094492A1 CN 2013085739 W CN2013085739 W CN 2013085739W WO 2014094492 A1 WO2014094492 A1 WO 2014094492A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
review
category
fusion
feature
Prior art date
Application number
PCT/CN2013/085739
Other languages
French (fr)
Chinese (zh)
Inventor
朱定局
Original Assignee
深圳先进技术研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳先进技术研究院 filed Critical 深圳先进技术研究院
Publication of WO2014094492A1 publication Critical patent/WO2014094492A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content

Definitions

  • the invention relates to the field of video censorship, in particular to an in-depth fusion video review method and an in-depth integrated video censorship system.
  • the present invention aims to provide an in-depth fusion video review method and an in-depth fusion video review system, which can improve the accuracy of video fusion review.
  • the video frames to be examined are classified by using a preset fusion review classification manner, and the video categories to which the video frames to be examined belong are obtained;
  • An in-depth fusion video review system that includes:
  • the video large class fusion determining module is configured to classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs;
  • a feature extraction module configured to extract various features in the video frame to be examined
  • a video subclass fusion determining module configured to respectively examine, according to various features in the to-be-reviewed video frame, features of the video category, a fusion parameter, and determine that the to-be-reviewed video frame belongs to each of the video categories.
  • the video frames to be examined are classified by using a preset fusion review classification manner, the video categories to which the video frames to be examined belong are obtained, and the fusion parameters are examined based on the characteristics of the video categories. And the various features in the video frame to be examined, determining the possibility that the video frame to be examined belongs to each video subclass under the determined video category, and comprehensively determining based on the possibility of belonging to each video subclass.
  • FIG. 1 is a schematic flow chart of an embodiment of an in-depth fusion video review method of the present invention
  • FIG. 2 is a schematic flow chart of determining feature review fusion parameters of each video category in a specific example
  • FIG. 3 is a schematic flow chart of an in-depth fusion video review in a specific example
  • FIG. 4 is a schematic structural diagram of an embodiment of the in-depth fusion video review system of the present invention.
  • FIG. 1 A schematic flow diagram of an embodiment of the in-depth fusion video review method of the present invention is shown in FIG. As shown in FIG. 1, the method in this embodiment includes the steps of:
  • Step S101 classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs.
  • Step S102 Extract various features in the video frame to be examined
  • Step S103 Examining the fusion parameters according to various features in the video frame to be examined and the characteristics of the video category, and determining the possibility that the video frame to be examined belongs to each video subclass under the video category;
  • Step S104 comprehensively determining the video subclass to which the video frame to be examined belongs according to the possibility that the video frame to be examined belongs to each video subclass under the above video category.
  • the video frame to be examined when reviewing, is classified by using a preset fusion review classification manner, and the video category to which the video frame to be examined belongs is obtained, and then based on the characteristics of the video category. Examining the fusion parameters and various features in the video frame to be examined, determining the possibility that the video frame to be examined belongs to each video subclass under the above determined video category, and based on the possibility of belonging to each video subclass And comprehensively determining the video subclass to which the video frame to be examined belongs.
  • the feature review fusion parameter of the video category may be determined based on the established video sample database.
  • FIG. 2 is a flow chart showing the determination of the feature review fusion parameters of each video category in a specific example.
  • the manner of determining the feature review fusion parameters of each video category in the specific example includes:
  • Step S201 classify each video frame in the video sample database by using the foregoing preset fusion review classification manner, and obtain a video frame of each video category that is merged and reviewed;
  • Step S202 classifying each video frame in the video sample database by using various feature review methods, and respectively obtaining video frames of each video category after each feature review;
  • Step S203 determining, according to the video frames of each video category and the video frames of each video category after each feature review, the accuracy of each feature review of each video category;
  • Step S204 Determine the feature review fusion parameters of each video category according to the accuracy of each type of feature review of each video category.
  • the following takes an example of determining the fusion parameters of each video category in FIG. 2 as an example, and a specific example thereof is described in detail.
  • the video major categories, various features, and various feature review methods may be different based on actual needs.
  • the video category includes yellow video, violent video, and reaction video as examples, and various features including text, sound, and image, and various feature reviews include text feature review and voice feature review.
  • the image feature review is described as an example, and the description is merely an exemplary description and is not intended to limit the present invention.
  • a schematic flow chart of this specific example is shown in FIG.
  • a certain number of video samples may be pre-stored in the video sample database, that is, a certain number of video frames are stored in the video sample database, and subsequent video categories are determined.
  • Feature review fusion parameters are described in conjunction with video frames in a video sample database.
  • the video frames in the video sample database are classified by using the preset fusion review classification method, and the video frames of each video category that are merged and examined are obtained, that is, video frames belonging to the yellow video and videos belonging to the violent video.
  • a frame a video frame belonging to a reaction video.
  • the video frames of the classified video categories can be respectively placed in the library of each video category of the fusion review, that is, the video frames of the classified yellow video. Put the video frame of the classified violence video into the violent video library (recorded as RB) of the merged review, and put the video frame of the classified reaction video into the yellow video library (recorded as RH).
  • Fusion review of the reactionary video library (marked as RF).
  • the above-mentioned preset fusion review classification method may be performed by any one of the existing and subsequent methods, as long as the video frame belongs to any of the yellow video, the violent video, the reaction video, and the like. Yes, we will not go into details here.
  • the video frames in the video sample database are classified by using various feature review methods, and the video frames of each video category after each feature review are respectively obtained.
  • various types of features include characters, sounds, and images
  • the details may be as follows.
  • the video feature frame is used to classify each video frame in the video sample video database, and the video frames of each video category that are subject to text feature review are obtained, that is, video frames belonging to the yellow video, video frames belonging to the violent video, and belonging to the reaction video.
  • Video frame After obtaining the video frames of each video category of the character feature review, the classified video frames of each video category can be respectively placed in the library of each video category of the character feature review, that is, the yellow video of the classified video.
  • the video frame is placed in the yellow video library (recorded as WH) of the text feature review, and the video frame of the classified violent video is placed in the violent video library of the character feature review (denoted as WB), and the classified reaction video is The video frame is placed in the reaction video library (denoted as WF) for text feature review.
  • the specific text feature review method may be performed in any manner that is currently available and may occur in the future, and will not be described in detail herein.
  • the video feature frame is used to classify each video frame in the video sample video database, and the video frames of each video category of the sound feature review are obtained, that is, the video frames belonging to the yellow video, the video frames belonging to the violent video, and the reaction video.
  • Video frame After obtaining the video frames of each video category of the sound feature review, the classified video frames of each video category can be respectively placed in the library of the video categories of the sound feature review, that is, the yellow video of the classified video.
  • the video frame is placed in the yellow video library (recorded as VH) of the sound feature review, and the video frame of the classified violent video is placed in the violent video library (recorded as VB) of the sound feature review, and the classified reaction video is The video frame is placed in the reaction video library (denoted as VF) of the sound feature review.
  • VH yellow video library
  • VB violent video library
  • VF reaction video library
  • the image feature review mode is used to classify each video frame in the video sample video database, and the video frames of each video category of the image feature review are obtained, that is, the video frames belonging to the yellow video, the video frames belonging to the violent video, and the reaction video.
  • Video frame After obtaining the video frames of each video category of the image feature review, the classified video frames of each video category may be respectively placed in a library of each video category of the image feature review, that is, the yellow video of the classified video.
  • the video frame is placed in the yellow video library (recorded as GH) of the image feature review, and the video frame of the classified violent video is placed in the violent video library (recorded as GB) of the image feature review, and the classified reaction video is The video frame is placed in the reaction video library (denoted as GF) of the image feature review.
  • the specific image feature review mode may be performed in any manner that is currently available and may occur in the future, and will not be described in detail herein.
  • the specific manner for determining the accuracy of each type of feature review of each video category may be:
  • the accuracy of the current class feature review of the current video category is determined.
  • each video category includes a yellow video, a violent video, and a reaction video
  • various feature reviews performed include text feature review, sound feature review, and image feature review. Therefore, in the end, the accuracy of the character review of the yellow video (violent video, reactionary video), the accuracy of the sound feature review, and the accuracy of the image feature review can be obtained, and a total of nine accuracy rates are obtained.
  • the specific process may be as follows.
  • the accuracy rate of the text review of the violent video can be determined by the violent video library RB of the fusion review, and the false positive rate and the missed rate of the violent video library WB of the text review are determined, and then based on the false positive rate and the leak rate.
  • the judgment rate comprehensively determines the accuracy of the text review of the violent video.
  • the false positive rate determine the number of video frames belonging to the violent video library WB of the text review, but not the violent video library RB of the merged review, and divide the number by the number of video frames of the violent video library RB of the merged review.
  • the value obtained is used as the false positive rate for violent video text review, namely:
  • the false positive rate of violent video text review (the number of video frames belonging to the violent video library WB of the text review, but not belonging to the violent video library RB of the fusion review) / (the number of video frames of the violent video library RB of the merged review) ).
  • the missed rate the number of video frames belonging to the violent video library RB of the merge review but not the violent video library WB of the text review is determined, and the number is divided by the number of video frames of the violent video library RB of the merged review, The resulting value is used as a missed rate for violent video text review, ie:
  • Missing rate of violent video text review (number of video frames belonging to the violent video library RB of the merge review but not violent video library WB of the text review) / (number of video frames of the bred video library RB of the merged review) ).
  • the accuracy of the violent video text review is comprehensively determined.
  • the false positive rate of the violent video text review the larger value, the smaller value, the average value, the weighted average value of the violent video text review, or the value calculated by other methods may be used as the misjudgment rate of the violent video text review.
  • the accuracy of the violent video text review, the specific comprehensive determination method can be different according to the actual application needs.
  • the accuracy of violent video sound review and the accuracy of violent video image review are similar to the above-described methods for determining the accuracy of violent video text review.
  • the violent video library RB of the fusion review may be used as the standard to determine the false positive rate and the missed rate of the violent video library VB of the voice review, and then based on the false positive rate and the missed rate. Comprehensively determine the accuracy of the sound review of violent videos.
  • the false positive rate determine the number of video frames belonging to the violent video library VB of the sound review, but not the violent video library RB of the fusion review, and divide the number by the number of video frames of the violent video library RB of the fusion review. , the value obtained as the false positive rate of violent video sound review, namely:
  • the false positive rate of violent video sound review (the number of video frames belonging to the violent video library VB of the sound review, but not belonging to the violent video library RB of the fusion review) / (the number of video frames of the violent video library RB of the merged review) ).
  • the missed rate the number of video frames belonging to the violent video library RB of the fusion review but not the violent video library VB of the sound review is determined, and the number is divided by the number of video frames of the violent video library RB of the merged review, The value obtained is used as the missed rate of violent video sound review, namely:
  • Missing rate of violent video sound review (number of video frames belonging to the violent video library RB of the fusion review but not violent video library VB of the sound review) / (number of video frames of the bred video library RB of the merged review) ).
  • the accuracy of the violent video sound review is determined.
  • the false positive rate of the violent video sound review the larger value, the smaller value, the average value, the weighted average value or the value calculated by other methods may be used as the false positive rate of the violent video sound review.
  • the accuracy of violent video sound review, the specific comprehensive determination method can be different according to the actual application needs.
  • the violent video library RB of the fusion review may be used as a standard to determine the false positive rate and the missed rate of the violent video library GB of the image review, and then based on the false positive rate and the missed rate. Comprehensively determine the accuracy of image review of violent videos.
  • the false positive rate determine the number of video frames belonging to the violent video library GB of the image review, but not belonging to the violent video library RB of the merge review, and divide the number by the number of video frames of the violent video library RB of the merged review. , the value obtained as the false positive rate of violent video image review, namely:
  • the false positive rate of violent video image review (the number of video frames belonging to the violent video library GB of the image review, but not belonging to the violent video library RB of the fusion review) / (the number of video frames of the violent video library RB of the merged review) ).
  • the number of video frames belonging to the violent video library RB of the merge review, but not the violent video library GB of the image review is determined, and the number is divided by the number of video frames of the violent video library RB of the merged review, The resulting value is used as a missed rate for violent video image review, namely:
  • Missing rate of violent video image review (the number of video frames belonging to the violent video library RB of the merged review but not belonging to the violent video library GB of the image review) / (the number of video frames of the violent video library RB of the merged review) ).
  • the accuracy of the violent video image review is comprehensively determined.
  • the false positive rate of the violent video image review the larger value, the smaller value, the average value, the weighted average value of the violent video image review, or the value calculated by other methods may be used as the misjudgment rate of the violent video image review.
  • the accuracy of violent video image review, the specific comprehensive determination method can be different according to the actual application needs.
  • the feature review fusion parameters of each video category can be comprehensively determined.
  • the accuracy rate of the various features of the above violent video (including the accuracy of violent video text review, the accuracy of violent video sound review, and the accuracy of violent video image review), for example, the characteristics of the violent video can be recorded.
  • rw, rv, rg where rw represents the fusion parameter of the violent video text review, rv represents the fusion parameter of the violent video sound review, and rg represents the fusion parameter of the violent video image review.
  • the parameters rw, rv, rg can be determined in the following manners, respectively:
  • Rw Accuracy rate of violent video text review/(accuracy of violent video text review+accuracy of violent video sound review, accuracy of violent video image review);
  • Rv Accuracy rate of violent video sound review/(accuracy of violent video text review+accuracy of violent video sound review, accuracy of violent video image review);
  • Rg Accuracy of violent video image review/(accuracy of violent video text review+accuracy of violent video sound review, accuracy of violent video image review).
  • the specific comprehensive determination may be similar to the determination method of the feature fusion parameter of the above violent video, and details are not described herein.
  • the feature fusion parameters of each of the video categories obtained above can be stored for subsequent review of the fusion of the video frames to be reviewed.
  • the video frame to be examined may be classified by using the foregoing preset fusion review classification manner, and the video category to which the video frame to be examined belongs belongs.
  • the video to which the subject to be reviewed belongs is a violent video.
  • the corresponding various types of features are extracted from the to-be-reviewed video frame, and specifically may include a text feature, a sound feature, and an image feature.
  • the violent video to which the video frame to be examined belongs belongs to the i-class
  • the violent video is further divided into N subclasses, which are respectively recorded as i1, i2, i3, ..., iN.
  • the character feature of the video frame to be examined it is determined that the video frame to be examined or the character feature belongs to the i1th subclass, the possibility wi1, the possibility i2 belonging to the i2th subclass, and belongs to the i3th subclass.
  • wi1+wi2+wi3+...+wiN 1.
  • the video frames to be examined belong to the subcategories of the violent video:
  • pi1 rw*wi1+rv+vi1+rg*gi1;
  • piN rw*wiN+rv+viN+rg*giN.
  • the video frame to be examined may be comprehensively determined.
  • Video subclass In general, the video subclass corresponding to the maximum value of pi1, pi2, pi3, ..., piN may be determined as the video subclass to which the video frame to be examined belongs.
  • the present invention also provides an in-depth fusion video review system.
  • the following describes an embodiment of the in-depth fusion video review system of the present invention.
  • FIG. 4 A block diagram of an embodiment of the in-depth fusion video review system of the present invention is shown in FIG. 4, the in-depth fusion video review system in this embodiment includes:
  • the video major class fusion determining module 401 is configured to classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs.
  • the feature extraction module 402 is configured to extract various features in the video frame to be examined
  • the video subclass fusion determining module 403 is configured to respectively review the fusion parameters according to the various features in the to-be-reviewed video frame and the features of the video category, and determine that the to-be-reviewed video frame belongs to the video category.
  • the video frame to be examined when reviewing, is classified by using a preset fusion review classification manner, and the video category to which the video frame to be examined belongs is obtained, and then based on the characteristics of the video category. Examining the fusion parameters and various features in the video frame to be examined, determining the possibility that the video frame to be examined belongs to each video subclass under the above determined video category, and based on the possibility of belonging to each video subclass And comprehensively determining the video subclass to which the video frame to be examined belongs.
  • the in-depth fusion video review system in this embodiment may further include: a fusion parameter determination module 404 for determining feature review fusion parameters of the video categories.
  • the fusion parameter determination module 404 includes:
  • the sample fusion review module 4041 is configured to classify each video frame in the video sample database by using the preset fusion review classification manner, and obtain a video frame of each video category that is merged and reviewed;
  • the sample classification review module 4042 is configured to classify each video frame in the video sample database by using various feature review methods, and obtain video frames of each video category after each feature review;
  • the sample accuracy determining module 4043 is configured to determine the accuracy of each type of feature review of each video category according to the video frames of each video category and the video frames of each video category after the feature review. ;
  • the fusion parameter comprehensive determination module 4044 determines the feature review fusion parameters of each video category according to the accuracy of each type of feature review of each video category.
  • the fusion parameter integration determining module 4044 may be a current ratio of the accuracy of the current class feature review of the current video category to the sum of the accuracy of the current video category.
  • the fusion parameters of the current class feature review of the video category, the feature fusion review parameters of the current video category include the fusion parameters of various feature reviews of the current video category.
  • the sample accuracy determining module 4043 may specifically include:
  • the false positive rate determining module 40431 is configured to respectively acquire the first video frame number of the video frame belonging to the current video category of the current class feature review but not belonging to the current video category of the merge review, and the first video frame The number is divided by the value of the sample number of the video frame of the current video category of the fusion review as the false positive rate of the current class feature review;
  • the missed rate determining module 40432 is configured to respectively acquire the number of second video frames belonging to the current video category of the merged review but not the current video category of the current class feature review, and the second video frame The number and the value of the number of samples of the video frame divided by the current video category of the fusion review are used as the missed rate of the current class feature review;
  • the accuracy rate determining module 40433 is configured to determine an accuracy rate of the current class feature review of the current video category according to the false positive rate and the missed rate of the current class feature review.
  • the accuracy determination module 40433 may use the error rate of the current class feature review, the average value of the missed rate, or the weighted average as the accuracy of the current class feature review of the current video category.
  • the video subclass fusion determining module 403 may specifically include:
  • the feature subclass likelihood determining module 4031 is configured to determine, respectively, the possibility that each type of feature in the video frame to be examined belongs to each video subclass under the video category;
  • the video subclass likelihood determining module 4032 is configured to determine the video frame to be reviewed according to the possibility that each type of feature belongs to each video subclass under the video category, and the feature review fusion parameter of the video category The possibility of belonging to each video subclass under the video category;
  • a subclass determining module configured to determine a maximum value of a likelihood that the video frame to be examined belongs to each video subclass under the video category, and determine a video subclass corresponding to the maximum value of the likelihood as the The video subclass to which the video frame to be reviewed belongs.
  • each module in the in-depth fusion video review system of the present invention may be the same as that in the above-mentioned in-depth fusion video review method of the present invention, and details are not described herein.

Abstract

A method and system for screening a depth fusion video, the method comprising the steps of: classifying a to-be-screened video frame by using a preset fusion screening classification method to acquire a video class of the to-be-screened video frame; extracting various characteristics of the to-be-screened video frame; determining the possibility that the to-be-screened video frame belongs to each video subclass of the video class according to various characteristics of the to-be-screened video frame and the characteristic screening fusion parameters of the video class respectively; and comprehensively determining the video subclass of the to-be-screened video frame according to the possibility that the to-be-screened video frame belongs to each video subclass of the video class. Based on various characteristics of the to-be-screened video frame and the characteristic screening fusion parameters of the video class, the solution of the present invention comprehensively considers the effects of various characteristics of a video frame on the video frames in the video class, and differentiates the effects of different characteristics on different types of video frames, thus improving video screening accuracy.

Description

深入融合视频审查方法和系统In-depth integration of video review methods and systems
【技术领域】[Technical Field]
本发明涉及视频审查领域,特别涉及一种深入融合视频审查方法、一种深入融合视频审查系统。The invention relates to the field of video censorship, in particular to an in-depth fusion video review method and an in-depth integrated video censorship system.
【背景技术】【Background technique】
视频内容的应用日益广泛,对视频内容的审查成为针对视频内容的处理中的重要的一部分。目前针对视频内容的审查,最普通、最简单的方式是通过人眼观看的方式,通过将视频文件的内容从头到尾观看一遍,据此审查视频内容是否是限制性发布或者不允许发布的视频内容。作为对这种人工审查方式的改进,出现了进行视频融合的审查方式,这种视频融合的审查方式中,是按照固定的方式对听觉、视觉等特征进行融合,例如采用加权平均进行融合。采用加权平均进行融合的方式中,假设该视频内容在文字上有60%的可能性为不雅裸露视频,在声音上有90%的可能性为不雅裸露视频,在图像上有30%的可能性为不雅裸露视频,则通过加权综合判定该视频内容有(60%+90%+30%)/3的可能性为不雅裸露视频。The application of video content is becoming more and more extensive, and the review of video content becomes an important part of the processing of video content. At present, the most common and simple way to review video content is to watch the content of the video file from beginning to end by human eyes, and check whether the video content is restricted or not allowed. content. As an improvement to this manual review method, a video fusion review method has emerged. In this video fusion review method, features such as auditory and visual are integrated in a fixed manner, for example, using a weighted average for fusion. In the way of using the weighted average for fusion, it is assumed that the video content has a 60% chance of being invisible to the video, and 90% of the sound is indecent to expose the video, and 30% of the image is on the image. The possibility is that the video is indecent, and the possibility of (60%+90%+30%)/3 of the video content is determined by weighted synthesis as an indecent naked video.
在目前的这种针对视频内容的融合审查中,是按照固定的方式对听觉、视觉等特征进行融合,而实际上,对不同类型的视频来说,视觉、听觉特征的明显度、可判度,例如文字、图像、声音特征的明显度与可判度,对视频审查中的可信度的共享是不同的。例如,声音在枪击类视频中所起的作用、融合时所占的比例应当远大于在裸体类视频中的作用与比例,而图像在裸体类视频中所起的作用、融合时所占的比例应当远大于在反动言论类视频中所起的作用与比例,而文字在反动言论类视频中所起的作用、融合时所占的比例应远大于斗殴类视频中所起的作用与比例。目前的针对视频内容的审查,并未对不同特征对不同类型视频文件的作用加以区分,从而导致审查的准确度大大降低。In the current fusion review of video content, features such as auditory and visual are merged in a fixed manner. In fact, for different types of video, the visibility and judgment of visual and auditory features are different. For example, the visibility and verifiability of text, image, and sound features are different for the sharing of credibility in video censorship. For example, the role played by sound in shooting video, the proportion of fusion should be much larger than the role and proportion in nude video, and the role of images in nude video, the proportion of fusion It should be much larger than the role and proportion in the reactionary speech class, and the role played by the text in the reactionary video should be much larger than the role and proportion of the video in the fight. The current review of video content does not distinguish between the different features of different types of video files, resulting in greatly reduced accuracy of the review.
【发明内容】[Summary of the Invention]
基于此,针对上述现有技术中存在的问题,本发明的目的在于提供一种深入融合视频审查方法、深入融合视频审查系统,其可以提高视频融合审查的准确度。Based on this, in view of the above problems in the prior art, the present invention aims to provide an in-depth fusion video review method and an in-depth fusion video review system, which can improve the accuracy of video fusion review.
为达到上述目的,本发明采用以下技术方案:In order to achieve the above object, the present invention adopts the following technical solutions:
一种深入融合视频审查方法,包括步骤:An in-depth fusion of video review methods, including steps:
采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类;The video frames to be examined are classified by using a preset fusion review classification manner, and the video categories to which the video frames to be examined belong are obtained;
提取所述待审查视频帧中的各类特征;Extracting various features in the video frame to be examined;
分别根据所述待审查视频帧中的各类特征、所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性;Determining, according to the various features in the to-be-reviewed video frame, and the feature of the video category, the fusion parameters, and determining the possibility that the video frame to be examined belongs to each video subclass under the video category;
根据所述待审查视频帧属于所述视频大类下的各视频小类的可能性,综合确定所述待审查视频帧所属的视频小类。Determining, according to the possibility that the video frame to be examined belongs to each video subclass under the video category, comprehensively determining a video subclass to which the to-be-reviewed video frame belongs.
一种深入融合视频审查系统,包括:An in-depth fusion video review system that includes:
视频大类融合确定模块,用于采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类;The video large class fusion determining module is configured to classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs;
特征提取模块,用于提取所述待审查视频帧中的各类特征;a feature extraction module, configured to extract various features in the video frame to be examined;
视频小类融合确定模块,用于分别根据所述待审查视频帧中的各类特征、所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性,并根据所述待审查视频帧属于所述视频大类下的各视频小类的可能性,综合确定所述待审查视频帧所属的视频小类。a video subclass fusion determining module, configured to respectively examine, according to various features in the to-be-reviewed video frame, features of the video category, a fusion parameter, and determine that the to-be-reviewed video frame belongs to each of the video categories. The possibility of the video subclass, and comprehensively determining the video subclass to which the video frame to be examined belongs according to the possibility that the video frame to be examined belongs to each video subclass under the video category.
根据发明方案,其在进行审查时,是先采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类,再基于该视频大类的特征审查融合参数、以及该待审查视频帧中的各类特征,确定出该待审查视频帧属于上述确定的视频大类下的各视频小类的可能性,并基于属于各视频小类的可能性,综合确定出上述待审查视频帧所属的视频小类。其基于该待审查视频帧中的各类特征、以及视频大类的特征审查融合参数,综合考虑了视频帧中的各类特征在该视频大类中的视频帧中所起的作用,对不同类特征对不同类型视频帧的作用加以区分,提高了视频审查的准确度。According to the invention, when reviewing, the video frames to be examined are classified by using a preset fusion review classification manner, the video categories to which the video frames to be examined belong are obtained, and the fusion parameters are examined based on the characteristics of the video categories. And the various features in the video frame to be examined, determining the possibility that the video frame to be examined belongs to each video subclass under the determined video category, and comprehensively determining based on the possibility of belonging to each video subclass. The video subclass to which the video frame to be examined belongs. It examines the fusion parameters based on various features in the video frame to be examined and the features of the video category, and comprehensively considers the role of various features in the video frame in the video frames in the video category, and different Class features distinguish between the effects of different types of video frames, improving the accuracy of video censorship.
【附图说明】[Description of the Drawings]
图1是本发明的深入融合视频审查方法实施例的流程示意图;1 is a schematic flow chart of an embodiment of an in-depth fusion video review method of the present invention;
图2是一个具体示例中确定各视频大类的特征审查融合参数的流程示意图;2 is a schematic flow chart of determining feature review fusion parameters of each video category in a specific example;
图3是一个具体示例中的深入融合视频审查的流程示意图;3 is a schematic flow chart of an in-depth fusion video review in a specific example;
图4是本发明的深入融合视频审查系统实施例的结构示意图。4 is a schematic structural diagram of an embodiment of the in-depth fusion video review system of the present invention.
【具体实施方式】 【detailed description】
以下结合其中的较佳实施例对本发明方案进行详细说明。在下述说明中,先对本发明的深入融合视频审查方法的实施例进行说明,再对本发明的深入融合视频审查系统的实施例进行说明。The solution of the present invention will be described in detail below in conjunction with the preferred embodiments thereof. In the following description, an embodiment of the in-depth fusion video review method of the present invention will be described first, and an embodiment of the in-depth fusion video review system of the present invention will be described.
图1中示出了本发明的深入融合视频审查方法实施例的流程示意图。如图1所示,本实施例中的方法包括步骤:A schematic flow diagram of an embodiment of the in-depth fusion video review method of the present invention is shown in FIG. As shown in FIG. 1, the method in this embodiment includes the steps of:
步骤S101:采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类;Step S101: classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs.
步骤S102:提取待审查视频帧中的各类特征;Step S102: Extract various features in the video frame to be examined;
步骤S103:分别根据待审查视频帧中的各类特征、上述视频大类的特征审查融合参数,确定待审查视频帧属于上述视频大类下的各视频小类的可能性;Step S103: Examining the fusion parameters according to various features in the video frame to be examined and the characteristics of the video category, and determining the possibility that the video frame to be examined belongs to each video subclass under the video category;
步骤S104:根据待审查视频帧属于上述视频大类下的各视频小类的可能性,综合确定待审查视频帧所属的视频小类。Step S104: comprehensively determining the video subclass to which the video frame to be examined belongs according to the possibility that the video frame to be examined belongs to each video subclass under the above video category.
根据本实施例中的方案,其在进行审查时,是先采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类,再基于该视频大类的特征审查融合参数、以及该待审查视频帧中的各类特征,确定出该待审查视频帧属于上述确定的视频大类下的各视频小类的可能性,并基于属于各视频小类的可能性,综合确定出上述待审查视频帧所属的视频小类。其基于该待审查视频帧中的各类特征、以及视频大类的特征审查融合参数,综合考虑了视频帧中的各类特征在该视频大类中的视频帧中所起的作用,对不同类特征对不同类型视频帧的作用加以区分,提高了视频审查的准确度。According to the solution in this embodiment, when reviewing, the video frame to be examined is classified by using a preset fusion review classification manner, and the video category to which the video frame to be examined belongs is obtained, and then based on the characteristics of the video category. Examining the fusion parameters and various features in the video frame to be examined, determining the possibility that the video frame to be examined belongs to each video subclass under the above determined video category, and based on the possibility of belonging to each video subclass And comprehensively determining the video subclass to which the video frame to be examined belongs. It examines the fusion parameters based on various features in the video frame to be examined and the features of the video category, and comprehensively considers the role of various features in the video frame in the video frames in the video category, and different Class features distinguish between the effects of different types of video frames, improving the accuracy of video censorship.
其中,在其中一个具体实现方式中,上述视频大类的特征审查融合参数,可以是基于建立的视频样本数据库来确定。图2中示出了一个具体示例中确定各视频大类的特征审查融合参数的流程示意图。In one specific implementation manner, the feature review fusion parameter of the video category may be determined based on the established video sample database. FIG. 2 is a flow chart showing the determination of the feature review fusion parameters of each video category in a specific example.
如图2所示,该具体示例中确定各视频大类的特征审查融合参数的方式包括:As shown in FIG. 2, the manner of determining the feature review fusion parameters of each video category in the specific example includes:
步骤S201:采用上述预设融合审查分类方式对视频样本数据库中的各视频帧进行分类,获得融合审查的各视频大类的视频帧;Step S201: classify each video frame in the video sample database by using the foregoing preset fusion review classification manner, and obtain a video frame of each video category that is merged and reviewed;
步骤S202:分别采用各类特征审查方法对上述视频样本数据库中的各视频帧进行分类,分别获得各类特征审查后的各视频大类的视频帧;Step S202: classifying each video frame in the video sample database by using various feature review methods, and respectively obtaining video frames of each video category after each feature review;
步骤S203:根据上述融合审查的各视频大类的视频帧、各类特征审查后的各视频大类的视频帧,确定各视频大类的各类特征审查的准确率;Step S203: determining, according to the video frames of each video category and the video frames of each video category after each feature review, the accuracy of each feature review of each video category;
步骤S204:根据各视频大类的各类特征审查的准确率,确定各视频大类的特征审查融合参数。Step S204: Determine the feature review fusion parameters of each video category according to the accuracy of each type of feature review of each video category.
以下结合图2中的确定各视频大类的特征审查融合参数为例,就其中的一个具体示例进行详细说明。The following takes an example of determining the fusion parameters of each video category in FIG. 2 as an example, and a specific example thereof is described in detail.
在本发明方案中,各视频大类、各类特征、各类特征审查方法基于实际需要的设定可以有所不同。在本发明的该具体示例中,以视频大类包括有黄色视频、暴力视频、反动视频为例,以各类特征包括有文字、声音、图像,各类特征审查包括文字特征审查、声音特征审查、图像特征审查为例进行说明,这种说明仅仅只是一种示例性的说明,并不用以对本发明方案进行限定。图3中示出了该具体示例的流程示意图。In the solution of the present invention, the video major categories, various features, and various feature review methods may be different based on actual needs. In this specific example of the present invention, the video category includes yellow video, violent video, and reaction video as examples, and various features including text, sound, and image, and various feature reviews include text feature review and voice feature review. The image feature review is described as an example, and the description is merely an exemplary description and is not intended to limit the present invention. A schematic flow chart of this specific example is shown in FIG.
在确定各视频大类的特征审查融合参数之前,在视频样本数据库中可预先存储有一定数量的视频样本,即在视频样本数据库中存储有一定数量的视频帧,后续的确定各视频大类的特征审查融合参数是结合视频样本数据库中的视频帧进行说明。Before determining the feature review fusion parameters of each video category, a certain number of video samples may be pre-stored in the video sample database, that is, a certain number of video frames are stored in the video sample database, and subsequent video categories are determined. Feature review fusion parameters are described in conjunction with video frames in a video sample database.
然后,首先,采用上述预设融合审查分类方式对视频样本数据库中的各视频帧进行分类,获得融合审查的各视频大类的视频帧,即分别属于黄色视频的视频帧、属于暴力视频的视频帧、属于反动视频的视频帧。在获得融合审查的各视频大类的视频帧之后,可以将分类后的各视频大类的视频帧分别放入在融合审查的各视频大类的库中,即将分类后的黄色视频的视频帧放入融合审查的黄色视频库(记为RH)中,将分类后的暴力视频的视频帧放入融合审查的暴力视频库(记为RB)中,将分类后的反动视频的视频帧放入融合审查的反动视频库(记为RF)中。其中,上述预设融合审查分类方式,可以是采用目前已有以及以后出现的任何一种方式进行,只要能够对视频帧是属于黄色视频、暴力视频、反动视频等当中的哪种视频大类即可,在此不予详加赘述。Then, firstly, the video frames in the video sample database are classified by using the preset fusion review classification method, and the video frames of each video category that are merged and examined are obtained, that is, video frames belonging to the yellow video and videos belonging to the violent video. A frame, a video frame belonging to a reaction video. After obtaining the video frames of each video category that are merged and reviewed, the video frames of the classified video categories can be respectively placed in the library of each video category of the fusion review, that is, the video frames of the classified yellow video. Put the video frame of the classified violence video into the violent video library (recorded as RB) of the merged review, and put the video frame of the classified reaction video into the yellow video library (recorded as RH). Fusion review of the reactionary video library (marked as RF). The above-mentioned preset fusion review classification method may be performed by any one of the existing and subsequent methods, as long as the video frame belongs to any of the yellow video, the violent video, the reaction video, and the like. Yes, we will not go into details here.
然后分别采用各类特征审查方法对上述视频样本数据库中的视频帧进行分类,分别获得各类特征审查后的各视频大类的视频帧。以上述各类特征包括文字、声音、图像的情况下,具体可以是如下所述。Then, the video frames in the video sample database are classified by using various feature review methods, and the video frames of each video category after each feature review are respectively obtained. In the case where the above various types of features include characters, sounds, and images, the details may be as follows.
采用文字特征审查方式对视频样本视频数据库中的各视频帧进行分类,获得文字特征审查的各视频大类的视频帧,即分别属于黄色视频的视频帧、属于暴力视频的视频帧、属于反动视频的视频帧。在获得文字特征审查的各视频大类的视频帧之后,可以将分类后的各视频大类的视频帧分别放入在文字特征审查的各视频大类的库中,即将分类后的黄色视频的视频帧放入文字特征审查的黄色视频库(记为WH)中,将分类后的暴力视频的视频帧放入文字特征审查的暴力视频库(记为WB)中,将分类后的反动视频的视频帧放入文字特征审查的反动视频库(记为WF)中。其中,具体的文字特征审查方式,可以是采用目前已有的以及以后可能出现的任何一种方式进行,在此不予详加赘述。The video feature frame is used to classify each video frame in the video sample video database, and the video frames of each video category that are subject to text feature review are obtained, that is, video frames belonging to the yellow video, video frames belonging to the violent video, and belonging to the reaction video. Video frame. After obtaining the video frames of each video category of the character feature review, the classified video frames of each video category can be respectively placed in the library of each video category of the character feature review, that is, the yellow video of the classified video. The video frame is placed in the yellow video library (recorded as WH) of the text feature review, and the video frame of the classified violent video is placed in the violent video library of the character feature review (denoted as WB), and the classified reaction video is The video frame is placed in the reaction video library (denoted as WF) for text feature review. The specific text feature review method may be performed in any manner that is currently available and may occur in the future, and will not be described in detail herein.
采用声音特征审查方式对视频样本视频数据库中的各视频帧进行分类,获得声音特征审查的各视频大类的视频帧,即分别属于黄色视频的视频帧、属于暴力视频的视频帧、属于反动视频的视频帧。在获得声音特征审查的各视频大类的视频帧之后,可以将分类后的各视频大类的视频帧分别放入在声音特征审查的各视频大类的库中,即将分类后的黄色视频的视频帧放入声音特征审查的黄色视频库(记为VH)中,将分类后的暴力视频的视频帧放入声音特征审查的暴力视频库(记为VB)中,将分类后的反动视频的视频帧放入声音特征审查的反动视频库(记为VF)中。其中,具体的文字特征审查方式,可以是采用目前已有的以及以后可能出现的任何一种方式进行,在此不予详加赘述。The video feature frame is used to classify each video frame in the video sample video database, and the video frames of each video category of the sound feature review are obtained, that is, the video frames belonging to the yellow video, the video frames belonging to the violent video, and the reaction video. Video frame. After obtaining the video frames of each video category of the sound feature review, the classified video frames of each video category can be respectively placed in the library of the video categories of the sound feature review, that is, the yellow video of the classified video. The video frame is placed in the yellow video library (recorded as VH) of the sound feature review, and the video frame of the classified violent video is placed in the violent video library (recorded as VB) of the sound feature review, and the classified reaction video is The video frame is placed in the reaction video library (denoted as VF) of the sound feature review. The specific text feature review method may be performed in any manner that is currently available and may occur in the future, and will not be described in detail herein.
采用图像特征审查方式对视频样本视频数据库中的各视频帧进行分类,获得图像特征审查的各视频大类的视频帧,即分别属于黄色视频的视频帧、属于暴力视频的视频帧、属于反动视频的视频帧。在获得图像特征审查的各视频大类的视频帧之后,可以将分类后的各视频大类的视频帧分别放入在图像特征审查的各视频大类的库中,即将分类后的黄色视频的视频帧放入图像特征审查的黄色视频库(记为GH)中,将分类后的暴力视频的视频帧放入图像特征审查的暴力视频库(记为GB)中,将分类后的反动视频的视频帧放入图像特征审查的反动视频库(记为GF)中。其中,具体的图像特征审查方式,可以是采用目前已有的以及以后可能出现的任何一种方式进行,在此不予详加赘述。The image feature review mode is used to classify each video frame in the video sample video database, and the video frames of each video category of the image feature review are obtained, that is, the video frames belonging to the yellow video, the video frames belonging to the violent video, and the reaction video. Video frame. After obtaining the video frames of each video category of the image feature review, the classified video frames of each video category may be respectively placed in a library of each video category of the image feature review, that is, the yellow video of the classified video. The video frame is placed in the yellow video library (recorded as GH) of the image feature review, and the video frame of the classified violent video is placed in the violent video library (recorded as GB) of the image feature review, and the classified reaction video is The video frame is placed in the reaction video library (denoted as GF) of the image feature review. The specific image feature review mode may be performed in any manner that is currently available and may occur in the future, and will not be described in detail herein.
然后,根据融合审查的各视频大类的视频帧、各类特征审查后的各视频大类的视频帧,确定各视频大类的各类特征审查的准确率。在其中一个具体实现方式中,具体的确定各视频大类的各类特征审查的准确率的方式可以是:Then, according to the video frames of each video category and the video frames of each video category after the feature review, the accuracy of each feature review of each video category is determined. In one specific implementation manner, the specific manner for determining the accuracy of each type of feature review of each video category may be:
分别获取属于当前类特征审查的当前视频大类、但不属于融合审查的当前视频大类的视频帧的第一视频帧数目;Obtaining, respectively, the number of first video frames of the video frame belonging to the current video category of the current class feature review but not belonging to the current video category of the fusion review;
将第一视频帧数目除以融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的误判率;Dividing the number of first video frames by the value of the number of samples of the video frame of the current video category of the merged review as the false positive rate of the current class feature review;
分别获取属于融合审查的当前视频大类的视频帧、但不属于当前类特征审查的当前视频大类的第二视频帧数目;Obtaining, respectively, the number of second video frames belonging to the current video category of the merged review, but not belonging to the current video category of the current class feature review;
将第二视频帧数目与除以融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的漏判率;Taking the value of the second video frame number and the number of samples of the video frame divided by the current video of the fusion review as the missed rate of the current class feature review;
根据当前类特征审查的误判率、漏判率确定当前视频大类的当前类特征审查的准确率。According to the misjudgment rate and the missed rate of the current class feature review, the accuracy of the current class feature review of the current video category is determined.
结合上述本发明的具体示例中的说明,各视频大类包括有黄色视频、暴力视频、反动视频,所进行的各类特征审查包括有文字特征审查、声音特征审查、图像特征审查。因而最后可以得到黄色视频(暴力视频、反动视频)的文字特征审查的准确率、声音特征审查的准确率、图像特征审查的准确率,共计九种准确率。In combination with the above description of the specific example of the present invention, each video category includes a yellow video, a violent video, and a reaction video, and various feature reviews performed include text feature review, sound feature review, and image feature review. Therefore, in the end, the accuracy of the character review of the yellow video (violent video, reactionary video), the accuracy of the sound feature review, and the accuracy of the image feature review can be obtained, and a total of nine accuracy rates are obtained.
针对确定暴力视频的各类特征审查的准确率为例,具体过程可以是如下所述。For the accuracy rate of various feature reviews for determining a violent video, the specific process may be as follows.
首先,对于暴力视频的文本审查的准确率,可以是以融合审查的暴力视频库RB为标准,确定文字审查的暴力视频库WB的误判率、漏判率,再基于该误判率、漏判率综合确定暴力视频的文本审查的准确率。First of all, the accuracy rate of the text review of the violent video can be determined by the violent video library RB of the fusion review, and the false positive rate and the missed rate of the violent video library WB of the text review are determined, and then based on the false positive rate and the leak rate. The judgment rate comprehensively determines the accuracy of the text review of the violent video.
对于误判率,确定属于文字审查的暴力视频库WB、但是却不属于融合审查的暴力视频库RB的视频帧的数目,再将该数目除以融合审查的暴力视频库RB的视频帧的数目,将所得到的值作为暴力视频文本审查的误判率,即:For the false positive rate, determine the number of video frames belonging to the violent video library WB of the text review, but not the violent video library RB of the merged review, and divide the number by the number of video frames of the violent video library RB of the merged review. The value obtained is used as the false positive rate for violent video text review, namely:
暴力视频文本审查的误判率=(属于文字审查的暴力视频库WB、但是却不属于融合审查的暴力视频库RB的视频帧的数目)/(融合审查的暴力视频库RB的视频帧的数目)。The false positive rate of violent video text review = (the number of video frames belonging to the violent video library WB of the text review, but not belonging to the violent video library RB of the fusion review) / (the number of video frames of the violent video library RB of the merged review) ).
对于漏判率,确定属于融合审查的暴力视频库RB、但是不属于文字审查的暴力视频库WB的视频帧的数目,再将该数目除以融合审查的暴力视频库RB的视频帧的数目,将所得到的值作为暴力视频文本审查的漏判率,即:For the missed rate, the number of video frames belonging to the violent video library RB of the merge review but not the violent video library WB of the text review is determined, and the number is divided by the number of video frames of the violent video library RB of the merged review, The resulting value is used as a missed rate for violent video text review, ie:
暴力视频文本审查的漏判率=(属于融合审查的暴力视频库RB、但是却不属于文本审查的暴力视频库WB的视频帧的数目)/(融合审查的暴力视频库RB的视频帧的数目)。Missing rate of violent video text review = (number of video frames belonging to the violent video library RB of the merge review but not violent video library WB of the text review) / (number of video frames of the bred video library RB of the merged review) ).
然后,根据暴力视频文本审查的误判率、漏判率综合确定暴力视频文本审查的准确率。具体综合确定时,可以是将暴力视频文本审查的误判率、暴力视频文本审查的漏判率中的较大值、较小值、平均值、加权平均值或者通过其他方式计算所得的值作为暴力视频文本审查的准确率,具体的综合确定方式根据实际应用需求的不同可以有所不同。Then, based on the false positive rate and the missed rate of the violent video text review, the accuracy of the violent video text review is comprehensively determined. When the specific comprehensive determination is made, the false positive rate of the violent video text review, the larger value, the smaller value, the average value, the weighted average value of the violent video text review, or the value calculated by other methods may be used as the misjudgment rate of the violent video text review. The accuracy of the violent video text review, the specific comprehensive determination method can be different according to the actual application needs.
暴力视频声音审查的准确率、暴力视频图像审查的准确率的确定方式与上述确定暴力视频文本审查的准确率的方式类似。The accuracy of violent video sound review and the accuracy of violent video image review are similar to the above-described methods for determining the accuracy of violent video text review.
对于暴力视频的声音审查的准确率,可以是以融合审查的暴力视频库RB为标准,确定声音审查的暴力视频库VB的误判率、漏判率,再基于该误判率、漏判率综合确定暴力视频的声音审查的准确率。For the accuracy of the voice review of the violent video, the violent video library RB of the fusion review may be used as the standard to determine the false positive rate and the missed rate of the violent video library VB of the voice review, and then based on the false positive rate and the missed rate. Comprehensively determine the accuracy of the sound review of violent videos.
对于误判率,确定属于声音审查的暴力视频库VB、但是却不属于融合审查的暴力视频库RB的视频帧的数目,再将该数目除以融合审查的暴力视频库RB的视频帧的数目,将所得到的值作为暴力视频声音审查的误判率,即:For the false positive rate, determine the number of video frames belonging to the violent video library VB of the sound review, but not the violent video library RB of the fusion review, and divide the number by the number of video frames of the violent video library RB of the fusion review. , the value obtained as the false positive rate of violent video sound review, namely:
暴力视频声音审查的误判率=(属于声音审查的暴力视频库VB、但是却不属于融合审查的暴力视频库RB的视频帧的数目)/(融合审查的暴力视频库RB的视频帧的数目)。The false positive rate of violent video sound review = (the number of video frames belonging to the violent video library VB of the sound review, but not belonging to the violent video library RB of the fusion review) / (the number of video frames of the violent video library RB of the merged review) ).
对于漏判率,确定属于融合审查的暴力视频库RB、但是不属于声音审查的暴力视频库VB的视频帧的数目,再将该数目除以融合审查的暴力视频库RB的视频帧的数目,将所得到的值作为暴力视频声音审查的漏判率,即:For the missed rate, the number of video frames belonging to the violent video library RB of the fusion review but not the violent video library VB of the sound review is determined, and the number is divided by the number of video frames of the violent video library RB of the merged review, The value obtained is used as the missed rate of violent video sound review, namely:
暴力视频声音审查的漏判率=(属于融合审查的暴力视频库RB、但是却不属于声音审查的暴力视频库VB的视频帧的数目)/(融合审查的暴力视频库RB的视频帧的数目)。Missing rate of violent video sound review = (number of video frames belonging to the violent video library RB of the fusion review but not violent video library VB of the sound review) / (number of video frames of the bred video library RB of the merged review) ).
然后,根据暴力视频声音审查的误判率、漏判率综合确定暴力视频声音审查的准确率。具体综合确定时,可以是将暴力视频声音审查的误判率、暴力视频声音审查的漏判率中的较大值、较小值、平均值、加权平均值或者通过其他方式计算所得的值作为暴力视频声音审查的准确率,具体的综合确定方式根据实际应用需求的不同可以有所不同。Then, based on the false positive rate and the missed rate of the violent video sound review, the accuracy of the violent video sound review is determined. When the specific comprehensive determination is made, the false positive rate of the violent video sound review, the larger value, the smaller value, the average value, the weighted average value or the value calculated by other methods may be used as the false positive rate of the violent video sound review. The accuracy of violent video sound review, the specific comprehensive determination method can be different according to the actual application needs.
对于暴力视频的图像审查的准确率,可以是以融合审查的暴力视频库RB为标准,确定图像审查的暴力视频库GB的误判率、漏判率,再基于该误判率、漏判率综合确定暴力视频的图像审查的准确率。For the accuracy of the image review of the violent video, the violent video library RB of the fusion review may be used as a standard to determine the false positive rate and the missed rate of the violent video library GB of the image review, and then based on the false positive rate and the missed rate. Comprehensively determine the accuracy of image review of violent videos.
对于误判率,确定属于图像审查的暴力视频库GB、但是却不属于融合审查的暴力视频库RB的视频帧的数目,再将该数目除以融合审查的暴力视频库RB的视频帧的数目,将所得到的值作为暴力视频图像审查的误判率,即:For the false positive rate, determine the number of video frames belonging to the violent video library GB of the image review, but not belonging to the violent video library RB of the merge review, and divide the number by the number of video frames of the violent video library RB of the merged review. , the value obtained as the false positive rate of violent video image review, namely:
暴力视频图像审查的误判率=(属于图像审查的暴力视频库GB、但是却不属于融合审查的暴力视频库RB的视频帧的数目)/(融合审查的暴力视频库RB的视频帧的数目)。The false positive rate of violent video image review = (the number of video frames belonging to the violent video library GB of the image review, but not belonging to the violent video library RB of the fusion review) / (the number of video frames of the violent video library RB of the merged review) ).
对于漏判率,确定属于融合审查的暴力视频库RB、但是不属于图像审查的暴力视频库GB的视频帧的数目,再将该数目除以融合审查的暴力视频库RB的视频帧的数目,将所得到的值作为暴力视频图像审查的漏判率,即:For the missed rate, the number of video frames belonging to the violent video library RB of the merge review, but not the violent video library GB of the image review, is determined, and the number is divided by the number of video frames of the violent video library RB of the merged review, The resulting value is used as a missed rate for violent video image review, namely:
暴力视频图像审查的漏判率=(属于融合审查的暴力视频库RB、但是却不属于图像审查的暴力视频库GB的视频帧的数目)/(融合审查的暴力视频库RB的视频帧的数目)。Missing rate of violent video image review = (the number of video frames belonging to the violent video library RB of the merged review but not belonging to the violent video library GB of the image review) / (the number of video frames of the violent video library RB of the merged review) ).
然后,根据暴力视频图像审查的误判率、漏判率综合确定暴力视频图像审查的准确率。具体综合确定时,可以是将暴力视频图像审查的误判率、暴力视频图像审查的漏判率中的较大值、较小值、平均值、加权平均值或者通过其他方式计算所得的值作为暴力视频图像审查的准确率,具体的综合确定方式根据实际应用需求的不同可以有所不同。Then, based on the false positive rate and the missed rate of the violent video image review, the accuracy of the violent video image review is comprehensively determined. When the specific comprehensive determination is made, the false positive rate of the violent video image review, the larger value, the smaller value, the average value, the weighted average value of the violent video image review, or the value calculated by other methods may be used as the misjudgment rate of the violent video image review. The accuracy of violent video image review, the specific comprehensive determination method can be different according to the actual application needs.
上述说明中,是以确定暴力视频文字审查的准确率、暴力视频声音审查的准确率、暴力视频图像审查的准确率进行说明。对于黄色视频、反动视频等其他视频大类来说,具体的确定各类特征审查的准确率的方式与上述类似,在此不予详加赘述。In the above description, it is to determine the accuracy of the violent video text review, the accuracy of the violent video sound review, and the accuracy of the violent video image review. For other video categories such as yellow video and reaction video, the specific method for determining the accuracy of each feature review is similar to the above, and will not be described in detail here.
然后,根据各视频大类的各类特征审查的准确率,可综合确定各视频大类的特征审查融合参数。Then, according to the accuracy rate of each feature review of each video category, the feature review fusion parameters of each video category can be comprehensively determined.
以上述暴力视频的各类特征审查的准确率(包括暴力视频文字审查的准确率、暴力视频声音审查的准确率、暴力视频图像审查的准确率)为例,暴力视频的特征审查融合参数可以记为(rw,rv,rg),其中rw表示暴力视频文字审查的融合参数,rv表示暴力视频声音审查的融合参数,rg表示暴力视频图像审查的融合参数。For example, the accuracy rate of the various features of the above violent video (including the accuracy of violent video text review, the accuracy of violent video sound review, and the accuracy of violent video image review), for example, the characteristics of the violent video can be recorded. For (rw, rv, rg), where rw represents the fusion parameter of the violent video text review, rv represents the fusion parameter of the violent video sound review, and rg represents the fusion parameter of the violent video image review.
在其中一个具体示例中,参数rw、rv、rg可以分别采用下述方式确定:In one specific example, the parameters rw, rv, rg can be determined in the following manners, respectively:
rw=暴力视频文字审查的准确率/(暴力视频文字审查的准确率+暴力视频声音审查的准确率、暴力视频图像审查的准确率);Rw=Accuracy rate of violent video text review/(accuracy of violent video text review+accuracy of violent video sound review, accuracy of violent video image review);
rv=暴力视频声音审查的准确率/(暴力视频文字审查的准确率+暴力视频声音审查的准确率、暴力视频图像审查的准确率);Rv=Accuracy rate of violent video sound review/(accuracy of violent video text review+accuracy of violent video sound review, accuracy of violent video image review);
rg=暴力视频图像审查的准确率/(暴力视频文字审查的准确率+暴力视频声音审查的准确率、暴力视频图像审查的准确率)。Rg=Accuracy of violent video image review/(accuracy of violent video text review+accuracy of violent video sound review, accuracy of violent video image review).
需要说明的是,这种确定方式仅仅只是一种示例性的说明,本领域技术人员可以预见还可以通过其他的方式来对融合参数进行综合性的确定,在此不予穷举。It should be noted that the determination manner is merely an exemplary description, and those skilled in the art can foresee that the fusion parameters can be comprehensively determined by other means, which is not exhaustive.
对于黄色视频、反动视频等其他视频大类的特征融合参数,具体的综合确定可以与上述暴力视频的特征融合参数的确定方式类似,在此不予详加赘述。For the feature fusion parameters of other video categories such as yellow video and reaction video, the specific comprehensive determination may be similar to the determination method of the feature fusion parameter of the above violent video, and details are not described herein.
上述获得的各视频大类的特征融合参数,可以予以储存,以便于后续对待审查视频帧的融合审查。The feature fusion parameters of each of the video categories obtained above can be stored for subsequent review of the fusion of the video frames to be reviewed.
在对待审查视频帧进行融合审查时,可先采用上述预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类。出于说明的目的,在此假设该待审查所属的视频大类为暴力视频。When performing the fusion review on the video frame to be examined, the video frame to be examined may be classified by using the foregoing preset fusion review classification manner, and the video category to which the video frame to be examined belongs belongs. For the purpose of illustration, it is assumed here that the video to which the subject to be reviewed belongs is a violent video.
然后,从该待审查视频帧中提取出相应的各类特征,具体可以包括有文字特征、声音特征、图像特征。Then, the corresponding various types of features are extracted from the to-be-reviewed video frame, and specifically may include a text feature, a sound feature, and an image feature.
然后,基于该待审查视频帧的文字特征、声音特征、图像特征,结合暴力视频的特征融合参数,确定该待审查视频帧属于暴力视频下的各视频小类的可能性。以下结合其中一个具体示例进行详细说明。Then, based on the character features, the sound features, and the image features of the video frame to be examined, combined with the feature fusion parameters of the violent video, the possibility that the video frame to be examined belongs to each video subclass under the violent video is determined. The following is a detailed description in conjunction with one of the specific examples.
假设该待审查视频帧所属的暴力视频的大类为i大类,该暴力视频又被划分为N个小类,分别记为i1、i2、i3、……、iN。Assume that the violent video to which the video frame to be examined belongs belongs to the i-class, and the violent video is further divided into N subclasses, which are respectively recorded as i1, i2, i3, ..., iN.
然后,根据该待审查视频帧的文字特征,判断该待审查视频帧或者该文字特征属于第i1个小类的可能性wi1、属于第i2个小类的可能性wi2、属于第i3个小类的可能性wi3、……、属于第iN个小类的可能性wiN。从而必然会有wi1+wi2+wi3+……+wiN=1。Then, according to the character feature of the video frame to be examined, it is determined that the video frame to be examined or the character feature belongs to the i1th subclass, the possibility wi1, the possibility i2 belonging to the i2th subclass, and belongs to the i3th subclass. The possibility wi3, ..., the possibility wiN belonging to the iNth subclass. Thus there must be wi1+wi2+wi3+...+wiN=1.
根据该待审查视频帧的声音特征,判断该待审查视频帧或者该声音特征属于第i1个小类的可能性vi1、属于第i2个小类的可能性vi2、属于第i3个小类的可能性vi3、……、属于第iN个小类的可能性viN。从而必然会有vi1+vi2+vi3+……+viN=1。Determining, according to the sound characteristics of the video frame to be examined, the possibility that the video frame to be examined or the sound feature belongs to the i1th subclass, the possibility vi2 belonging to the i2th subclass, and the possibility of belonging to the i3th subclass Sex vi3, ..., the possibility viN belonging to the iNth subclass. Therefore, there will inevitably be vi1+vi2+vi3+...+viN=1.
根据该待审查视频帧的图像特征,判断该待审查视频帧或者该图像特征属于第i1个小类的可能性gi1、属于第i2个小类的可能性gi2、属于第i3个小类的可能性gi3、……、属于第iN个小类的可能性giN。从而必然会有gi1+gi2+gi3+……+giN=1。Determining, according to the image feature of the video frame to be examined, the possibility that the video frame to be examined or the image feature belongs to the i1th subclass gi1, the possibility gi2 belonging to the i2th subclass, and the possibility of belonging to the i3th subclass Sex gi3, ..., the probability giN belonging to the iNth subclass. Thus there must be gi1+gi2+gi3+...+giN=1.
从而基于上述获得的结果,可以得到上述待审查视频帧属于暴力视频下各小类的可能性分别为:Therefore, based on the obtained results, it can be obtained that the video frames to be examined belong to the subcategories of the violent video:
该待审查视频帧属于第i1个小类的可能性为:pi1=rw*wi1+rv+vi1+rg*gi1;The possibility that the video frame to be examined belongs to the i1th subclass is: pi1=rw*wi1+rv+vi1+rg*gi1;
该待审查视频帧属于第i2个小类的可能性为:pi2=rw*wi2+rv+vi2+rg*gi2;The possibility that the video frame to be examined belongs to the i2th subclass is: pi2=rw*wi2+rv+vi2+rg*gi2;
该待审查视频帧属于第i3个小类的可能性为:pi3=rw*wi3+rv+vi3+rg*gi3;The possibility that the video frame to be examined belongs to the i3th subclass is: pi3=rw*wi3+rv+vi3+rg*gi3;
……......
该待审查视频帧属于第iN个小类的可能性为:piN=rw*wiN+rv+viN+rg*giN。The possibility that the video frame to be examined belongs to the i-th subclass is: piN=rw*wiN+rv+viN+rg*giN.
从而,根据该待审查视频帧属于暴力视频下的第i1、i2、i3、……、iN个小类的可能性pi1、pi2、pi3、……、piN,可以综合确定该待审查视频帧所属的视频小类。一般情况下,可以将pi1、pi2、pi3、……、piN中的最大值对应的视频小类确定为该待审查视频帧所属的视频小类。Therefore, according to the possibility that the video frame to be examined belongs to the i1, i2, i3, ..., iN subclasses under the violent video, pi1, pi2, pi3, ..., piN, the video frame to be examined may be comprehensively determined. Video subclass. In general, the video subclass corresponding to the maximum value of pi1, pi2, pi3, ..., piN may be determined as the video subclass to which the video frame to be examined belongs.
根据上述本发明的深入融合视频审查方法,本发明还提供一种深入融合视频审查系统,以下对本发明的深入融合视频审查系统的实施例进行详细说明。According to the in-depth fusion video review method of the present invention described above, the present invention also provides an in-depth fusion video review system. The following describes an embodiment of the in-depth fusion video review system of the present invention.
图4中示出了本发明的深入融合视频审查系统实施例的结构示意图。如图4所示,本实施例中的深入融合视频审查系统包括有:A block diagram of an embodiment of the in-depth fusion video review system of the present invention is shown in FIG. As shown in FIG. 4, the in-depth fusion video review system in this embodiment includes:
视频大类融合确定模块401,用于采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类;The video major class fusion determining module 401 is configured to classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs.
特征提取模块402,用于提取所述待审查视频帧中的各类特征;The feature extraction module 402 is configured to extract various features in the video frame to be examined;
视频小类融合确定模块403,用于分别根据所述待审查视频帧中的各类特征、所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性,并根据所述待审查视频帧属于所述视频大类下的各视频小类的可能性,综合确定所述待审查视频帧所属的视频小类。The video subclass fusion determining module 403 is configured to respectively review the fusion parameters according to the various features in the to-be-reviewed video frame and the features of the video category, and determine that the to-be-reviewed video frame belongs to the video category. The possibility of each video subclass, and comprehensively determining the video subclass to which the video frame to be examined belongs according to the possibility that the video frame to be examined belongs to each video subclass under the video category.
根据本实施例中的方案,其在进行审查时,是先采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类,再基于该视频大类的特征审查融合参数、以及该待审查视频帧中的各类特征,确定出该待审查视频帧属于上述确定的视频大类下的各视频小类的可能性,并基于属于各视频小类的可能性,综合确定出上述待审查视频帧所属的视频小类。其基于该待审查视频帧中的各类特征、以及视频大类的特征审查融合参数,综合考虑了视频帧中的各类特征在该视频大类中的视频帧中所起的作用,对不同类特征对不同类型视频帧的作用加以区分,提高了视频审查的准确度。According to the solution in this embodiment, when reviewing, the video frame to be examined is classified by using a preset fusion review classification manner, and the video category to which the video frame to be examined belongs is obtained, and then based on the characteristics of the video category. Examining the fusion parameters and various features in the video frame to be examined, determining the possibility that the video frame to be examined belongs to each video subclass under the above determined video category, and based on the possibility of belonging to each video subclass And comprehensively determining the video subclass to which the video frame to be examined belongs. It examines the fusion parameters based on various features in the video frame to be examined and the features of the video category, and comprehensively considers the role of various features in the video frame in the video frames in the video category, and different Class features distinguish between the effects of different types of video frames, improving the accuracy of video censorship.
图4所示中,本实施例中的深入融合视频审查系统,还可以包括有:用于确定所述各视频大类的特征审查融合参数的融合参数确定模块404。As shown in FIG. 4, the in-depth fusion video review system in this embodiment may further include: a fusion parameter determination module 404 for determining feature review fusion parameters of the video categories.
如图4所示,该融合参数确定模块404包括有:As shown in FIG. 4, the fusion parameter determination module 404 includes:
样本融合审查模块4041,用于采用所述预设融合审查分类方式对视频样本数据库中的各视频帧进行分类,获得融合审查的各视频大类的视频帧;The sample fusion review module 4041 is configured to classify each video frame in the video sample database by using the preset fusion review classification manner, and obtain a video frame of each video category that is merged and reviewed;
样本分类审查模块4042,用于分别采用各类特征审查方法对所述视频样本数据库中的各视频帧进行分类,分别获得各类特征审查后的各视频大类的视频帧;The sample classification review module 4042 is configured to classify each video frame in the video sample database by using various feature review methods, and obtain video frames of each video category after each feature review;
样本准确率确定模块4043,用于根据所述融合审查的各视频大类的视频帧、各类特征审查后的各视频大类的视频帧,确定各视频大类的各类特征审查的准确率;The sample accuracy determining module 4043 is configured to determine the accuracy of each type of feature review of each video category according to the video frames of each video category and the video frames of each video category after the feature review. ;
融合参数综合确定模块4044,根据各视频大类的各类特征审查的准确率,确定各视频大类的特征审查融合参数。The fusion parameter comprehensive determination module 4044 determines the feature review fusion parameters of each video category according to the accuracy of each type of feature review of each video category.
在其中一个具体示例中,上述融合参数综合确定模块4044,可以是将当前视频大类的当前类特征审查的准确率相对于当前视频大类的各类特征审查的准确率之和的比例作为当前视频大类的当前类特征审查的融合参数,当前视频大类的特征融合审查参数包括当前视频大类的各类特征审查的融合参数。In one specific example, the fusion parameter integration determining module 4044 may be a current ratio of the accuracy of the current class feature review of the current video category to the sum of the accuracy of the current video category. The fusion parameters of the current class feature review of the video category, the feature fusion review parameters of the current video category include the fusion parameters of various feature reviews of the current video category.
其中,如图4所示,上述样本准确率确定模块4043具体可以包括有:As shown in FIG. 4, the sample accuracy determining module 4043 may specifically include:
误判率确定模块40431,用于分别获取属于当前类特征审查的当前视频大类、但不属于融合审查的当前视频大类的视频帧的第一视频帧数目,并将所述第一视频帧数目除以所述融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的误判率;The false positive rate determining module 40431 is configured to respectively acquire the first video frame number of the video frame belonging to the current video category of the current class feature review but not belonging to the current video category of the merge review, and the first video frame The number is divided by the value of the sample number of the video frame of the current video category of the fusion review as the false positive rate of the current class feature review;
漏判率确定模块40432,用于分别获取属于融合审查的当前视频大类的视频帧、但不属于当前类特征审查的当前视频大类的第二视频帧数目,并将所述第二视频帧数目与除以所述融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的漏判率;The missed rate determining module 40432 is configured to respectively acquire the number of second video frames belonging to the current video category of the merged review but not the current video category of the current class feature review, and the second video frame The number and the value of the number of samples of the video frame divided by the current video category of the fusion review are used as the missed rate of the current class feature review;
准确率确定模块40433,用于根据当前类特征审查的误判率、漏判率确定当前视频大类的当前类特征审查的准确率。The accuracy rate determining module 40433 is configured to determine an accuracy rate of the current class feature review of the current video category according to the false positive rate and the missed rate of the current class feature review.
在其中一个具体示例中,上述准确率确定模块40433,可以将当前类特征审查的误判率、漏判率的平均值或者加权平均值作为当前视频大类的当前类特征审查的准确率。In one specific example, the accuracy determination module 40433 may use the error rate of the current class feature review, the average value of the missed rate, or the weighted average as the accuracy of the current class feature review of the current video category.
如图4所示,在其中一个示例中,上述视频小类融合确定模块403具体可以包括:As shown in FIG. 4, in one example, the video subclass fusion determining module 403 may specifically include:
特征小类可能性确定模块4031,用于分别判断待审查视频帧中的各类特征属于所述视频大类下的各视频小类的可能性;The feature subclass likelihood determining module 4031 is configured to determine, respectively, the possibility that each type of feature in the video frame to be examined belongs to each video subclass under the video category;
视频小类可能性确定模块4032,用于根据各类特征属于所述视频大类下的各视频小类的可能性,以及所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性;The video subclass likelihood determining module 4032 is configured to determine the video frame to be reviewed according to the possibility that each type of feature belongs to each video subclass under the video category, and the feature review fusion parameter of the video category The possibility of belonging to each video subclass under the video category;
小类确定模块,用于确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性的最大值,并将该可能性的最大值对应的视频小类确定为所述待审查视频帧所属的视频小类。a subclass determining module, configured to determine a maximum value of a likelihood that the video frame to be examined belongs to each video subclass under the video category, and determine a video subclass corresponding to the maximum value of the likelihood as the The video subclass to which the video frame to be reviewed belongs.
本发明的深入融合视频审查系统中各模块的具体实现方式,可以与上述本发明的深入融合视频审查方法中的相同,在此不予赘述。The specific implementation manner of each module in the in-depth fusion video review system of the present invention may be the same as that in the above-mentioned in-depth fusion video review method of the present invention, and details are not described herein.
以上所述实施例仅表达了本发明的几种实施方式,其描述较为具体和详细,但并不能因此而理解为对本发明专利范围的限制。应当指出的是,对于本领域的普通技术人员来说,在不脱离本发明构思的前提下,还可以做出若干变形和改进,这些都属于本发明的保护范围。因此,本发明专利的保护范围应以所附权利要求为准。The above-mentioned embodiments are merely illustrative of several embodiments of the present invention, and the description thereof is more specific and detailed, but is not to be construed as limiting the scope of the invention. It should be noted that a number of variations and modifications may be made by those skilled in the art without departing from the spirit and scope of the invention. Therefore, the scope of the invention should be determined by the appended claims.

Claims (14)

  1. 一种深入融合视频审查方法,其特征在于,包括步骤:An in-depth fusion video review method, characterized in that it comprises the steps of:
    采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类;The video frames to be examined are classified by using a preset fusion review classification manner, and the video categories to which the video frames to be examined belong are obtained;
    提取所述待审查视频帧中的各类特征;Extracting various features in the video frame to be examined;
    分别根据所述待审查视频帧中的各类特征、所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性;Determining, according to the various features in the to-be-reviewed video frame, and the feature of the video category, the fusion parameters, and determining the possibility that the video frame to be examined belongs to each video subclass under the video category;
    根据所述待审查视频帧属于所述视频大类下的各视频小类的可能性,综合确定所述待审查视频帧所属的视频小类。Determining, according to the possibility that the video frame to be examined belongs to each video subclass under the video category, comprehensively determining a video subclass to which the to-be-reviewed video frame belongs.
  2. 根据权利要求1所述的深入融合视频审查方法,其特征在于,视频大类的特征审查融合参数的确定方式包括:The in-depth fusion video review method according to claim 1, wherein the method for determining the feature review fusion parameters of the video category comprises:
    采用所述预设融合审查分类方式对视频样本数据库中的各视频帧进行分类,获得融合审查的各视频大类的视频帧;Using the preset fusion review classification method to classify each video frame in the video sample database, and obtain a video frame of each video category that is merged and reviewed;
    分别采用各类特征审查方法对所述视频样本数据库中的各视频帧进行分类,分别获得各类特征审查后的各视频大类的视频帧;Each video frame in the video sample database is classified by using various feature review methods, and video frames of each video category after each feature review are respectively obtained;
    根据所述融合审查的各视频大类的视频帧、各类特征审查后的各视频大类的视频帧,确定各视频大类的各类特征审查的准确率;Determining the accuracy of each type of feature review of each video category according to the video frames of each video category reviewed by the fusion review and the video frames of each video category after review of various features;
    根据各视频大类的各类特征审查的准确率,确定各视频大类的特征审查融合参数。According to the accuracy rate of each type of feature review of each video category, the feature review fusion parameters of each video category are determined.
  3. 根据权利要求2所述的深入融合视频审查方法,其特征在于,确定各视频大类的各类特征审查的准确率的方式包括:The in-depth fusion video review method according to claim 2, wherein the manner of determining the accuracy of each feature review of each video category comprises:
    分别获取属于当前类特征审查的当前视频大类、但不属于融合审查的当前视频大类的视频帧的第一视频帧数目;Obtaining, respectively, the number of first video frames of the video frame belonging to the current video category of the current class feature review but not belonging to the current video category of the fusion review;
    将所述第一视频帧数目除以所述融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的误判率;Dividing the number of the first video frame by the value of the sample number of the video frame of the current video category of the fusion review as the false positive rate of the current class feature review;
    分别获取属于融合审查的当前视频大类的视频帧、但不属于当前类特征审查的当前视频大类的第二视频帧数目;Obtaining, respectively, the number of second video frames belonging to the current video category of the merged review, but not belonging to the current video category of the current class feature review;
    将所述第二视频帧数目与除以所述融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的漏判率;And determining, by the second video frame number, a value of a sample number of the video frame divided by the current video of the fusion review as a missed rate of the current class feature review;
    根据当前类特征审查的误判率、漏判率确定当前视频大类的当前类特征审查的准确率。According to the misjudgment rate and the missed rate of the current class feature review, the accuracy of the current class feature review of the current video category is determined.
  4. 根据权利要求3所述的深入融合视频审查方法,其特征在于,将当前类特征审查的误判率、漏判率的平均值或者加权平均值作为当前视频大类的当前类特征审查的准确率。The in-depth fusion video review method according to claim 3, wherein the error rate of the current class feature review, the average value of the missed rate, or the weighted average value is used as the accuracy rate of the current class feature review of the current video category. .
  5. 根据权利要求2所述的深入融合视频审查方法,其特征在于,根据各视频大类的各类特征审查的准确率,确定各视频大类的特征审查融合参数的方式包括:The in-depth fusion video review method according to claim 2, wherein the method for determining the feature review fusion parameters of each video category according to the accuracy of each type of feature review of each video category includes:
    将当前视频大类的当前类特征审查的准确率相对于当前视频大类的各类特征审查的准确率之和的比例作为当前视频大类的当前类特征审查的融合参数;The ratio of the accuracy of the current class feature review of the current video category to the sum of the accuracy of the various feature reviews of the current video category is taken as the fusion parameter of the current class feature review of the current video category;
    当前视频大类的特征融合审查参数包括当前视频大类的各类特征审查的融合参数。The feature fusion review parameters of the current video category include the fusion parameters of various feature reviews of the current video category.
  6. 根据权利要求1至5任意一项所述的深入融合视频审查方法,其特征在于,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性的方式包括:The in-depth fusion video review method according to any one of claims 1 to 5, wherein the manner of determining the possibility that the video frame to be examined belongs to each video subclass under the video category comprises:
    分别判断待审查视频帧中的各类特征属于所述视频大类下的各视频小类的可能性;Determining, respectively, the possibility that each type of feature in the video frame to be examined belongs to each video subclass under the video category;
    根据各类特征属于所述视频大类下的各视频小类的可能性,以及所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性。Determining, according to the possibility that each type of feature belongs to each video subclass under the video category, and the feature review fusion parameter of the video category, determining that the video frame to be examined belongs to each video under the video category The possibility of a class.
  7. 根据权利要求1至5任意一项所述的深入融合视频审查方法,其特征在于,综合确定所述待审查视频帧所属的视频小类的方式包括:The in-depth fusion video review method according to any one of claims 1 to 5, wherein the method for comprehensively determining the video subclass to which the video frame to be examined belongs includes:
    确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性的最大值;Determining a maximum value of the likelihood that the video frame to be examined belongs to each video subclass under the video category;
    将该可能性的最大值对应的视频小类确定为所述待审查视频帧所属的视频小类。The video subclass corresponding to the maximum value of the likelihood is determined as the video subclass to which the video frame to be examined belongs.
  8. 一种深入融合视频审查系统,其特征在于,包括:An in-depth fusion video review system featuring:
    视频大类融合确定模块,用于采用预设融合审查分类方式对待审查视频帧进行分类,获得该待审查视频帧所属的视频大类;The video large class fusion determining module is configured to classify the video frames to be examined by using a preset fusion review classification manner, and obtain a video category to which the video frame to be examined belongs;
    特征提取模块,用于提取所述待审查视频帧中的各类特征;a feature extraction module, configured to extract various features in the video frame to be examined;
    视频小类融合确定模块,用于分别根据所述待审查视频帧中的各类特征、所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性,并根据所述待审查视频帧属于所述视频大类下的各视频小类的可能性,综合确定所述待审查视频帧所属的视频小类。a video subclass fusion determining module, configured to respectively examine, according to various features in the to-be-reviewed video frame, features of the video category, a fusion parameter, and determine that the to-be-reviewed video frame belongs to each of the video categories. The possibility of the video subclass, and comprehensively determining the video subclass to which the video frame to be examined belongs according to the possibility that the video frame to be examined belongs to each video subclass under the video category.
  9. 根据权利要求8所述的深入融合视频审查系统,其特征在于,还包括:The in-depth fusion video review system according to claim 8, further comprising:
    融合参数确定模块,用于确定所述各视频大类的特征审查融合参数。The fusion parameter determining module is configured to determine a feature review fusion parameter of each of the video categories.
  10. 根据权利要求9所述的深入融合视频审查系统,其特征在于,所述融合参数确定模块包括:The in-depth fusion video review system according to claim 9, wherein the fusion parameter determination module comprises:
    样本融合审查模块,用于采用所述预设融合审查分类方式对视频样本数据库中的各视频帧进行分类,获得融合审查的各视频大类的视频帧;a sample fusion review module, configured to classify each video frame in the video sample database by using the preset fusion review classification manner, and obtain a video frame of each video category that is merged and reviewed;
    样本分类审查模块,用于分别采用各类特征审查方法对所述视频样本数据库中的各视频帧进行分类,分别获得各类特征审查后的各视频大类的视频帧;The sample classification review module is configured to classify each video frame in the video sample database by using various feature review methods, and obtain video frames of each video category after each feature review;
    样本准确率确定模块,用于根据所述融合审查的各视频大类的视频帧、各类特征审查后的各视频大类的视频帧,确定各视频大类的各类特征审查的准确率;a sample accuracy determining module, configured to determine an accuracy rate of each type of feature review of each video category according to the video frame of each video category and the video frames of each video category after the feature review;
    融合参数综合确定模块,根据各视频大类的各类特征审查的准确率,确定各视频大类的特征审查融合参数。The fusion parameter comprehensive determination module determines the feature review fusion parameters of each video category according to the accuracy of each type of feature review of each video category.
  11. 根据权利要求10所述的深入融合视频审查方法,其特征在于,所述样本准确率确定模块包括:The in-depth fusion video review method according to claim 10, wherein the sample accuracy determination module comprises:
    误判率确定模块,用于分别获取属于当前类特征审查的当前视频大类、但不属于融合审查的当前视频大类的视频帧的第一视频帧数目,并将所述第一视频帧数目除以所述融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的误判率;The false positive rate determining module is configured to respectively obtain the first video frame number of the current video category that belongs to the current video category of the current class feature review but does not belong to the current video category of the fusion review, and the number of the first video frame Dividing the value of the sample number of the video frame of the current video category of the fusion review as the false positive rate of the current class feature review;
    漏判率确定模块,用于分别获取属于融合审查的当前视频大类的视频帧、但不属于当前类特征审查的当前视频大类的第二视频帧数目,并将所述第二视频帧数目与除以所述融合审查的当前视频大类的视频帧的样本数目的值作为该当前类特征审查的漏判率;a missing judgment rate determining module, configured to respectively acquire a number of second video frames belonging to a current video category of the merged review, but not belonging to a current video category of the current class feature review, and the number of the second video frames The value of the sample number of the video frame divided by the current video category of the fusion review is used as the missed rate of the current class feature review;
    准确率确定模块,用于根据当前类特征审查的误判率、漏判率确定当前视频大类的当前类特征审查的准确率。The accuracy determination module is configured to determine the accuracy of the current class feature review of the current video category according to the false positive rate and the missed rate of the current class feature review.
  12. 根据权利要求11所述的深入融合视频审查系统,其特征在于,所述准确率确定模块,将当前类特征审查的误判率、漏判率的平均值或者加权平均值作为当前视频大类的当前类特征审查的准确率。The in-depth fusion video review system according to claim 11, wherein the accuracy determination module determines the false positive rate, the average value of the missed rate, or the weighted average of the current class feature review as the current video category. The accuracy of the current class feature review.
  13. 根据权利要求10所述的深入融合视频审查系统,其特征在于,所述融合参数综合确定模块,将当前视频大类的当前类特征审查的准确率相对于当前视频大类的各类特征审查的准确率之和的比例作为当前视频大类的当前类特征审查的融合参数,当前视频大类的特征融合审查参数包括当前视频大类的各类特征审查的融合参数。The in-depth fusion video review system according to claim 10, wherein the fusion parameter comprehensive determination module examines the accuracy of the current class feature review of the current video category with respect to various features of the current video category. The ratio of the sum of the accuracy rates is used as the fusion parameter of the current class feature review of the current video category. The feature fusion review parameters of the current video category include the fusion parameters of various feature reviews of the current video category.
  14. 根据权利要求8至13任意一项所述的深入融合视频审查系统,其特征在于,视频小类融合确定模块包括:The in-depth fusion video review system according to any one of claims 8 to 13, wherein the video subclass fusion determining module comprises:
    特征小类可能性确定模块,用于分别判断待审查视频帧中的各类特征属于所述视频大类下的各视频小类的可能性;a feature subclass likelihood determining module, configured to respectively determine, according to the possibility that each type of feature in the video frame to be examined belongs to each video subclass under the video category;
    视频小类可能性确定模块,用于根据各类特征属于所述视频大类下的各视频小类的可能性,以及所述视频大类的特征审查融合参数,确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性;a video subclass likelihood determining module, configured to determine, according to the possibility that each type of feature belongs to each video subclass under the video category, and the feature review fusion parameter of the video category, to determine that the to-be-reviewed video frame belongs to The possibility of each video subclass under the video category;
    小类确定模块,用于确定所述待审查视频帧属于所述视频大类下的各视频小类的可能性的最大值,并将该可能性的最大值对应的视频小类确定为所述待审查视频帧所属的视频小类。a subclass determining module, configured to determine a maximum value of a likelihood that the video frame to be examined belongs to each video subclass under the video category, and determine a video subclass corresponding to the maximum value of the likelihood as the The video subclass to which the video frame to be reviewed belongs.
PCT/CN2013/085739 2012-12-22 2013-10-23 Method and system for screening depth fusion video WO2014094492A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201210563525.7A CN103049530B (en) 2012-12-22 2012-12-22 Deeply merge video checking method and system
CN201210563525.7 2012-12-22

Publications (1)

Publication Number Publication Date
WO2014094492A1 true WO2014094492A1 (en) 2014-06-26

Family

ID=48062171

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2013/085739 WO2014094492A1 (en) 2012-12-22 2013-10-23 Method and system for screening depth fusion video

Country Status (2)

Country Link
CN (1) CN103049530B (en)
WO (1) WO2014094492A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111754544A (en) * 2019-03-29 2020-10-09 杭州海康威视数字技术股份有限公司 Video frame fusion method and device and electronic equipment

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103049530B (en) * 2012-12-22 2015-12-23 深圳先进技术研究院 Deeply merge video checking method and system
CN109495766A (en) * 2018-11-27 2019-03-19 广州市百果园信息技术有限公司 A kind of method, apparatus, equipment and the storage medium of video audit
CN110418161A (en) * 2019-08-02 2019-11-05 广州虎牙科技有限公司 Video reviewing method and device, electronic equipment and readable storage medium storing program for executing

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5128757A (en) * 1990-06-18 1992-07-07 Zenith Electronics Corporation Video transmission system using adaptive sub-band coding
CN101834982A (en) * 2010-05-28 2010-09-15 上海交通大学 Hierarchical screening method of violent videos based on multiplex mode
CN101853377A (en) * 2010-05-13 2010-10-06 复旦大学 Method for identifying content of digital video
CN103049530A (en) * 2012-12-22 2013-04-17 深圳先进技术研究院 System and method for deep fused video examination

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101021857A (en) * 2006-10-20 2007-08-22 鲍东山 Video searching system based on content analysis
CN101470897B (en) * 2007-12-26 2011-04-20 中国科学院自动化研究所 Sensitive film detection method based on audio/video amalgamation policy

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5128757A (en) * 1990-06-18 1992-07-07 Zenith Electronics Corporation Video transmission system using adaptive sub-band coding
CN101853377A (en) * 2010-05-13 2010-10-06 复旦大学 Method for identifying content of digital video
CN101834982A (en) * 2010-05-28 2010-09-15 上海交通大学 Hierarchical screening method of violent videos based on multiplex mode
CN103049530A (en) * 2012-12-22 2013-04-17 深圳先进技术研究院 System and method for deep fused video examination

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111754544A (en) * 2019-03-29 2020-10-09 杭州海康威视数字技术股份有限公司 Video frame fusion method and device and electronic equipment
CN111754544B (en) * 2019-03-29 2023-09-05 杭州海康威视数字技术股份有限公司 Video frame fusion method and device and electronic equipment

Also Published As

Publication number Publication date
CN103049530A (en) 2013-04-17
CN103049530B (en) 2015-12-23

Similar Documents

Publication Publication Date Title
WO2019037197A1 (en) Method and device for training topic classifier, and computer-readable storage medium
WO2014094492A1 (en) Method and system for screening depth fusion video
WO2015003480A1 (en) Information recommendation method and apparatus in social media
WO2011021907A2 (en) Metadata tagging system, image searching method and device, and method for tagging a gesture thereof
WO2018058959A1 (en) Sql auditing method and apparatus, server and storage device
WO2013174172A1 (en) File information previewing method and system
WO2020207035A1 (en) Crank call interception method, apparatus, and device, and storage medium
WO2013183811A1 (en) Portable device and method for controlling the same
WO2016163755A1 (en) Quality measurement-based face recognition method and apparatus
WO2014046401A1 (en) Device and method for changing shape of lips on basis of automatic word translation
WO2014193041A1 (en) System and method for analyzing yield by utilizing sensor data of manufacturing equipment
WO2020186773A1 (en) Call request monitoring method, device, apparatus, and storage medium
WO2017173838A1 (en) Verification-based message display method and communication terminal
WO2020082562A1 (en) Symbol identification method, apparatus, device, and storage medium
WO2020119384A1 (en) Medical insurance abnormity detection method, apparatus and device based on big data analysis, and medium
WO2019024485A1 (en) Data sharing method and device and computer readable storage medium
WO2018086219A1 (en) Method, device and terminal for storing call information
WO2013165083A1 (en) System and method for providing image-based video service
WO2016123898A1 (en) Short message managing method and mobile terminal thereof
WO2015129983A1 (en) Device and method for recommending movie on basis of distributed mining of fuzzy association rules
WO2017138766A1 (en) Hybrid-based image clustering method and server for operating same
EP2435935A2 (en) Method of providing by-viewpoint patent map and system thereof
WO2017067290A1 (en) Fingerprint entry method, apparatus, and terminal device
WO2020010568A1 (en) Big data artificial intelligence analysis device
WO2020062641A1 (en) Method for identifying user role, and user equipment, storage medium, and apparatus for identifying user role

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13865140

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13865140

Country of ref document: EP

Kind code of ref document: A1