CN105260398A - Quick sorting method for movie types based on poster and plot summary - Google Patents

Quick sorting method for movie types based on poster and plot summary Download PDF

Info

Publication number
CN105260398A
CN105260398A CN201510592018.XA CN201510592018A CN105260398A CN 105260398 A CN105260398 A CN 105260398A CN 201510592018 A CN201510592018 A CN 201510592018A CN 105260398 A CN105260398 A CN 105260398A
Authority
CN
China
Prior art keywords
film
placard
sigma
text
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510592018.XA
Other languages
Chinese (zh)
Inventor
胡卫明
付志康
李兵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201510592018.XA priority Critical patent/CN105260398A/en
Publication of CN105260398A publication Critical patent/CN105260398A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Abstract

The invention discloses a quick sorting method for movie types based on posters and plot summaries. The method comprises: determining a type assembly a movie belongs to, establishing poster training sets and plot summary training sets of various kinds of movies; extracting characteristics of a poster of a to-be-tested movie, using the obtained characteristics of each poster and a label training support vector machine corresponding to the poster to obtain a sorting model of posters; extracting characteristics of a text of the plot summary of the to-be-tested movie, using the obtained characteristics of each text and the corresponding label training support vector machine to obtain a sorting model of the texts; using the sorting model of the posters to predict the poster of the to-be-tested movie, to obtain a result Y1, and then calling the sorting model of the texts to predict the plot summary of the to-be-tested movie, to obtain a result Y2; and finally performing OR operation on the Y1 and Y2, to obtain the type of the to-be-tested movie. The method can rapidly and accurately predict the type of a movie under a condition of no movie videos.

Description

A kind of rapid classification method of the film types based on placard and story introduction
Technical field
The present invention relates to area of pattern recognition, particularly the detection technique of film types.
Background technology
Along with the fast development of internet, film has become an indispensable part in people's free life.Do not make unified regulation to the kind of film so far, also, the classification of film is divided into substantially: terror, love, action, comedy, science fiction etc.Web film all manually can put on class label to film, so, the Fast Classification that realizes film is necessary.
The detection of film types is all detect based on video content itself substantially.The detection of video content comprises: the detection of shot boundary, the detection of camera lens key frame of video and the detection of audio frequency characteristics.The basic assumption of shot boundary detector is that the content of adjacent two camera lenses exists larger difference.Therefore, the border of camera lens can be determined by the difference degree measured between consecutive frame.The feature of camera lens key frame of video comprises: the color of key frame of video, contrast, lightness, Texture eigenvalue.Detect by extracting the key frame of these features to video.Audio frequency characteristics mainly contains: temporal signatures, frequency domain character and acoustics Perception Features etc.
There is following problem in the detection based on video content: the data volume of needs is large, video detecting comparatively slowly, obviously cannot complete Detection task when not having video content itself, and accuracy rate is not very high simultaneously.
Summary of the invention
(1) technical matters that will solve
The object of the invention is to propose a kind ofly quickly and easily to the method that film detects, thus separated film fast to be achieved when there is no film video.
Two) technical scheme
In order to solve the problems of the technologies described above, the present invention proposes a kind of rapid classification method of the film types based on placard and story introduction, the method comprises the following steps: step 1: determine the type set belonging to film, sets up the placard training set of various types of film and the training set of story introduction;
Step 2: the feature extracting the placard of film to be measured, the feature of every width placard that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of placard;
Step 3: the feature extracting the text of the story introduction of film to be measured, the feature of each text that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of text;
Step 4: with the disaggregated model of placard, prediction is carried out to the placard of film to be measured and obtains type Y1, and then the disaggregated model calling text carries out prediction to the story introduction of film to be measured and obtains type Y2; Finally Y1 and Y2 is carried out OR operation; Namely contrast with the type label of film to be measured, if one to predict the outcome be correct, then predict that correct type is used as the type of last film to be measured; Otherwise using the type of Y1 as last film to be measured.
(3) beneficial effect
The present invention detects in conjunction with the placard of film and the type of story introduction to film, can to realize fast, the detection of high-accuracy when not having film video to the type of film.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the rapid classification method of the film types based on placard and story introduction of the present invention.
Fig. 2 is the type determining film of the present invention, and obtains the method flow diagram of type set.
Fig. 3 is the method flow diagram of acquisition placard disaggregated model of the present invention.
Fig. 4 is the method flow diagram of acquisition textual classification model of the present invention.
Fig. 5 is the method flow diagram of acquisition of the present invention film types to be measured.
Embodiment
For making the object, technical solutions and advantages of the present invention clearly understand, below in conjunction with specific embodiment, and with reference to accompanying drawing, the present invention is described in further detail.
The hardware of method carrying out practically of the present invention and programming language are also unrestricted, can realize method of the present invention by any language compilation.The present invention adopts a computing machine with 2.67G hertz central processing unit and 4G byte of memory, and by the C Plus Plus establishment program arrived involved in the present invention, achieves method of the present invention.
Fig. 1 is the process flow diagram of the rapid classification method of film types based on placard and story introduction.
Step 101: collect China and foreign countries' video website, determine the type set belonging to film, collect placard corresponding to film as much as possible and story introduction, set up the training set of film poster and the training set of story introduction, idiographic flow as shown in Figure 2.
First collect the film common type in China and foreign countries' video website, determine that common film types set is: horror film, romance movie, comedy and action movie etc.Then placard and the story introduction of the film of this Four types as much as possible is collected.Set up the training set of film poster and the training set of film story introduction respectively.
Step 102: the feature extracting placard, the feature of every width placard that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of placard, and idiographic flow as shown in Figure 3.
1): the feature extracting placard.
The feature of the placard that every width placard extracts is comprised: the number of face in color affective characteristics, color harmony analysis matrix feature, edge feature, textural characteristics, color variation characteristic and placard.
The computing method of color affective characteristics are as follows.Color emotion is commonly used to the emotion of Description Image.In the calculating of color affective characteristics, first, by RGB color space conversion in CLELAB and CLELCH color space, three factors relevant with color affective characteristics are: temperature (heat), importance (weight) and activity (activity), and the computing method of these three factors are as follows:
a c t i v i t y = - 2.1 + 0.06 [ ( a * - 3 ) 2 + ( L * - 50 ) 2 + ( b * - 17 1.4 ) 2 ] 1 / 2
weight=-1.8+0.45cos(h-10°)+0.04(100-L *)
heat=-0.5+0.02(C *) 1.07cos(h-50°)
Wherein, (L*, C*, h) and (L*, a *, b *) be the color component of color space CIELCH and CIELAB respectively.
Color affective characteristics EI (x, y) that the present invention adopts is defined as:
E I ( x , y ) = activity 2 + weight 2 + heat 2
The computing method of color harmony analysis matrix feature are as follows.The harmony analysis matrix feature of color is commonly used to the emotion of Description Image equally.In the calculating of color affective characteristics, first, by RGB color space conversion to CLELAB color space, the harmony analysis matrix factor relevant to color harmony analysis matrix feature comprises: tone factor H h(hueeffect), luminance factor H land saturation degree factor H (lightnesseffect) c(chromaticeffect):
H L=H Lsum+H ΔL
H Lsum=0.28+0.54tanh(-3.88+0.029ΔL sum)
L sum=L 1 *+L 2 *
H ΔL=0.14+0.15tanh(-2+0.2ΔL)
ΔL=|L * 1-L* 2|
H H=H SY1+H SY2
H SY=E C(H S+E Y)
E C=0.5+0.5tanh(-2+0.5C ab *)
H s=0.08-0.14sin(h ab+50°)-0.07sin(2h ab+90°)
Wherein, h aband C * abwhat represent is tone in CIELAB color space and color saturation, Δ C ab *with Δ H * abthen two kinds of colors look secondary colo(u)rs in cielab color space respectively
Saturation degree difference, L 1 *and L * 2color pair brightness value in cielab color space respectively.
Overall color harmony analysis matrix feature is then by tone factor H h, luminance factor H lwith saturation degree factor H cjoin together to obtain:
CH=H H+H C+H L
The computing method of edge feature are as follows.Research shows, the perception of the mankind to color can be stated more accurately than RGB color space in hsv color space.So, first by image by RGB color space conversion to hsv color space, then, V passage carries out filtering through Gaussian filter, subsequently the result obtained and edge detector is carried out mask, finally calculates and exceedes the number of the pixel of threshold value.
The computing method of textural characteristics are as follows.Textural characteristics and Image emotional semantic have close relationship.The spatial texture feature of scene meets the distribution of Wei cloth:
w b ( y ) = γ β ( x β ) γ - 1 e - 1 γ ( y β ) γ
Wherein, x is stochastic variable, and (beta, gamma) is Wei cloth distribution parameter.The space structure of parameter in Weibull distribution to image texture has very complete expression, the contrast of what parameter beta represented is image, and the larger picture contrast of its value is larger; The granularity of what parameter γ then represented is image, its value larger expression image granularity is less.
The computing method of color variation characteristic are as follows.Research shows, Luv color space has space unitarity.The present invention's determinant Δ f=det (ρ) represents color variation characteristic.In the computation process of color variation characteristic, first by RGB color space conversion to Luv color space, then obtain colour switching matrix:
ρ = σ L 2 σ L u 2 σ L v 2 σ L u 2 σ u 2 σ u v 2 σ L v 2 σ u v 2 σ v 2
Wherein, σ i 2represent the variance at Luv space i passage, represent the covariance at Luv space i and j passage.
The computation process calculating the number of face in placard is as follows.Do not have normal face in terrified placard, be two faces mostly in love placard, in comedy placard, face number is greater than two.Therefore, the present invention extracts the number of the face in placard to embody the difference of dissimilar film.In the number process calculating face in placard, the model adopting opencv to carry detects the number of face in placard.
2): the feature of every width placard that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of placard.
Step 103: the feature extracting the text of story introduction, the feature of each text that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of text, and idiographic flow as shown in Figure 4.
1) pre-service of the text of story introduction.
First to remove the punctuation mark in the text and stop word.The film that the present embodiment relates to is foreign film, so their story introduction English is write, therefore will carry out the reduction of morphological pattern to English word.
Then word bag model is built.Build word bag model and need Feature Words.Present invention employs the method for information gain to obtain Feature Words.The computing formula of information gain is as follows:
IG(T)=H(C)-H(C|T)
H ( C ) = - Σ i = 1 n P ( C i ) 1 og 2 P ( C i )
H ( C | T ) = - P ( t ) Σ i = 1 n P ( C i | t ) 1 og 2 P ( C i | t ) - P ( t ‾ ) Σ i = 1 n P ( C i | t ‾ ) 1 og 2 P ( C i | t ‾ )
Wherein, p (c i) represent film types c ithe probability occurred, the probability that p (t) representation feature T occurs.P (c i| time t) there is T in expression, class c ithe probability occurred.H (C) represents the entropy of system when there is n kind film types, and H (C|T) represents the entropy reduction of the system when learning feature T.
2): the story introduction word bag model representation of each film is become space vector.The feature of each text that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of text.
Step 104:
With the disaggregated model of placard, prediction is carried out to the placard of film to be measured and obtains type Y1, and then the disaggregated model calling text carries out prediction to the story introduction of film to be measured and obtains type Y2.Finally Y1 and Y2 is carried out " or " operation, namely contrast with the type label of film to be measured, obtain the type of last film to be measured, idiographic flow is as shown in Figure 5.
As long as it is correct for having one to predict the outcome in Y1 and Y2, then will detect that correct result as last predicting the outcome; Otherwise, the result Y1 of placard model prediction is used as last predicting the outcome.
Above-described specific embodiment; object of the present invention, technical scheme and beneficial effect are further described; be understood that; the foregoing is only specific embodiments of the invention; be not limited to the present invention; within the spirit and principles in the present invention all, any amendment made, equivalent replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (5)

1., based on a rapid classification method for the film types of placard and story introduction, the method comprises the following steps:
Step 1: determine the type set belonging to film, sets up the placard training set of various types of film and the training set of story introduction;
Step 2: the feature extracting the placard of film to be measured, the feature of every width placard that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of placard;
Step 3: the feature extracting the text of the story introduction of film to be measured, the feature of each text that utilization obtains and the label Training Support Vector Machines of correspondence thereof obtain the disaggregated model of text;
Step 4: with the disaggregated model of placard, prediction is carried out to the placard of film to be measured and obtains result Y1, and then the disaggregated model calling text carries out prediction to the story introduction of film to be measured and obtains result Y2; Finally Y1 and Y2 is carried out OR operation, namely contrasts with the type label of film to be measured, if one to predict the outcome be correct, then predict that correct type is used as the type of last film to be measured; Otherwise using the type of Y1 as last film to be measured.
2. method according to claim 1, is characterized in that, collects the type of the film in China and foreign countries' video website, determines the type set belonging to film; Collect placard corresponding to film and story introduction, set up the training set of placard and the training set of text respectively.
3. method according to claim 2, is characterized in that, the feature of described placard comprises: the number of face in color affective characteristics, color harmony analysis matrix feature, edge feature, textural characteristics, color variation characteristic and placard.
4. according to method according to claim 3, it is characterized in that, when the text of story introduction is English, the feature extracting the text of story introduction comprises:
Step 4a: remove the punctuation mark in text and stop word;
Step 4b: reduction morphological pattern;
Step 4c: selected characteristic word, sets up word bag model;
Step 4d: the story introduction word bag model representation of each film is become vector space model.
5. according to method according to claim 3, it is characterized in that, use determinant
Δ f=det (ρ) represents described color variation characteristic, in the computation process of color variation characteristic, first by RGB color space conversion to Luv color space, then obtain colour switching matrix; Image is in Luv space, and the covariance matrix that each pixel 3 passages produce is:
ρ = σ L 2 σ L u 2 σ L v 2 σ L u 2 σ u 2 σ u v 2 σ L v 2 σ u v 2 σ v 2
Wherein, represent the variance at Luv space i passage, represent the covariance at Luv space i and j passage.
CN201510592018.XA 2015-09-17 2015-09-17 Quick sorting method for movie types based on poster and plot summary Pending CN105260398A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510592018.XA CN105260398A (en) 2015-09-17 2015-09-17 Quick sorting method for movie types based on poster and plot summary

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510592018.XA CN105260398A (en) 2015-09-17 2015-09-17 Quick sorting method for movie types based on poster and plot summary

Publications (1)

Publication Number Publication Date
CN105260398A true CN105260398A (en) 2016-01-20

Family

ID=55100090

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510592018.XA Pending CN105260398A (en) 2015-09-17 2015-09-17 Quick sorting method for movie types based on poster and plot summary

Country Status (1)

Country Link
CN (1) CN105260398A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108153908A (en) * 2018-01-18 2018-06-12 逄泽沐风 Film poster analysis method and system
CN108898165A (en) * 2018-06-12 2018-11-27 浙江大学 A kind of recognition methods of billboard style
CN109151563A (en) * 2018-08-31 2019-01-04 李叶 Audio intensity modifies platform automatically

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101923561A (en) * 2010-05-24 2010-12-22 中国科学技术信息研究所 Automatic document classifying method
CN101937445A (en) * 2010-05-24 2011-01-05 中国科学技术信息研究所 Automatic file classification system
CN103473340A (en) * 2013-09-23 2013-12-25 江苏刻维科技信息有限公司 Classifying method for internet multimedia contents based on video image
CN104657468A (en) * 2015-02-12 2015-05-27 中国科学院自动化研究所 Fast video classification method based on images and texts

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101923561A (en) * 2010-05-24 2010-12-22 中国科学技术信息研究所 Automatic document classifying method
CN101937445A (en) * 2010-05-24 2011-01-05 中国科学技术信息研究所 Automatic file classification system
CN103473340A (en) * 2013-09-23 2013-12-25 江苏刻维科技信息有限公司 Classifying method for internet multimedia contents based on video image
CN104657468A (en) * 2015-02-12 2015-05-27 中国科学院自动化研究所 Fast video classification method based on images and texts

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108153908A (en) * 2018-01-18 2018-06-12 逄泽沐风 Film poster analysis method and system
CN108898165A (en) * 2018-06-12 2018-11-27 浙江大学 A kind of recognition methods of billboard style
CN109151563A (en) * 2018-08-31 2019-01-04 李叶 Audio intensity modifies platform automatically

Similar Documents

Publication Publication Date Title
CN104657468A (en) Fast video classification method based on images and texts
CN110135269B (en) Fire image detection method based on mixed color model and neural network
CN108875602A (en) Monitor the face identification method based on deep learning under environment
CN106097366B (en) A kind of image processing method based on improved Codebook foreground detection
CN101477633B (en) Method for automatically estimating visual significance of image and video
CN104408745A (en) Real-time smog scene detection method based on video image
CN104268590B (en) The blind image quality evaluating method returned based on complementary combination feature and multiphase
CN102236796A (en) Method and system for sorting defective contents of digital video
CN104978565B (en) A kind of pictograph extracting method of universality
CN110570420B (en) No-reference contrast distortion image quality evaluation method
CN104700405B (en) A kind of foreground detection method and system
CN110119688A (en) A kind of Image emotional semantic classification method using visual attention contract network
CN105260398A (en) Quick sorting method for movie types based on poster and plot summary
CN106127234A (en) The non-reference picture quality appraisement method of feature based dictionary
CN108985298A (en) A kind of human body clothing dividing method based on semantic consistency
CN103985130A (en) Image significance analysis method for complex texture images
CN105488475A (en) Method for detecting human face in mobile phone
CN106910195A (en) A kind of web page layout monitoring method and device
CN103489012A (en) Crowd density detecting method and system based on support vector machine
Fu et al. Fast film genres classification combining poster and synopsis
CN104680189A (en) Pornographic image detection method based on improved bag-of-words model
CN100548030C (en) A kind of news main broadcaster's lens detection method based on the space-time strip pattern analysis
Wang et al. Low-light Images In-the-wild: A Novel Visibility Perception-guided Blind Quality Indicator
Cowie et al. An intelligent system for facial emotion recognition
CN104581379A (en) Video preview image selecting method and device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160120

WD01 Invention patent application deemed withdrawn after publication