CN100559879C - A kind of movie action scene detection method based on story line development model analysis - Google Patents

A kind of movie action scene detection method based on story line development model analysis Download PDF

Info

Publication number
CN100559879C
CN100559879C CN 200710099727 CN200710099727A CN100559879C CN 100559879 C CN100559879 C CN 100559879C CN 200710099727 CN200710099727 CN 200710099727 CN 200710099727 A CN200710099727 A CN 200710099727A CN 100559879 C CN100559879 C CN 100559879C
Authority
CN
China
Prior art keywords
camera lens
frame
calculate
plot
action scene
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN 200710099727
Other languages
Chinese (zh)
Other versions
CN101316362A (en
Inventor
刘安安
李锦涛
张勇东
唐胜
宋砚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huan Rui century film and television media Limited by Share Ltd
Original Assignee
Institute of Computing Technology of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Computing Technology of CAS filed Critical Institute of Computing Technology of CAS
Priority to CN 200710099727 priority Critical patent/CN100559879C/en
Publication of CN101316362A publication Critical patent/CN101316362A/en
Application granted granted Critical
Publication of CN100559879C publication Critical patent/CN100559879C/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)
  • Studio Devices (AREA)

Abstract

The invention discloses a kind of movie action scene detection method, comprising: video is done pretreatment operation based on story line development model analysis; Calculate the lens length of each camera lens; Calculate the mean motion intensity of camera lens; Utilize lens length and camera lens mean motion intensity to calculate the move edit factor; Calculate the audio power in short-term of each audio frame and the average audio energy of camera lens; Calculate the mean motion decentralization of camera lens; Utilize camera lens average audio energy and camera lens mean motion decentralization, calculate the human perception factor; According to the move edit factor and the human perception factor, set up the plot progressions model, and generate plot development flow graph according to time sequencing; According to the action scene in the plot progressions model detection film.The invention has the advantages that from move edit gimmick and two angles of human perception and consider vision and sense of hearing factor, set up the plot progressions model, thereby the development and change of simulation plot have realized the accurate detection of action scene in the film.

Description

A kind of movie action scene detection method based on story line development model analysis
Technical field
The present invention relates to video analysis and detection range, particularly a kind of movie action scene detection method based on story line development model analysis.
Background technology
Along with current Digital Media and rapid development of network, people can be in home-confinedly and enjoy various multimedia messagess.Nowadays in the world film industry become video production one enliven strength, probably have every year 4500 New cinema to occur.Therefore, how these magnanimity informations are effectively analyzed, manage, inquired about and retrieve, become present problem demanding prompt solution.
In film, because action scene has strong dramatic conflicts and fierce visual effect usually, be the emphasis of move edit therefore, also be the scene that spectators are most interested in.Detect the action scene in the film need be by the method for movie contents analysis.Present movie contents analytical method is mainly divided two kinds:
1, detects the certain semantic incident by analyzing low-level features such as move edit Rule Extraction video/audio;
2, extract low-level feature detection certain semantic incidents such as video/audio by analyzing human attention change factor.
The groundwork of early stage content-based video analysis is that video structural is retrieved with similar fragment, and video data is mainly news, the video of structure comparison rule such as physical culture.But because " semantic wide gap " between vision and sense of hearing low-level feature and the human senior semanteme, this video analysis and retrieval mode can not satisfy human demand.Along with to video semanteme analysis and research deeply and in news, physical culture etc. have application in the video of regular texture, the researcher is placed on more attention on the semantic fact retrieval of film.But film video is a kind of artistic rendition of height, have very complicated plot and potential edit pattern, the incomprehensive semantic event detection accuracy rate that caused that existing method is described certain semantic is lower, causes failing to reflect the of photoplay plot development own comprehensively.
Generally speaking, main at present accuracy and the robustness that has semantic event detection in two The key factor restriction films:
1, to the deep understanding of move edit technology and human attention model, low-level features such as excavation video/audio characterize these factors;
2, comprehensive objectively move edit mode and subjective spectators' perception two aspect factors are set up the development that rational model characterizes plot.
Summary of the invention
The objective of the invention is to overcome existing movie contents analytical method and pay close attention to the semantic incident of film from move edit gimmick or human perception angle merely, the defective low to the semantic event detection accuracy rate, thereby provide a kind of fusion move edit gimmick and human perception two angles, can realize movie contents analytical method to the movie action scene detection, have high accuracy and robustness, help the move edit personnel the editor of film and spectators selection to scene of interest in the film.
To achieve these goals, the invention provides a kind of movie action scene detection method, carry out according to the following steps order based on story line development model analysis:
Step 10), extract frame of video, frame of video is done pretreatment operation, obtain key frame, the scene in camera lens, the camera lens, the macroblock motion vector of video image by pretreatment operation from original video;
Step 20), calculate the number that comprises frame of video in each camera lens, thereby the acquisition lens length;
Step 30), the macroblock motion vector that utilizes step 10) to obtain, calculate the mean motion intensity of camera lens;
Step 40), utilize step 20) resulting lens length and step 30) the resulting camera lens mean motion intensity calculating move edit factor;
Step 50), extract audio frame, calculate the audio power in short-term of each audio frame from original video, and by calculating in the camera lens average audio energy of the mean value computation camera lens of audio power in short-term;
Step 60), calculate the mean motion decentralization of camera lens;
Step 70), utilize step 50) the camera lens average audio energy and the step 60 that obtain) the camera lens mean motion decentralization that obtains, calculate the human perception factor;
Step 80), according to step 40) the move edit factor and the step 70 that obtain) the human perception factor that obtains, set up the plot progressions model, and generate plot development flow graph according to time sequencing;
Step 90), according to step 80) plot that obtains development flow graph detects the action scene in the film.
In the technique scheme, in described step 30) in, the mean motion intensity of calculating camera lens specifically may further comprise the steps:
Step 31), calculate the energy of the motion vector of macro block in interior all the P frames of camera lens;
Step 32), set up a template, for foreground macro block in the image and background macro block are given different weights;
Step 33), according to step 32) in the weights given for macro block, calculate the exercise intensity of each P frame in the camera lens;
Step 34), with the exercise intensity sum of all P frames in the camera lens divided by the P frame number in the camera lens, obtain the mean motion intensity of camera lens.
In described step 31) in, ask that macroblock motion vector energy calculation formula is in the P frame:
E MV ( i , j ) = x i , j 2 + y i , j 2
Wherein, i, j represent the residing position of macro block.
Described step 32) specifically comprise following operation:
Step 32-1), with the left and right edges of P two field picture background parts as image, calculate the average of motion vector energy in the background, obtain the background motion intensity of P frame;
Step 32-2), calculate the average μ and the variances sigma of all P frame background motion intensity in the camera lens, obtain the threshold value Th of described template according to result calculated, the computing formula of this threshold value is as follows:
Th=μ+a*σ
Wherein, a is an empirical value, and value is 3;
Step 32-3), by step 32-2) resulting threshold value Th, set up template, in template, give different weights for the different macro blocks in the image, described template is:
If E MV ( i , j ) > Th , weight ( x i , j , y i , j ) = 2 ; If E MV ( i , j ) < = Th , weight ( x i , j , y i , j ) = 1 ;
Weight (x wherein I, j, y I, j) for giving the weights of corresponding macro block in the template.
In described step 33) in, the exercise intensity that calculates the P frame is meant: the summation of the product of the weights that the motion vector energy of each macro block of P frame is corresponding with it is as the exercise intensity of this P frame, and the computing formula of described exercise intensity is:
E MV AVE ( P w ) = &Sigma; i = 1 M &Sigma; j = 1 N E MV ( i , j ) * weight ( x i , j , y i , j )
Wherein, M and N represent the macroblock number of level and vertical direction respectively.
In the technique scheme, in described step 40) in, before calculating the move edit factor, described lens length and described camera lens mean motion intensity are done normalized.
In the technique scheme, the calculating of the described move edit factor is the product that the product of lens length and α is added camera lens mean intensity and β, and the value of described α and β is 0.5.
Described step 60) concrete operations are as follows:
Step 61), calculation procedure 32) the weights in the template set up be the direction of the motion vector of the macro block of " 2 ";
Step 62), two dimensional surface is divided into [90,0), [0,90), [90,180), [180,270) four sub spaces, calculate the four-dimensional direction histogram of every two field picture, each H[i of Wesy] (i=1,2,3,4) expression;
Step 63), calculate the motion decentralization of each frame, the motion decentralization of frame represents that with MD its computing formula is:
MD = &Pi; i = 1 4 H [ i ] ;
Step 64), according to step 63) result of calculation, ask the mean motion decentralization of camera lens, camera lens mean motion decentralization MD AVEExpression, then its computing formula is:
MD AVE = 1 Q &Sigma; w = 1 Q MD ( P w )
Wherein, Q represents P frame number in each camera lens.
In the technique scheme, in described step 70) in, before calculating the described human perception factor, described camera lens average audio energy and described camera lens mean motion decentralization are done normalized.
In the technique scheme, the calculating of the described human perception factor is the product that the product of camera lens average audio energy and γ is added camera lens mean motion decentralization and λ, and wherein, described γ and λ value all are 0.5.
In the technique scheme, described step 80) may further comprise the steps:
Step 81), the described move edit factor and the described human perception factor are done linear the fusion, set up the plot progressions model, described plot progressions model is the product that the product of the described move edit factor and φ is added the described human perception factor and ψ, wherein, the value of described φ and ψ is 0.5.
Step 82), with step 81) the plot progressions model set up changes continuously with the No.1 n of mirror, forms plot development flow graph.
With the level and smooth described plot development flow graph of Gauss's template.
The coefficient of described Gauss's template is: window size is 9, and standard deviation is 1.5.
In the technique scheme, described step 90) comprising: at first handle resulting scene information, add up the number N that plot development flow graph value in the scene surpasses first threshold according to the video preprocessor of step 10); Whether judge described N then greater than second threshold value, if greater than, then this scene is an action scene, otherwise this scene is non-action scene.
In the value of plot that described first thresholding is camera lens development flow graph 1/3rd of all local peak maximum, described second threshold value is 2.
The invention has the advantages that from move edit gimmick and two angles of human perception and take all factors into consideration vision and sense of hearing factor, set up the plot progressions model, thereby the development and change of simulation plot have realized the accurate detection of action scene in the film.
Description of drawings
Fig. 1 is the flow chart of the movie action scene detection method based on story line development model analysis of the present invention;
Fig. 2 is to image background schematic diagram partly when setting up template in the embodiment of the invention;
Fig. 3 (a) is the schematic diagram of plot development flow graph;
Fig. 3 (b) is the schematic diagram of the plot development flow graph after level and smooth through Gauss.
Embodiment
Below in conjunction with the drawings and specific embodiments the present invention is described in further detail:
As shown in Figure 1, the movie action scene detection method based on story line development model analysis of the present invention specifically may further comprise the steps:
Step 10, film video is done pretreatment operation.In complete film video, comprise frame of video that is used for token image and the audio frame that is used to characterize sound, in this step, mainly be that the frame of video in the film video is partly carried out preliminary treatment.Pretreatment operation specifically may further comprise the steps:
Step 11, realization are cut apart camera lens, and the camera lens among the present invention is meant the continuous video frames that has similar content in the video;
Key frame in step 12, the extraction camera lens, key frame is the frame of video that best embodies lens features in the camera lens, includes at least one key frame in a camera lens;
Step 13, realization are cut apart film scene, and film scene is the combination with continuous several camera lenses of similar content;
Macroblock motion vector in step 14, the extraction video image.
Above-mentioned steps 11 pretreatment operation to video in the step 14 all belongs to ripe prior art, at list of references 1 " Yueting Zhuang, Yong Rui, Thomas S.Huang et al.Adaptive key frameextraction using unsupervised clustering.Image Processing, ICIP 1998 and Zeeshan Rasheed, Mubarak Shah.Detection and Representation of Scenes inVideos.IEEE Transaction on Multimedia, Vol7, NO.6, December, 2005 " in pair detailed description of step 11-13 is arranged.And step 14 can realize by extract motion vector from MPEG video compression territory.
Step 20, calculate the number that comprises frame of video in each camera lens, thus acquisition lens length information.
The mean motion intensity of step 30, calculating camera lens, the computational methods of camera lens mean motion intensity are as follows:
Step 31, calculate in the camera lens energy of the motion vector of macro block in all P frames, in MPEG video compression territory, the P frame is the frame that is used for forward prediction, asks in the P frame macroblock motion vector energy calculation formula as follows:
E MV ( i , j ) = x i , j 2 + y i , j 2
Wherein, i, j represent the residing position of macro block.
Step 32, set up a template, give different weights to foreground macro block and background macro block, its implementation method is as follows:
Step 32-1, calculate the background motion intensity of each P frame, as shown in Figure 2, dash area in the drawings is considered as the background parts of a P frame, represents with F, and then calculating background motion intensity is exactly the average of calculating motion vector energy among the F of shadow region, uses BE MV AVE(P k) expression, its computing formula is as follows:
BE MV AVE ( P k ) = 1 L &Sigma; ( x i , j , y i , j ) &Element; F E MV ( i , j )
Step 32-2, calculate the average μ and the variances sigma of all P frame background motion intensity in the video, obtain the threshold value of template according to result calculated, the computing formula of this threshold value is as follows:
Th=μ+a*σ
Wherein, a is an empirical value, and value is 3.
Step 32-3, by the resulting threshold value Th of step 32-2, set up as lower bolster:
If E MV ( i , j ) > Th , weight ( x i , j , y i , j ) = 2 ; If E MV ( i , j ) < = Th , weight ( x i , j , y i , j ) = 1 ;
Weight (x wherein I, j, y I, j) for giving the weights of corresponding macro block in the template.
Step 33, according to the weights of giving for macro block in the step 32, calculate the exercise intensity of each P frame in the camera lens, suppose that a certain P frame is W P frame in the camera lens, then the computing formula of its exercise intensity is:
E MV AVE ( P w ) = &Sigma; i = 1 M &Sigma; j = 1 N E MV ( i , j ) * weight ( x i , j , y i , j )
Wherein, M and N represent the macroblock number of level and vertical direction respectively.
Step 34, according to the result of calculation of step 33, calculate the mean motion intensity of camera lens.The exercise intensity sum that when calculating the mean motion intensity of camera lens is exactly all P frames in the camera lens is divided by the P frame number in the camera lens.Its computing formula is as follows:
E MV AVE = 1 Q &Sigma; w = 1 Q E MV AVE ( P w )
Wherein, Q represents P frame number in each camera lens.
Step 40, resulting lens length of step 20 and the resulting camera lens mean motion of step 30 intensity are done normalized respectively, set up the move edit factor according to two features after the normalization then.With s (n) expression lens length, with m (n) expression camera lens mean motion intensity, represent frame number with n, use P 1(n) the expression move edit factor, then the computing formula of this factor is as follows:
P 1(n)=α*s(n)+β*m(n)
α=β=0.5
Resulting move edit factor representation the influence that movie action scene is detected of move edit technology.
Step 50, from film video, extract audio frame, calculate the audio power in short-term of each audio frame, and by calculating in the camera lens mean value computation camera lens average audio energy of audio power in short-term.
Audio power in short-term in this step is meant the energy summation of each all sampling point of audio frequency short time frame, and its computational methods are at list of references 2 " Bai Liang; Hu Yaali, Feature analysis and extraction for audioautomatic classification, Proc.of IEEE International Conference on Systems, Man andCybernetics, vol.1, pp:767-772,2005. " be documented in.
Step 60, calculating camera lens mean motion decentralization, camera lens mean motion decentralization is used for representing the average complexity that this camera lens vision content changes.Calculating to camera lens mean motion decentralization comprises following specific implementation:
Step 61, calculation procedure 32 are set up the direction for the motion vector of the macro block of " 2 " of weights in the template, and its computing formula is as follows:
If x i , j > 0 , y &NotEqual; 0 , &theta; = arctan ( y i , j / x i , j ) ; If x i , j < 0 , y &NotEqual; 0 , &theta; = arctan ( y i , j / x i , j ) + 180 ; If x i , j = 0 , y i , j > 0 , &theta; = 90 ; If x i , j = 0 , y i , j < 0 , &theta; = - 90 ; If y i , j = 0 , x i , j > 0 , &theta; = 0 ; If y i , j = 0 , x i , j = < 0 , &theta; = 180 ; If x i , j = 0 , y i , j = 0 , omitting&theta; ;
θ wherein represents the direction of motion vector.
Step 62, two dimensional surface is divided into [90,0), [0,90), [90,180), [180,270) four sub spaces, calculate the four-dimensional direction histogram of every two field picture, each H[i of Wesy] (i=1,2,3,4) expression.Calculating four-dimensional direction histogram is exactly to calculate the number of the motion vector of angle in each sub spaces and the ratio of motion vector sum.
Step 63, calculate the motion decentralization of each frame, the motion decentralization of frame represents that with MD its computing formula is as follows:
MD = &Pi; i = 1 4 H [ i ]
Step 64, according to the result of calculation of step 63, ask the mean motion decentralization of camera lens, camera lens mean motion decentralization MD AVEExpression, then its computing formula is as follows:
MD AVE = 1 Q &Sigma; w = 1 Q MD ( P w )
Wherein, Q represents P frame number in each camera lens.
The camera lens mean motion decentralization that step 70, the camera lens average audio energy that step 50 is obtained and step 60 obtain is done the normalization operation respectively, sets up the human perception factor according to two features after the normalization then.With a (n) expression camera lens average audio energy, with d (n) expression camera lens mean motion decentralization, represent that with n mirror is No.1, use P 2(n) the human perception factor of expression, the computing formula of the then human perception factor is as follows:
P 2(n)=γ*a(n)+λ*d(n)
γ=λ=0.5
Resulting human perception factor representation human attention influence that movie action scene is detected.
The human perception factor that step 80, the move edit factor that obtains according to step 40 and step 70 obtain is set up the plot progressions model.The plot progressions model can be used for characterizing the importance of each construction unit content in the video and the mankind being attracted intensity of force.The specific implementation step of setting up the plot progressions model is as follows:
Step 81, the move edit factor and the human perception factor are done linear the fusion, set up the plot progressions model, its computing formula is as follows:
M(n)=φ*P 1(n)+ψ*P 2(n)
φ=ψ=0.5
Step 82, the plot progressions model set up form plot development flow graph along with the continuous variation of the No.1 n of mirror.Plot development flow graph has reflected the importance of different camera lenses in the whole video, and it compares the importance of whole video different units in chronological order, embodies the difference of importance of different camera lenses, has showed the development and change of plot.Fig. 3 (a) is exactly the example of a plot development flow graph.
Step 83, the level and smooth plot development of usefulness Gauss template flow graph, wherein the coefficient of Gauss's template is: window size is 9, standard deviation is 1.5.The plot development flow graph of Fig. 3 (b) after to be Fig. 3 (a) through Gauss level and smooth.
Action scene in step 90, the detection film.During action scene in detecting film, scene carve information at first handling according to video preprocessor add up in the scene plot development flow graph M (n) value above threshold value Th 1Number N; Judge that then N is greater than threshold value Th 2, then this scene is an action scene; Otherwise, be non-action scene.Th wherein 1And Th 2Determine by experiment, in the present embodiment, Th 1Be 1/3rd of all local peak maximum in each camera lens M (n) value; Th 1Value be 2.

Claims (15)

1, a kind of movie action scene detection method based on story line development model analysis, carry out according to the following steps order:
Step 10), extract frame of video, frame of video is done pretreatment operation, obtain key frame, the scene in camera lens, the camera lens, the macroblock motion vector of video image by pretreatment operation from original video;
Step 20), calculate the number that comprises frame of video in each camera lens, thereby the acquisition lens length;
Step 30), the macroblock motion vector that utilizes step 10) to obtain, calculate the mean motion intensity of camera lens;
Step 40), utilize step 20) resulting lens length and step 30) the resulting camera lens mean motion intensity calculating move edit factor;
Step 50), extract audio frame, calculate the audio power in short-term of each audio frame from original video, and by calculating in the camera lens average audio energy of the mean value computation camera lens of audio power in short-term;
Step 60), calculate the mean motion decentralization of camera lens;
Step 70), utilize step 50) the camera lens average audio energy and the step 60 that obtain) the camera lens mean motion decentralization that obtains, calculate the human perception factor;
Step 80), according to step 40) the move edit factor and the step 70 that obtain) the human perception factor that obtains, set up the plot progressions model, and generate plot development flow graph according to time sequencing;
Step 90), according to step 80) plot that obtains development flow graph detects the action scene in the film.
2, the movie action scene detection method based on story line development model analysis according to claim 1 is characterized in that, in described step 30) in, the mean motion intensity of calculating camera lens specifically may further comprise the steps:
Step 31), calculate the energy of the motion vector of macro block in interior all the P frames of camera lens;
Step 32), set up a template, for foreground macro block in the image and background macro block are given different weights;
Step 33), according to step 32) in the weights given for macro block, calculate the exercise intensity of each P frame in the camera lens;
Step 34), with the exercise intensity sum of all P frames in the camera lens divided by the P frame number in the camera lens, obtain the mean motion intensity of camera lens.
3, the movie action scene detection method based on story line development model analysis according to claim 2 is characterized in that, in described step 31) in, ask that macroblock motion vector energy calculation formula is in the P frame:
E MV ( i , j ) = x i , j 2 + y i , j 2
Wherein, i, j represent the residing position of macro block.
4, the movie action scene detection method based on story line development model analysis according to claim 2 is characterized in that, described step 32) specifically comprise following operation:
Step 32-1), with the left and right edges of P two field picture background parts as image, calculate the average of motion vector energy in the background, obtain the background motion intensity of P frame;
Step 32-2), calculate the average μ and the variances sigma of all P frame background motion intensity in the camera lens, obtain the threshold value Th of described template according to result calculated, the computing formula of this threshold value is as follows:
Th=μ+a*σ
Wherein, a is an empirical value, and value is 3;
Step 32-3), by step 32-2) resulting threshold value Th, set up template, in template, give different weights for the different macro blocks in the image, described template is:
If E MV ( i , j ) > Th , weight ( x i , j , y i , j ) = 2 ; If E MV ( i , j ) < = Th , weight ( x i , j , y i , j ) = 1 ;
Weight (x wherein I, j, y I, j) for giving the weights of corresponding macro block in the template.
5, the movie action scene detection method based on story line development model analysis according to claim 4, it is characterized in that, in described step 33) in, the exercise intensity that calculates the P frame is meant: the summation of the product of the weights that the motion vector energy of each macro block of P frame is corresponding with it is as the exercise intensity of this P frame, and the computing formula of described exercise intensity is:
E MV AVE ( P w ) = &Sigma; i = 1 M &Sigma; j = 1 M E MV ( i , j ) * weight ( x i , j , y i , j )
Wherein, M and N represent the macroblock number of level and vertical direction respectively.
6, the movie action scene detection method based on story line development model analysis according to claim 1, it is characterized in that, in described step 40) in, before calculating the move edit factor, described lens length and described camera lens mean motion intensity are done normalized.
7, according to claim 1 or 6 described movie action scene detection methods based on story line development model analysis, it is characterized in that, the calculating of the described move edit factor is the product that the product of lens length and α is added camera lens mean intensity and β, and the value of described α and β is 0.5.
8, the movie action scene detection method based on story line development model analysis according to claim 4 is characterized in that, described step 60) concrete operations as follows:
Step 61), calculation procedure 32) the weights in the template set up be the direction of the motion vector of the macro block of " 2 ";
Step 62), two dimensional surface is divided into [90,0), [0,90), [90,180), [180,270) four sub spaces, calculate the four-dimensional direction histogram of every two field picture, each H[i of Wesy] (i=1,2,3,4) expression;
Step 63), calculate the motion decentralization of each frame, the motion decentralization of frame represents that with MD its computing formula is:
MD = &Pi; i = 1 4 H [ i ] ;
Step 64), according to step 63) result of calculation, ask the mean motion decentralization of camera lens, camera lens mean motion decentralization MD AVEExpression, then its computing formula is:
MD AVE = 1 Q &Sigma; w = 1 Q MD ( P w )
Wherein, Q represents P frame number in each camera lens.
9, the movie action scene detection method based on story line development model analysis according to claim 1, it is characterized in that, in described step 70) in, before calculating the described human perception factor, described camera lens average audio energy and described camera lens mean motion decentralization are done normalized.
10, according to claim 1 or 9 described movie action scene detection methods based on story line development model analysis, it is characterized in that, the calculating of the described human perception factor is the product that the product of camera lens average audio energy and γ is added camera lens mean motion decentralization and λ, wherein, described γ and λ value all are 0.5.
11, the movie action scene detection method based on story line development model analysis according to claim 1 is characterized in that, described step 80) may further comprise the steps:
Step 81), the described move edit factor and the described human perception factor are done linear the fusion, set up the plot progressions model, described plot progressions model is the product that the product of the described move edit factor and φ is added the described human perception factor and ψ, wherein, the value of described φ and ψ is 0.5.
Step 82), with step 81) the plot progressions model set up changes continuously with the No.1 n of mirror, forms plot development flow graph.
12, the movie action scene detection method based on story line development model analysis according to claim 11 is characterized in that, with the level and smooth described plot development flow graph of Gauss's template.
13, the movie action scene detection method based on story line development model analysis according to claim 12 is characterized in that, the coefficient of described Gauss's template is: window size is 9, and standard deviation is 1.5.
14, the movie action scene detection method based on story line development model analysis according to claim 1, it is characterized in that, described step 90) comprising: at first handle resulting scene information, add up the number N that the development of plot in scene flow graph value surpasses first threshold according to the video preprocessor of step 10); Whether judge described N then greater than second threshold value, if greater than, then this scene is an action scene, otherwise this scene is non-action scene.
15, the movie action scene detection method based on story line development model analysis according to claim 14, it is characterized in that, in the value of plot that described first thresholding is camera lens development flow graph 1/3rd of all local peak maximum, described second threshold value is 2.
CN 200710099727 2007-05-29 2007-05-29 A kind of movie action scene detection method based on story line development model analysis Expired - Fee Related CN100559879C (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 200710099727 CN100559879C (en) 2007-05-29 2007-05-29 A kind of movie action scene detection method based on story line development model analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 200710099727 CN100559879C (en) 2007-05-29 2007-05-29 A kind of movie action scene detection method based on story line development model analysis

Publications (2)

Publication Number Publication Date
CN101316362A CN101316362A (en) 2008-12-03
CN100559879C true CN100559879C (en) 2009-11-11

Family

ID=40107177

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 200710099727 Expired - Fee Related CN100559879C (en) 2007-05-29 2007-05-29 A kind of movie action scene detection method based on story line development model analysis

Country Status (1)

Country Link
CN (1) CN100559879C (en)

Families Citing this family (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101477798B (en) * 2009-02-17 2011-01-05 北京邮电大学 Method for analyzing and extracting audio data of set scene
CN102724385B (en) * 2012-06-21 2016-05-11 浙江宇视科技有限公司 A kind of Intelligent video analysis method and device
CN103458279B (en) * 2013-08-27 2016-08-10 西北工业大学 A kind of captivation decision method based on advertisement video image data
CN104713187B (en) * 2013-12-11 2017-07-21 海尔集团公司 Air-conditioning sets off the control method and air-conditioning of scene by contrast according to picture
CN104008175B (en) * 2014-06-06 2017-03-08 福州大学 A kind of key frame of video self-adaptation extraction method under emotional
CN105847964A (en) * 2016-03-28 2016-08-10 乐视控股(北京)有限公司 Movie and television program processing method and movie and television program processing system
CN107295284B (en) * 2017-08-03 2020-08-25 浙江大学 Method and device for generating, retrieving and playing video file consisting of audio and picture
CN109525892B (en) * 2018-12-03 2021-09-10 易视腾科技股份有限公司 Video key scene extraction method and device
EP3984235A4 (en) 2019-06-11 2022-12-28 Wemovie Technologies Production-as-service systems for making movies, tv shows and multimedia contents
WO2021022499A1 (en) 2019-08-07 2021-02-11 WeMovie Technologies Adaptive marketing in cloud-based content production
WO2021068105A1 (en) 2019-10-08 2021-04-15 WeMovie Technologies Pre-production systems for making movies, tv shows and multimedia contents
CN111191010B (en) * 2019-12-31 2023-08-08 天津外国语大学 Movie script multi-element information extraction method
WO2021225608A1 (en) 2020-05-08 2021-11-11 WeMovie Technologies Fully automated post-production editing for movies, tv shows and multimedia contents
US11070888B1 (en) 2020-08-27 2021-07-20 WeMovie Technologies Content structure aware multimedia streaming service for movies, TV shows and multimedia contents
US11812121B2 (en) 2020-10-28 2023-11-07 WeMovie Technologies Automated post-production editing for user-generated multimedia contents
US11166086B1 (en) 2020-10-28 2021-11-02 WeMovie Technologies Automated post-production editing for user-generated multimedia contents
US11330154B1 (en) 2021-07-23 2022-05-10 WeMovie Technologies Automated coordination in multimedia content production
US11321639B1 (en) 2021-12-13 2022-05-03 WeMovie Technologies Automated evaluation of acting performance using cloud services

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1510907A (en) * 2002-12-24 2004-07-07 中国科学院计算技术研究所 Method for detecting exciting part in sports game video frequency

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1510907A (en) * 2002-12-24 2004-07-07 中国科学院计算技术研究所 Method for detecting exciting part in sports game video frequency

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
A Model of Saliency-Based Visual Attention for Rapid Scene Analysis. Laurent Itti et al.IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,Vol.20 No.11. 1998 *
Effective Video Scene Detection Approach Based on Cinematic Rules. R.Khosla et al.KES 2005,LNAI 3682. 2005 *
Two Important Action Scenes Detection Based on Probability Neural Networks. J. Wang et al.ISNN 2006 LNCS 3972. 2006 *
武打片中的动作场景检测方法. 程文刚等.电子学报,第34卷第5期. 2006 *

Also Published As

Publication number Publication date
CN101316362A (en) 2008-12-03

Similar Documents

Publication Publication Date Title
CN100559879C (en) A kind of movie action scene detection method based on story line development model analysis
CN101299241B (en) Method for detecting multi-mode video semantic conception based on tensor representation
CN110097026B (en) Paragraph association rule evaluation method based on multi-dimensional element video segmentation
CN102903119B (en) A kind of method for tracking target and device
US20070030391A1 (en) Apparatus, medium, and method segmenting video sequences based on topic
CN114186069B (en) Depth video understanding knowledge graph construction method based on multi-mode different-composition attention network
CN101316327B (en) Multimode amalgamation covering lens detection method
CN111488487A (en) Advertisement detection method and detection system for all-media data
CN111145222A (en) Fire detection method combining smoke movement trend and textural features
CN111027440B (en) Crowd abnormal behavior detection device and detection method based on neural network
Naphade et al. Semantic video indexing using a probabilistic framework
CN113823323B (en) Audio processing method and device based on convolutional neural network and related equipment
CN101827224B (en) Detection method of anchor shot in news video
Zhu et al. Video scene segmentation and semantic representation using a novel scheme
Yang et al. LCSED: A low complexity CNN based SED model for IoT devices
CN111444913A (en) License plate real-time detection method based on edge-guided sparse attention mechanism
CN102752479B (en) Scene detection method of vegetable diseases
CN113158720A (en) Video abstraction method and device based on dual-mode feature and attention mechanism
Kim et al. An effective news anchorperson shot detection method based on adaptive audio/visual model generation
Prabakaran et al. Key Frame Extraction Analysis Based on Optimized Convolution Neural Network (OCNN) using Intensity Feature Selection (IFS)
Beigh et al. Object-Based Key Frame Extraction in Videos
CN115294359B (en) High-resolution image saliency target detection method based on deep learning
Huang et al. Learning Where to Cut from Edited Videos
Kumano et al. Video editing support system based on video content analysis
Vuorinen et al. Unsupervised speaker change detection for mobile device recorded speech

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
ASS Succession or assignment of patent right

Owner name: H+R CENTURY PICTURES CO., LTD.

Free format text: FORMER OWNER: INSTITUTE OF COMPUTING TECHNOLOGY, CHINESE ACADEMY OF SCIENCES

Effective date: 20121019

C41 Transfer of patent application or patent right or utility model
COR Change of bibliographic data

Free format text: CORRECT: ADDRESS; FROM: 100080 HAIDIAN, BEIJING TO: 322118 JINHUA, ZHEJIANG PROVINCE

TR01 Transfer of patent right

Effective date of registration: 20121019

Address after: 322118 experimentation area, Hengdian Film & TV industry, Dongyang, Zhejiang, Jinhua C1-008

Patentee after: Huan Rui century film and television media Limited by Share Ltd

Address before: 100080 Haidian District, Zhongguancun Academy of Sciences, South Road, No. 6, No.

Patentee before: Institute of Computing Technology, Chinese Academy of Sciences

C17 Cessation of patent right
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20091111

Termination date: 20130529