CN108600865A - A kind of video abstraction generating method based on super-pixel segmentation - Google Patents

A kind of video abstraction generating method based on super-pixel segmentation Download PDF

Info

Publication number
CN108600865A
CN108600865A CN201810456341.8A CN201810456341A CN108600865A CN 108600865 A CN108600865 A CN 108600865A CN 201810456341 A CN201810456341 A CN 201810456341A CN 108600865 A CN108600865 A CN 108600865A
Authority
CN
China
Prior art keywords
pixel
super
image
video
merging technique
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810456341.8A
Other languages
Chinese (zh)
Other versions
CN108600865B (en
Inventor
金海燕
李喻蒙
肖照林
李秀秀
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shaanxi Zhisheng Desen Elevator Co.,Ltd.
Original Assignee
Xian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Technology filed Critical Xian University of Technology
Priority to CN201810456341.8A priority Critical patent/CN108600865B/en
Publication of CN108600865A publication Critical patent/CN108600865A/en
Application granted granted Critical
Publication of CN108600865B publication Critical patent/CN108600865B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8549Creating video summaries, e.g. movie trailer

Landscapes

  • Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of video abstraction generating methods based on super-pixel segmentation, the segmentation of video boundaries by slightly arrive it is thin in a manner of complete, partitioning boundary tends to Local Minimum moving region, make boundary alignment to the part for the position for being suitable for cutting, the partitioning boundary of video is extracted as the key frame extracted, indicate that the key activities between action sequence are extracted, represent the effective information of video, the operand and complexity of video can be greatly lowered, the real-time for improving video analysis has a significant impact;The image block that will have many characteristics, such as similar grain, color, the adjacent pixel composition of brightness using the method for super-pixel region merging technique, by the similarity degree of feature between pixel by group pixels, the redundancy that image can be obtained reduces the complexity of subsequent image processing task;The Similarity measures of image are carried out using spatial organization's relationship between pixel and generate video frequency abstract to eliminate redundancy key frames, and video frequency abstract effect is preferable.

Description

A kind of video abstraction generating method based on super-pixel segmentation
Technical field
The invention belongs to technical field of image processing, and in particular to a kind of video frequency abstract generation side based on super-pixel segmentation Method.
Background technology
The research early start of video summarization technique is in the Informedia engineerings of Carnegie Mellon University in 1994.From that After, researcher is added more and more, and the technology of video frequency abstract related fields is studied and explored, overall For, for video summarization technique according to the abstract form of output, common video frequency abstract generating algorithm mainly has data clusters at present Method, curve law of planning and machine learning method.
In the prior art, IEEE International Conference on Acoustics, Speech and Signal Processing.IEEE,2014:1260-1264. is disclosed original video is decomposed into image sequence after, extraction it is each The color moment characteristics of frame image, and video is divided into several camera lenses, then use the spectral clustering based on rough set theory To clustering shots.But such algorithm existing defects:First, most of clustering algorithm is required for presetting the number of cluster Mesh, optimal clusters number, generally requires constantly to test repeatedly in order to obtain.And optimal clusters number is dynamic change , and input original video length and type have relationship, cause algorithm that must be executed in the case of manual intervention.The Two, whether the feature of extraction is representative often to influence Clustering Effect.It is only extracted the color characteristic of image, has ignored figure The shape and Texture eigenvalue of picture, and single feature cannot comprehensively express the visual information of image.
Proceedings of IEEE International Conference on Multimedia and Expo.Washington DC,USA:IEEE,2005:In 670-673, after original video is decomposed into image sequence, calculate Pixel difference between frame and frame, and pixel difference is fitted to curve, in order to simplify curve, broken line is divided into the collection of a broken line It closes, and the small point of correlation is deleted from set.Curve law of planning intuitively shows video using the curve in coordinate system Content, can simplify the processing to video, however this method may only reflecting video content variation, it is not possible to complete expression Semantic information.
Nature and Biologically Inspired Computing.IEEE,2011:In 7-11, SVM pairs is utilized Playback mark in football video is trained and learns, and identification playback camera lens, the generation of goal event is corresponded to be indicated in playback Occur the scenes such as stopping, portrait attachment, crowd shots, goal area and audio excitement successively before, and then can after playback indicates There are the scenes such as full length shot, scoreboard.By the playback camera lens detected and judge whether its front and back scene meets the above-mentioned fact To carry out football video Context event monitoring.Machine learning method is established model and is required in the case of the overwhelming majority manually Participate in auxiliary modeling, and this method to modeling when Feature Selection it is more sensitive, if suitable feature is had chosen, using machine Device learning method can get the high-level semantics information for meeting human cognitive.Conversely, unsuitable Feature Selection will make learning outcome not People's will to the greatest extent.
Invention content
The purpose of the present invention is to provide a kind of video abstraction generating methods based on super-pixel segmentation, can significantly drop The operand and complexity of low video, the real-time for improving video analysis have a significant impact.
The technical solution adopted by the present invention is a kind of video abstraction generating method based on super-pixel segmentation, specifically according to Following steps are implemented:
Step 1 obtains video image, and using video image as input data, extracts video image frame sequence;
Step 2 carries out initial isometric segmentation to video image frame sequence, is divided into multiple isometric frame sequences;
Meanwhile the light stream of video image frame sequence is calculated, and use the light stream offset estimation between video image frame sequence Motion amplitude as the function of time;
Step 3 is directed to each isometric frame sequence, and the motion amplitude for passing through the function of time finds initial isometric partitioning boundary frame The frame of partial gradient value minimum is as key frame;
Step 4 carries out Local Clustering to the pixel in every width key frame images, using the similarity pair of feature between pixel Key frame images carry out super-pixel segmentation, obtain multiple super-pixel;
Step 5 carries out region merging technique to multiple super-pixel, obtains super-pixel region merging technique image;
Step 6 carries out similarity measurement to adjacent two width super-pixel region merging technique image, judges adjacent Liang Fu super-pixel area Whether domain merges image similar;
One in step 7, the adjacent similar super-pixel region merging technique image of deletion, by remaining super-pixel region merging technique figure As generating video frequency abstract after combination.
The present invention is further characterized in that:
Motion amplitude detailed process of the step 2 using the light stream offset estimation between successive frame as the function of time be:It is logical The light stream in all pixels of each video image frame in aggregation level and vertical direction is crossed to calculate video image frame sequence The amplitude of movement, calculation formula are as follows:
Wherein, OFx (i, j, t) is the x-component of light stream of the pixel (i, j) between frame t and t-1, and OFy (i, j, t) is picture The y-component of the light stream of plain (i, j) between frame t and t-1;Light stream tracks all the points over time, and summation is between frame Amount of exercise estimation.
The detailed process of step 4 is:
Step 4.1, by every width key frame formation image be converted to 5 dimensional features under CIELAB color spaces and XY coordinates to Then amount constructs module to 5 dimensional feature vectors:
Assuming that it is the super-pixel of K identical sizes that image, which has N number of pixel, pre-segmentation, the size of each super-pixel is N/ K, the super-pixel central point for selecting each pre-segmentation is seed point, and the distance of each two seed point is approximately
The seed point, is moved to the position of Grad minimum by step 4.2, the window that 3 × 3 are divided centered on seed point It sets, while an individual label is distributed for each seed;
Step 4.3, for each pixel, calculate separately the similarity degree for each seed point that distance is nearest therewith, will most phase It is assigned to the pixel like the label of seed point, the pixel with same label forms a super-pixel.
The formula that step 4.3 calculates the similarity degree for each seed point that distance is nearest therewith is as follows:
Wherein, dlabColor distortion between pixel, dxySpace length between pixel, DiFor the phase of two pixels Like degree;S is the spacing of seed point, and m is balance parameters, for weighing the ratio of color value and spatial information in measuring similarity Weight, DiValue is bigger, illustrates that two pixels are more similar.
Step 5 detailed process is:
Step 5.1 presets region merging technique threshold value, it is assumed that super-pixel has K, using non-directed graph G=(V, E) to super picture Plain syntople is recorded;
Wherein V={ S1, S2..., SKBe all K vertex set, E is the set on all boundaries, figure G=(V, E) In, each region is represented as a node in figure;
Step 5.2 chooses any two region (Si, Sj) ∈ V, if Si, SjIt is adjacent, then between corresponding node just There are a lines to be connected, and each side assigns certain weight, calculates weight and corresponds to the cost value for merging two regions;
The weight of cost value minimum merges in step 5.3, selection adjacent area, constantly updates relevant edge weight, And judge whether the areal after the merging is equal to and preset region merging technique threshold value, when equal to presetting region merging technique threshold Value, then merge termination, obtain multiple super-pixel block;
The image that step 5.4, multiple super-pixel block are formed is super-pixel region merging technique image.
Step 5.2 calculates weight:
Wherein, N indicates that region S areas, μ indicate the spectrum mean value in region, and l is the Border of region i and j, and λ is shape Shape parameter.
Step 6 carries out similarity measurement detailed process to adjacent two width super-pixel region merging technique image:
Preset difference value threshold value;
Calculate the difference value between each pixel of adjacent two width super-pixel region merging technique image:
Wherein, i indicates that picture number, i+1 indicate the adjacent image of i images, and a indicates pixel, to any in image The coordinate of pixel a is Ia(x, y), where super-pixel segmentation merge after region be:Ω a=sp (Ia(x, y)), M is indicated The pixel number in the region;
The difference value of all respective pixel points place super-pixel of two width super-pixel region merging technique images is calculated, then adjacent two width Super-pixel region merging technique figure difference value calculation formula is as follows:
Wherein, the boundary of the pixel difference value value during threshold is indicated per adjacent two super-pixel region merging technique image Value.
Step 6 judges whether similar process is adjacent two width super-pixel region merging technique image:Judge adjacent two width super-pixel Whether the difference value of region merging technique image is less than threshold value:
If difference value is less than threshold value, two width super-pixel region merging technique images are similar;
If difference value is not less than threshold value, two width super-pixel region merging technique images are dissimilar.
A kind of video abstraction generating method advantageous effect based on super-pixel segmentation of the present invention is:
(1) present invention in video boundaries segmentation by slightly arrive it is thin in a manner of complete, partitioning boundary tend to Local Minimum transport Dynamic region makes boundary alignment extract the partitioning boundary of video as the key extracted to the part for the position for being suitable for cutting Frame indicates that the key activities between action sequence are extracted, represents the effective information of video, and video can be greatly lowered Operand and complexity, for improve video analysis real-time have a significant impact;
(2) will have many characteristics, such as similar grain, color, the phase of brightness using the method for super-pixel region merging technique in the present invention Adjacent pixel constitutes block of pixels can obtain the redundancy of image by the similarity degree of feature between pixel by group pixels, Largely reduce the complexity of subsequent image processing task;Image is carried out using spatial organization's relationship between pixel Similarity measures generate video frequency abstract to eliminate redundancy key frames, and video frequency abstract effect is preferable.
Description of the drawings
Fig. 1 is a kind of video abstraction generating method flow chart based on super-pixel segmentation of the present invention;
Fig. 2 is key-frame extraction display diagram in the present invention;
Fig. 3 is super-pixel region merging technique process schematic of the present invention.
Specific implementation mode
The following describes the present invention in detail with reference to the accompanying drawings and specific embodiments.
A kind of video abstraction generating method based on super-pixel segmentation of the present invention, as shown in Figure 1, specifically according to the following steps Implement:
Step 1 obtains video image, and using video image as input data, extracts video image frame sequence;
Step 2 carries out initial isometric segmentation to video image frame sequence, is divided into multiple isometric frame sequences;
Meanwhile the light stream of video image frame sequence is calculated, and use the light stream offset estimation between video image frame sequence Motion amplitude as the function of time;
Motion amplitude detailed process using the light stream offset estimation between successive frame as the function of time is:By every What the light stream in all pixels of a video image frame in aggregation level and vertical direction moved to calculate video image frame sequence Amplitude, calculation formula are as follows:
Wherein, OFx (i, j, t) is the x-component of light stream of the pixel (i, j) between frame t and t-1, and OFy (i, j, t) is picture The y-component of the light stream of plain (i, j) between frame t and t-1;Light stream tracks all the points over time, and summation is between frame Amount of exercise estimation.
Step 3 is directed to each isometric frame sequence, and the motion amplitude for passing through the function of time finds initial isometric partitioning boundary frame The frame of partial gradient value minimum is as key frame, as shown in Figure 2;
Step 4 carries out Local Clustering to the pixel in every width key frame images, using the similarity pair of feature between pixel Key frame images carry out super-pixel segmentation, obtain multiple super-pixel;
Detailed process is:
Step 4.1, by every width key frame formation image be converted to 5 dimensional features under CIELAB color spaces and XY coordinates to Then amount constructs module to 5 dimensional feature vectors:
Assuming that it is the super-pixel of K identical sizes that image, which has N number of pixel, pre-segmentation, the size of each super-pixel is N/ K, the super-pixel central point for selecting each pre-segmentation is seed point, and the distance of each two seed point is approximately
The seed point, is moved to the position of Grad minimum by step 4.2, the window that 3 × 3 are divided centered on seed point It sets, while an individual label is distributed for each seed;
Step 4.3, as shown in figure 3, for each pixel, calculate separately the similar journey for each seed point that distance is nearest therewith Degree, the pixel is assigned to by the label of most like seed point, and the pixel with same label forms a super-pixel;
The formula for calculating the similarity degree for each seed point that distance is nearest therewith is as follows:
Wherein, dlabColor distortion between pixel, dxySpace length between pixel, DiFor the phase of two pixels Like degree;S is the spacing of seed point, and m is balance parameters, for weighing the ratio of color value and spatial information in measuring similarity Weight, DiValue is bigger, illustrates that two pixels are more similar.
Step 5 carries out region merging technique to multiple super-pixel, obtains super-pixel region merging technique image;
Detailed process is:
Step 5.1 presets region merging technique threshold value, it is assumed that super-pixel has K, using non-directed graph G=(V, E) to super picture Plain syntople is recorded;
Wherein V={ S1, S2..., SKBe all K vertex set, E is the set on all boundaries, figure G=(V, E) In, each region is represented as a node in figure;
Step 5.2 chooses any two region (Si, Sj) ∈ V, if Si, SjIt is adjacent, then between corresponding node just There are a lines to be connected, and each side assigns certain weight, calculates weight and corresponds to the cost value for merging two regions;
It calculates weight and corresponds to and merge the calculation formula of cost value in two regions and be:
Wherein, N indicates that region S areas, μ indicate the spectrum mean value in region, and l is the Border of region i and j, and λ is shape Shape parameter;
The weight of cost value minimum merges in step 5.3, selection adjacent area, constantly updates relevant edge weight, And judge whether the areal after the merging is equal to and preset region merging technique threshold value, when equal to presetting region merging technique threshold Value, then merge termination, obtain multiple super-pixel block;
The image that step 5.4, multiple super-pixel block are formed is super-pixel region merging technique image.
Step 6 carries out similarity measurement to adjacent two width super-pixel region merging technique image, and detailed process is:
Preset difference value threshold value;
Calculate the difference value between each pixel of adjacent two width super-pixel region merging technique image:
Wherein, i indicates that picture number, i+1 indicate the adjacent image of i images, and a indicates pixel, to any in image The coordinate of pixel a is Ia(x, y), where super-pixel segmentation merge after region be:Ω a=sp (Ia(x, y)), M is indicated The pixel number in the region;
The difference value of all respective pixel points place super-pixel of two width super-pixel region merging technique images is calculated, then adjacent two width Super-pixel region merging technique figure difference value calculation formula is as follows:
Wherein, the boundary of the pixel difference value value during threshold is indicated per adjacent two super-pixel region merging technique image Value.
Judge whether adjacent two width super-pixel region merging technique image is similar, and process is again:Judge adjacent Liang Fu super-pixel area Whether the difference value that domain merges image is less than threshold value:
If difference value is less than threshold value, two width super-pixel region merging technique images are similar;
If difference value is not less than threshold value, two width super-pixel region merging technique images are dissimilar.
One in step 7, the adjacent similar super-pixel region merging technique image of deletion, by remaining super-pixel region merging technique figure As generating video frequency abstract after combination.
Embodiment
Feasibility verification is carried out to the scheme of invention with reference to specific calculation formula, it is described below:
The validity of the put forward algorithm of YouTube database authentications is chosen in experiment.YouTube databases include from video website 50 videos that (such as YouTube) is collected, these videos are also because of multiple types (such as caricature, news, sport, advertisement, TV Program and home videos) and it is different, the duration minute is differed from 1 minute to 10.
The present invention is evaluated using the objective evaluation standard of mainstream, i.e. precision (Precision), recall rate (Recall) With F values (F-snore).The calculation formula difference of precision, recall rate and F values is as follows:
Wherein Nmatched indicates autoabstract and the matched length of User Summary, i.e., in autoabstract in User Summary Identical key frame number, definition are less than specified threshold when the manhatton distance of the color histogram of two key framesWhen, it is believed that Two key frames are matched, this experiment willValue is set as 0.5;NAS indicates to automatically generate the length of abstract;Nus indicates user Length of summarization.Precision reflects autoabstract and plucks the ability for selecting matching key frame, and recall rate reflects matching key frame and hits The ability of User Summary, F values are the balances to precision and recall rate, are an overall evaluations to video frequency abstract quality.
For each individual video, when obtaining autoabstract result, the present invention first by by the frame detected with The annotation frame of each user is compared to calculate F values, all F values that then annotation by comparing each user is obtained into Row is average.The average F fractional values are used as each final assessment result summarized.The average F fractional values of algorithms of different such as 1 institute of table Show.
1 the method for the present invention of table and other video frequency abstract algorithm objective evaluation Comparison of standards results
As can be seen from the table, the F value highests that method of the invention calculates, mean apparent 0.54, it is clear that reach best Performance.The more key frames of algorithms selection proposed, although precision is relatively low, however, since recall rate is higher, it is proposed by the present invention The F values of method are not much less, it is proposed by the present invention based on the video summarization method of super-pixel segmentation still better than all Comparison algorithm.
By the above-mentioned means, a kind of video abstraction generating method based on super-pixel segmentation of the present invention, point of video boundaries Cut by slightly arrive it is thin in a manner of complete, partitioning boundary tends to Local Minimum moving region, make boundary alignment to be suitable for cutting The part of position extracts the partitioning boundary of video as the key frame extracted, indicates the key activities quilt between action sequence Extraction, represents the effective information of video, and the operand and complexity of video can be greatly lowered, for improving video point The real-time of analysis has a significant impact;To have many characteristics, such as similar grain, color, brightness using the method for super-pixel region merging technique The image block that adjacent pixel is constituted can obtain the redundancy of image by the similarity degree of feature between pixel by group pixels Information largely reduces the complexity of subsequent image processing task;Using spatial organization's relationship between pixel into The Similarity measures of row image generate video frequency abstract to eliminate redundancy key frames, and video frequency abstract effect is preferable.

Claims (8)

1. a kind of video abstraction generating method based on super-pixel segmentation, which is characterized in that be specifically implemented according to the following steps:
Step 1 obtains video image, and using video image as input data, extracts video image frame sequence;
Step 2 carries out initial isometric segmentation to video image frame sequence, is divided into multiple isometric frame sequences;
Meanwhile the light stream of video image frame sequence is calculated, and use the light stream offset estimation conduct between video image frame sequence The motion amplitude of the function of time;
Step 3 is directed to each isometric frame sequence, and the motion amplitude for passing through the function of time finds initial isometric partitioning boundary frame part The frame of Grad minimum is as key frame;
Step 4 carries out Local Clustering to the pixel in every width key frame images, using the similarity of feature between pixel to key Frame image carries out super-pixel segmentation, obtains multiple super-pixel;
Step 5 carries out region merging technique to multiple super-pixel, obtains super-pixel region merging technique image;
Step 6 carries out similarity measurement to adjacent two width super-pixel region merging technique image, judges that adjacent two width super-pixel region is closed And whether image is similar;
One in step 7, the adjacent similar super-pixel region merging technique image of deletion, by remaining super-pixel region merging technique image group Video frequency abstract is generated after conjunction.
2. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 2 The light stream offset estimation using between successive frame is as the motion amplitude detailed process of the function of time:By being regarded each Light stream in all pixels of frequency picture frame in aggregation level and vertical direction calculates the amplitude of video image frame sequence movement, Calculation formula is as follows:
Wherein, OFx(i, j, t) is the x-component of light stream of the pixel (i, j) between frame t and t-1, OFy(i, j, t) be pixel (i, J) y-component of the light stream between frame t and t-1;Light stream tracks all the points over time, and summation is the movement between frame The estimation of amount.
3. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 4 Detailed process be:
Every width key frame formation image is converted to 5 dimensional feature vectors under CIELAB color spaces and XY coordinates by step 4.1, Then module is constructed to 5 dimensional feature vectors:
Assuming that it is the super-pixel of K identical sizes that image, which has N number of pixel, pre-segmentation, the size of each super-pixel is N/K, choosing The super-pixel central point of fixed each pre-segmentation is seed point, and the distance of each two seed point is approximately
The seed point, is moved to the position of Grad minimum, together by step 4.2, the window that 3 × 3 are divided centered on seed point When for each seed distribute an individual label;
Step 4.3, for each pixel, the similarity degree for each seed point that distance is nearest therewith is calculated separately, by most like kind The label of son point is assigned to the pixel, and the pixel with same label forms a super-pixel.
4. a kind of video abstraction generating method based on super-pixel segmentation according to claim 3, which is characterized in that step The formula of 4.3 similarity degrees for calculating each seed point that distance is nearest therewith is as follows:
Wherein, dlabColor distortion between pixel, dxySpace length between pixel, DiFor the similarity of two pixels; S is the spacing of seed point, and m is balance parameters, for weighing color value and proportion of the spatial information in measuring similarity, DiIt takes Value is bigger, illustrates that two pixels are more similar.
5. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 5 Detailed process is:
Step 5.1 presets region merging technique threshold value, it is assumed that super-pixel has K, using non-directed graph G=(V, E) to super-pixel neighbour The relationship of connecing is recorded;
Wherein V={ S1, S2..., SKBe all K vertex set, E is the set on all boundaries, in figure G=(V, E), often One region is represented as a node in figure;
Step 5.2 chooses any two region (Si, Sj) ∈ V, if Si, SjIt is adjacent, then there is one between corresponding node Side is connected, and each side assigns certain weight, calculates weight and corresponds to the cost value for merging two regions;
The weight of cost value minimum merges in step 5.3, selection adjacent area, constantly updates relevant edge weight, and sentence Whether the areal after the merging that breaks, which is equal to, presets region merging technique threshold value, when equal to presetting region merging technique threshold value, Then merge termination, obtains multiple super-pixel block;
The image that step 5.4, multiple super-pixel block are formed is super-pixel region merging technique image.
6. a kind of video abstraction generating method based on super-pixel segmentation according to claim 5, which is characterized in that step 5.2 it is described calculate weights and correspond to merge the calculation formula of cost value in two regions and be:
Wherein, N indicates that region S areas, μ indicate that the spectrum mean value in region, l are the Borders of region i and j, and λ is shape ginseng Number.
7. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 6 It is described to be to adjacent two width super-pixel region merging technique image progress similarity measurement detailed process:
Preset difference value threshold value;
Calculate the difference value between each pixel of adjacent two width super-pixel region merging technique image:
Wherein, i indicates that picture number, i+1 indicate the adjacent image of i images, and a indicates pixel, to any pixel in image The coordinate of point a is Ia(x, y), where super-pixel segmentation merge after region be:Ω a=sp (Ia(x, y)), M indicates the area The pixel number in domain;
The difference value of super-pixel, calculation formula are as follows where calculating all respective pixel points of two width super-pixel region merging technique images:
Wherein, the dividing value of the pixel difference value value during threshold is indicated per adjacent two super-pixel region merging technique image.
8. a kind of video abstraction generating method based on super-pixel segmentation according to claim 7, which is characterized in that step 6 It is described to judge whether similar process is adjacent two width super-pixel region merging technique image:Judge adjacent two width super-pixel region merging technique figure Whether the difference value of picture is less than threshold value:
If difference value is less than threshold value, two width super-pixel region merging technique images are similar;
If difference value is not less than threshold value, two width super-pixel region merging technique images are dissimilar.
CN201810456341.8A 2018-05-14 2018-05-14 A kind of video abstraction generating method based on super-pixel segmentation Active CN108600865B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810456341.8A CN108600865B (en) 2018-05-14 2018-05-14 A kind of video abstraction generating method based on super-pixel segmentation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810456341.8A CN108600865B (en) 2018-05-14 2018-05-14 A kind of video abstraction generating method based on super-pixel segmentation

Publications (2)

Publication Number Publication Date
CN108600865A true CN108600865A (en) 2018-09-28
CN108600865B CN108600865B (en) 2019-07-23

Family

ID=63637370

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810456341.8A Active CN108600865B (en) 2018-05-14 2018-05-14 A kind of video abstraction generating method based on super-pixel segmentation

Country Status (1)

Country Link
CN (1) CN108600865B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109726765A (en) * 2019-01-02 2019-05-07 京东方科技集团股份有限公司 A kind of sample extraction method and device of visual classification problem
CN110347870A (en) * 2019-06-19 2019-10-18 西安理工大学 The video frequency abstract generation method of view-based access control model conspicuousness detection and hierarchical clustering method
CN110399847A (en) * 2019-07-30 2019-11-01 北京字节跳动网络技术有限公司 Extraction method of key frame, device and electronic equipment
CN111050210A (en) * 2018-10-12 2020-04-21 奥多比公司 Video patching via confidence weighted motion estimation
CN111625683A (en) * 2020-05-07 2020-09-04 山东师范大学 Video abstract automatic generation method and system based on graph structure difference analysis
CN111931811A (en) * 2020-06-29 2020-11-13 南京巨鲨显示科技有限公司 Calculation method based on super-pixel image similarity
CN112463385A (en) * 2020-12-02 2021-03-09 深圳市互盟科技股份有限公司 AI data scheduling and distributing method for cloud computing and related products
CN112583900A (en) * 2020-12-02 2021-03-30 深圳市互盟科技股份有限公司 Data processing method for cloud computing and related product
WO2023056835A1 (en) * 2021-10-09 2023-04-13 北京字节跳动网络技术有限公司 Video cover generation method and apparatus, and electronic device and readable medium
CN116805316A (en) * 2023-08-25 2023-09-26 深圳市鹏顺兴包装制品有限公司 Degradable plastic processing quality detection method based on image enhancement

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102289847A (en) * 2011-08-02 2011-12-21 浙江大学 Interaction method for quickly extracting video object
CN102637253A (en) * 2011-12-30 2012-08-15 清华大学 Video foreground object extracting method based on visual saliency and superpixel division
CN103065153A (en) * 2012-12-17 2013-04-24 西南科技大学 Video key frame extraction method based on color quantization and clusters
US8605795B2 (en) * 2008-09-17 2013-12-10 Intel Corporation Video editing methods and systems
CN104185089A (en) * 2013-05-23 2014-12-03 三星电子(中国)研发中心 Video summary generation method, server and client-terminal
CN106851437A (en) * 2017-01-17 2017-06-13 南通同洲电子有限责任公司 A kind of method for extracting video frequency abstract
US20170358090A1 (en) * 2016-06-09 2017-12-14 The Penn State Research Foundation Systems and methods for detection of significant and attractive components in digital images
CN107844779A (en) * 2017-11-21 2018-03-27 重庆邮电大学 A kind of video key frame extracting method

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8605795B2 (en) * 2008-09-17 2013-12-10 Intel Corporation Video editing methods and systems
CN102289847A (en) * 2011-08-02 2011-12-21 浙江大学 Interaction method for quickly extracting video object
CN102637253A (en) * 2011-12-30 2012-08-15 清华大学 Video foreground object extracting method based on visual saliency and superpixel division
CN103065153A (en) * 2012-12-17 2013-04-24 西南科技大学 Video key frame extraction method based on color quantization and clusters
CN104185089A (en) * 2013-05-23 2014-12-03 三星电子(中国)研发中心 Video summary generation method, server and client-terminal
US20170358090A1 (en) * 2016-06-09 2017-12-14 The Penn State Research Foundation Systems and methods for detection of significant and attractive components in digital images
CN106851437A (en) * 2017-01-17 2017-06-13 南通同洲电子有限责任公司 A kind of method for extracting video frequency abstract
CN107844779A (en) * 2017-11-21 2018-03-27 重庆邮电大学 A kind of video key frame extracting method

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
HANA GHARBI, ET AL.: "Key frames extraction using graph modularity clustering for efficient video summarization", 《 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)》 *
冀中,樊帅飞.: "基于超图排序算法的视频摘要", 《电子学报》 *
尚佳敏.: "基于特征聚类的视频摘要生成技术研究", 《中国硕士研究生优秀毕业论文库》 *
张玉培.: "基于层次结构的视频摘要", 《中国硕士研究生优秀毕业论文库》 *
贺宏遵.: "视频摘要技术的研究", 《中国硕士研究生优秀毕业论文库》 *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111050210A (en) * 2018-10-12 2020-04-21 奥多比公司 Video patching via confidence weighted motion estimation
CN111050210B (en) * 2018-10-12 2023-01-17 奥多比公司 Method of performing operations, video processing system, and non-transitory computer readable medium
CN109726765A (en) * 2019-01-02 2019-05-07 京东方科技集团股份有限公司 A kind of sample extraction method and device of visual classification problem
CN110347870A (en) * 2019-06-19 2019-10-18 西安理工大学 The video frequency abstract generation method of view-based access control model conspicuousness detection and hierarchical clustering method
CN110399847B (en) * 2019-07-30 2021-11-09 北京字节跳动网络技术有限公司 Key frame extraction method and device and electronic equipment
CN110399847A (en) * 2019-07-30 2019-11-01 北京字节跳动网络技术有限公司 Extraction method of key frame, device and electronic equipment
CN111625683A (en) * 2020-05-07 2020-09-04 山东师范大学 Video abstract automatic generation method and system based on graph structure difference analysis
CN111625683B (en) * 2020-05-07 2023-05-23 山东师范大学 Automatic video abstract generation method and system based on graph structure difference analysis
WO2022001571A1 (en) * 2020-06-29 2022-01-06 南京巨鲨显示科技有限公司 Computing method based on super-pixel image similarity
CN111931811A (en) * 2020-06-29 2020-11-13 南京巨鲨显示科技有限公司 Calculation method based on super-pixel image similarity
CN111931811B (en) * 2020-06-29 2024-03-29 南京巨鲨显示科技有限公司 Calculation method based on super-pixel image similarity
CN112583900A (en) * 2020-12-02 2021-03-30 深圳市互盟科技股份有限公司 Data processing method for cloud computing and related product
CN112463385A (en) * 2020-12-02 2021-03-09 深圳市互盟科技股份有限公司 AI data scheduling and distributing method for cloud computing and related products
WO2023056835A1 (en) * 2021-10-09 2023-04-13 北京字节跳动网络技术有限公司 Video cover generation method and apparatus, and electronic device and readable medium
CN116805316A (en) * 2023-08-25 2023-09-26 深圳市鹏顺兴包装制品有限公司 Degradable plastic processing quality detection method based on image enhancement
CN116805316B (en) * 2023-08-25 2023-11-28 深圳市鹏顺兴包装制品有限公司 Degradable plastic processing quality detection method based on image enhancement

Also Published As

Publication number Publication date
CN108600865B (en) 2019-07-23

Similar Documents

Publication Publication Date Title
CN108600865B (en) A kind of video abstraction generating method based on super-pixel segmentation
CN104063883B (en) A kind of monitor video abstraction generating method being combined based on object and key frame
US11861848B2 (en) System and method for generating trackable video frames from broadcast video
JP5686800B2 (en) Method and apparatus for processing video
CN106446015A (en) Video content access prediction and recommendation method based on user behavior preference
WO2012071696A1 (en) Method and system for pushing individual advertisement based on user interest learning
CN109064484A (en) Crowd movement's Activity recognition method with momentum Fusion Features is divided based on subgroup
CN107247919A (en) The acquisition methods and system of a kind of video feeling content
CN107358141B (en) Data identification method and device
Omidyeganeh et al. Video keyframe analysis using a segment-based statistical metric in a visually sensitive parametric space
CN104794446B (en) Human motion recognition method and system based on synthesis description
Yuan et al. Key frame extraction based on global motion statistics for team-sport videos
Martin et al. Optimal choice of motion estimation methods for fine-grained action classification with 3d convolutional networks
Zhu et al. Action recognition in broadcast tennis video using optical flow and support vector machine
CN110188625B (en) Video fine structuring method based on multi-feature fusion
Li et al. An efficient spatiotemporal attention model and its application to shot matching
Sandhu et al. Summarizing Videos by Key frame extraction using SSIM and other Visual Features
CN103971100A (en) Video-based camouflage and peeping behavior detection method for automated teller machine
Mademlis et al. Exploiting stereoscopic disparity for augmenting human activity recognition performance
CN108765384B (en) Significance detection method for joint manifold sequencing and improved convex hull
Turchini et al. Understanding sport activities from correspondences of clustered trajectories
Wang et al. Group Activity Recognition based on Temporal Semantic Sub-Graph Network
Dimou et al. A user-centric approach for event-driven summarization of surveillance videos
Park et al. Extraction of visual information in basketball broadcasting video for event segmentation system
CN111191524A (en) Sports people counting method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20211231

Address after: 710000 No. 29, Xiangyang North Road, Hongqing street, Baqiao District, Xi'an City, Shaanxi Province

Patentee after: Shaanxi Zhisheng Desen Elevator Co.,Ltd.

Address before: 710048 No. 5 Jinhua South Road, Shaanxi, Xi'an

Patentee before: XI'AN University OF TECHNOLOGY