CN108600865A - A kind of video abstraction generating method based on super-pixel segmentation - Google Patents
A kind of video abstraction generating method based on super-pixel segmentation Download PDFInfo
- Publication number
- CN108600865A CN108600865A CN201810456341.8A CN201810456341A CN108600865A CN 108600865 A CN108600865 A CN 108600865A CN 201810456341 A CN201810456341 A CN 201810456341A CN 108600865 A CN108600865 A CN 108600865A
- Authority
- CN
- China
- Prior art keywords
- pixel
- super
- image
- video
- merging technique
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8549—Creating video summaries, e.g. movie trailer
Landscapes
- Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Computer Security & Cryptography (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a kind of video abstraction generating methods based on super-pixel segmentation, the segmentation of video boundaries by slightly arrive it is thin in a manner of complete, partitioning boundary tends to Local Minimum moving region, make boundary alignment to the part for the position for being suitable for cutting, the partitioning boundary of video is extracted as the key frame extracted, indicate that the key activities between action sequence are extracted, represent the effective information of video, the operand and complexity of video can be greatly lowered, the real-time for improving video analysis has a significant impact;The image block that will have many characteristics, such as similar grain, color, the adjacent pixel composition of brightness using the method for super-pixel region merging technique, by the similarity degree of feature between pixel by group pixels, the redundancy that image can be obtained reduces the complexity of subsequent image processing task;The Similarity measures of image are carried out using spatial organization's relationship between pixel and generate video frequency abstract to eliminate redundancy key frames, and video frequency abstract effect is preferable.
Description
Technical field
The invention belongs to technical field of image processing, and in particular to a kind of video frequency abstract generation side based on super-pixel segmentation
Method.
Background technology
The research early start of video summarization technique is in the Informedia engineerings of Carnegie Mellon University in 1994.From that
After, researcher is added more and more, and the technology of video frequency abstract related fields is studied and explored, overall
For, for video summarization technique according to the abstract form of output, common video frequency abstract generating algorithm mainly has data clusters at present
Method, curve law of planning and machine learning method.
In the prior art, IEEE International Conference on Acoustics, Speech and
Signal Processing.IEEE,2014:1260-1264. is disclosed original video is decomposed into image sequence after, extraction it is each
The color moment characteristics of frame image, and video is divided into several camera lenses, then use the spectral clustering based on rough set theory
To clustering shots.But such algorithm existing defects:First, most of clustering algorithm is required for presetting the number of cluster
Mesh, optimal clusters number, generally requires constantly to test repeatedly in order to obtain.And optimal clusters number is dynamic change
, and input original video length and type have relationship, cause algorithm that must be executed in the case of manual intervention.The
Two, whether the feature of extraction is representative often to influence Clustering Effect.It is only extracted the color characteristic of image, has ignored figure
The shape and Texture eigenvalue of picture, and single feature cannot comprehensively express the visual information of image.
Proceedings of IEEE International Conference on Multimedia and
Expo.Washington DC,USA:IEEE,2005:In 670-673, after original video is decomposed into image sequence, calculate
Pixel difference between frame and frame, and pixel difference is fitted to curve, in order to simplify curve, broken line is divided into the collection of a broken line
It closes, and the small point of correlation is deleted from set.Curve law of planning intuitively shows video using the curve in coordinate system
Content, can simplify the processing to video, however this method may only reflecting video content variation, it is not possible to complete expression
Semantic information.
Nature and Biologically Inspired Computing.IEEE,2011:In 7-11, SVM pairs is utilized
Playback mark in football video is trained and learns, and identification playback camera lens, the generation of goal event is corresponded to be indicated in playback
Occur the scenes such as stopping, portrait attachment, crowd shots, goal area and audio excitement successively before, and then can after playback indicates
There are the scenes such as full length shot, scoreboard.By the playback camera lens detected and judge whether its front and back scene meets the above-mentioned fact
To carry out football video Context event monitoring.Machine learning method is established model and is required in the case of the overwhelming majority manually
Participate in auxiliary modeling, and this method to modeling when Feature Selection it is more sensitive, if suitable feature is had chosen, using machine
Device learning method can get the high-level semantics information for meeting human cognitive.Conversely, unsuitable Feature Selection will make learning outcome not
People's will to the greatest extent.
Invention content
The purpose of the present invention is to provide a kind of video abstraction generating methods based on super-pixel segmentation, can significantly drop
The operand and complexity of low video, the real-time for improving video analysis have a significant impact.
The technical solution adopted by the present invention is a kind of video abstraction generating method based on super-pixel segmentation, specifically according to
Following steps are implemented:
Step 1 obtains video image, and using video image as input data, extracts video image frame sequence;
Step 2 carries out initial isometric segmentation to video image frame sequence, is divided into multiple isometric frame sequences;
Meanwhile the light stream of video image frame sequence is calculated, and use the light stream offset estimation between video image frame sequence
Motion amplitude as the function of time;
Step 3 is directed to each isometric frame sequence, and the motion amplitude for passing through the function of time finds initial isometric partitioning boundary frame
The frame of partial gradient value minimum is as key frame;
Step 4 carries out Local Clustering to the pixel in every width key frame images, using the similarity pair of feature between pixel
Key frame images carry out super-pixel segmentation, obtain multiple super-pixel;
Step 5 carries out region merging technique to multiple super-pixel, obtains super-pixel region merging technique image;
Step 6 carries out similarity measurement to adjacent two width super-pixel region merging technique image, judges adjacent Liang Fu super-pixel area
Whether domain merges image similar;
One in step 7, the adjacent similar super-pixel region merging technique image of deletion, by remaining super-pixel region merging technique figure
As generating video frequency abstract after combination.
The present invention is further characterized in that:
Motion amplitude detailed process of the step 2 using the light stream offset estimation between successive frame as the function of time be:It is logical
The light stream in all pixels of each video image frame in aggregation level and vertical direction is crossed to calculate video image frame sequence
The amplitude of movement, calculation formula are as follows:
Wherein, OFx (i, j, t) is the x-component of light stream of the pixel (i, j) between frame t and t-1, and OFy (i, j, t) is picture
The y-component of the light stream of plain (i, j) between frame t and t-1;Light stream tracks all the points over time, and summation is between frame
Amount of exercise estimation.
The detailed process of step 4 is:
Step 4.1, by every width key frame formation image be converted to 5 dimensional features under CIELAB color spaces and XY coordinates to
Then amount constructs module to 5 dimensional feature vectors:
Assuming that it is the super-pixel of K identical sizes that image, which has N number of pixel, pre-segmentation, the size of each super-pixel is N/
K, the super-pixel central point for selecting each pre-segmentation is seed point, and the distance of each two seed point is approximately
The seed point, is moved to the position of Grad minimum by step 4.2, the window that 3 × 3 are divided centered on seed point
It sets, while an individual label is distributed for each seed;
Step 4.3, for each pixel, calculate separately the similarity degree for each seed point that distance is nearest therewith, will most phase
It is assigned to the pixel like the label of seed point, the pixel with same label forms a super-pixel.
The formula that step 4.3 calculates the similarity degree for each seed point that distance is nearest therewith is as follows:
Wherein, dlabColor distortion between pixel, dxySpace length between pixel, DiFor the phase of two pixels
Like degree;S is the spacing of seed point, and m is balance parameters, for weighing the ratio of color value and spatial information in measuring similarity
Weight, DiValue is bigger, illustrates that two pixels are more similar.
Step 5 detailed process is:
Step 5.1 presets region merging technique threshold value, it is assumed that super-pixel has K, using non-directed graph G=(V, E) to super picture
Plain syntople is recorded;
Wherein V={ S1, S2..., SKBe all K vertex set, E is the set on all boundaries, figure G=(V, E)
In, each region is represented as a node in figure;
Step 5.2 chooses any two region (Si, Sj) ∈ V, if Si, SjIt is adjacent, then between corresponding node just
There are a lines to be connected, and each side assigns certain weight, calculates weight and corresponds to the cost value for merging two regions;
The weight of cost value minimum merges in step 5.3, selection adjacent area, constantly updates relevant edge weight,
And judge whether the areal after the merging is equal to and preset region merging technique threshold value, when equal to presetting region merging technique threshold
Value, then merge termination, obtain multiple super-pixel block;
The image that step 5.4, multiple super-pixel block are formed is super-pixel region merging technique image.
Step 5.2 calculates weight:
Wherein, N indicates that region S areas, μ indicate the spectrum mean value in region, and l is the Border of region i and j, and λ is shape
Shape parameter.
Step 6 carries out similarity measurement detailed process to adjacent two width super-pixel region merging technique image:
Preset difference value threshold value;
Calculate the difference value between each pixel of adjacent two width super-pixel region merging technique image:
Wherein, i indicates that picture number, i+1 indicate the adjacent image of i images, and a indicates pixel, to any in image
The coordinate of pixel a is Ia(x, y), where super-pixel segmentation merge after region be:Ω a=sp (Ia(x, y)), M is indicated
The pixel number in the region;
The difference value of all respective pixel points place super-pixel of two width super-pixel region merging technique images is calculated, then adjacent two width
Super-pixel region merging technique figure difference value calculation formula is as follows:
Wherein, the boundary of the pixel difference value value during threshold is indicated per adjacent two super-pixel region merging technique image
Value.
Step 6 judges whether similar process is adjacent two width super-pixel region merging technique image:Judge adjacent two width super-pixel
Whether the difference value of region merging technique image is less than threshold value:
If difference value is less than threshold value, two width super-pixel region merging technique images are similar;
If difference value is not less than threshold value, two width super-pixel region merging technique images are dissimilar.
A kind of video abstraction generating method advantageous effect based on super-pixel segmentation of the present invention is:
(1) present invention in video boundaries segmentation by slightly arrive it is thin in a manner of complete, partitioning boundary tend to Local Minimum transport
Dynamic region makes boundary alignment extract the partitioning boundary of video as the key extracted to the part for the position for being suitable for cutting
Frame indicates that the key activities between action sequence are extracted, represents the effective information of video, and video can be greatly lowered
Operand and complexity, for improve video analysis real-time have a significant impact;
(2) will have many characteristics, such as similar grain, color, the phase of brightness using the method for super-pixel region merging technique in the present invention
Adjacent pixel constitutes block of pixels can obtain the redundancy of image by the similarity degree of feature between pixel by group pixels,
Largely reduce the complexity of subsequent image processing task;Image is carried out using spatial organization's relationship between pixel
Similarity measures generate video frequency abstract to eliminate redundancy key frames, and video frequency abstract effect is preferable.
Description of the drawings
Fig. 1 is a kind of video abstraction generating method flow chart based on super-pixel segmentation of the present invention;
Fig. 2 is key-frame extraction display diagram in the present invention;
Fig. 3 is super-pixel region merging technique process schematic of the present invention.
Specific implementation mode
The following describes the present invention in detail with reference to the accompanying drawings and specific embodiments.
A kind of video abstraction generating method based on super-pixel segmentation of the present invention, as shown in Figure 1, specifically according to the following steps
Implement:
Step 1 obtains video image, and using video image as input data, extracts video image frame sequence;
Step 2 carries out initial isometric segmentation to video image frame sequence, is divided into multiple isometric frame sequences;
Meanwhile the light stream of video image frame sequence is calculated, and use the light stream offset estimation between video image frame sequence
Motion amplitude as the function of time;
Motion amplitude detailed process using the light stream offset estimation between successive frame as the function of time is:By every
What the light stream in all pixels of a video image frame in aggregation level and vertical direction moved to calculate video image frame sequence
Amplitude, calculation formula are as follows:
Wherein, OFx (i, j, t) is the x-component of light stream of the pixel (i, j) between frame t and t-1, and OFy (i, j, t) is picture
The y-component of the light stream of plain (i, j) between frame t and t-1;Light stream tracks all the points over time, and summation is between frame
Amount of exercise estimation.
Step 3 is directed to each isometric frame sequence, and the motion amplitude for passing through the function of time finds initial isometric partitioning boundary frame
The frame of partial gradient value minimum is as key frame, as shown in Figure 2;
Step 4 carries out Local Clustering to the pixel in every width key frame images, using the similarity pair of feature between pixel
Key frame images carry out super-pixel segmentation, obtain multiple super-pixel;
Detailed process is:
Step 4.1, by every width key frame formation image be converted to 5 dimensional features under CIELAB color spaces and XY coordinates to
Then amount constructs module to 5 dimensional feature vectors:
Assuming that it is the super-pixel of K identical sizes that image, which has N number of pixel, pre-segmentation, the size of each super-pixel is N/
K, the super-pixel central point for selecting each pre-segmentation is seed point, and the distance of each two seed point is approximately
The seed point, is moved to the position of Grad minimum by step 4.2, the window that 3 × 3 are divided centered on seed point
It sets, while an individual label is distributed for each seed;
Step 4.3, as shown in figure 3, for each pixel, calculate separately the similar journey for each seed point that distance is nearest therewith
Degree, the pixel is assigned to by the label of most like seed point, and the pixel with same label forms a super-pixel;
The formula for calculating the similarity degree for each seed point that distance is nearest therewith is as follows:
Wherein, dlabColor distortion between pixel, dxySpace length between pixel, DiFor the phase of two pixels
Like degree;S is the spacing of seed point, and m is balance parameters, for weighing the ratio of color value and spatial information in measuring similarity
Weight, DiValue is bigger, illustrates that two pixels are more similar.
Step 5 carries out region merging technique to multiple super-pixel, obtains super-pixel region merging technique image;
Detailed process is:
Step 5.1 presets region merging technique threshold value, it is assumed that super-pixel has K, using non-directed graph G=(V, E) to super picture
Plain syntople is recorded;
Wherein V={ S1, S2..., SKBe all K vertex set, E is the set on all boundaries, figure G=(V, E)
In, each region is represented as a node in figure;
Step 5.2 chooses any two region (Si, Sj) ∈ V, if Si, SjIt is adjacent, then between corresponding node just
There are a lines to be connected, and each side assigns certain weight, calculates weight and corresponds to the cost value for merging two regions;
It calculates weight and corresponds to and merge the calculation formula of cost value in two regions and be:
Wherein, N indicates that region S areas, μ indicate the spectrum mean value in region, and l is the Border of region i and j, and λ is shape
Shape parameter;
The weight of cost value minimum merges in step 5.3, selection adjacent area, constantly updates relevant edge weight,
And judge whether the areal after the merging is equal to and preset region merging technique threshold value, when equal to presetting region merging technique threshold
Value, then merge termination, obtain multiple super-pixel block;
The image that step 5.4, multiple super-pixel block are formed is super-pixel region merging technique image.
Step 6 carries out similarity measurement to adjacent two width super-pixel region merging technique image, and detailed process is:
Preset difference value threshold value;
Calculate the difference value between each pixel of adjacent two width super-pixel region merging technique image:
Wherein, i indicates that picture number, i+1 indicate the adjacent image of i images, and a indicates pixel, to any in image
The coordinate of pixel a is Ia(x, y), where super-pixel segmentation merge after region be:Ω a=sp (Ia(x, y)), M is indicated
The pixel number in the region;
The difference value of all respective pixel points place super-pixel of two width super-pixel region merging technique images is calculated, then adjacent two width
Super-pixel region merging technique figure difference value calculation formula is as follows:
Wherein, the boundary of the pixel difference value value during threshold is indicated per adjacent two super-pixel region merging technique image
Value.
Judge whether adjacent two width super-pixel region merging technique image is similar, and process is again:Judge adjacent Liang Fu super-pixel area
Whether the difference value that domain merges image is less than threshold value:
If difference value is less than threshold value, two width super-pixel region merging technique images are similar;
If difference value is not less than threshold value, two width super-pixel region merging technique images are dissimilar.
One in step 7, the adjacent similar super-pixel region merging technique image of deletion, by remaining super-pixel region merging technique figure
As generating video frequency abstract after combination.
Embodiment
Feasibility verification is carried out to the scheme of invention with reference to specific calculation formula, it is described below:
The validity of the put forward algorithm of YouTube database authentications is chosen in experiment.YouTube databases include from video website
50 videos that (such as YouTube) is collected, these videos are also because of multiple types (such as caricature, news, sport, advertisement, TV
Program and home videos) and it is different, the duration minute is differed from 1 minute to 10.
The present invention is evaluated using the objective evaluation standard of mainstream, i.e. precision (Precision), recall rate (Recall)
With F values (F-snore).The calculation formula difference of precision, recall rate and F values is as follows:
Wherein Nmatched indicates autoabstract and the matched length of User Summary, i.e., in autoabstract in User Summary
Identical key frame number, definition are less than specified threshold when the manhatton distance of the color histogram of two key framesWhen, it is believed that
Two key frames are matched, this experiment willValue is set as 0.5;NAS indicates to automatically generate the length of abstract;Nus indicates user
Length of summarization.Precision reflects autoabstract and plucks the ability for selecting matching key frame, and recall rate reflects matching key frame and hits
The ability of User Summary, F values are the balances to precision and recall rate, are an overall evaluations to video frequency abstract quality.
For each individual video, when obtaining autoabstract result, the present invention first by by the frame detected with
The annotation frame of each user is compared to calculate F values, all F values that then annotation by comparing each user is obtained into
Row is average.The average F fractional values are used as each final assessment result summarized.The average F fractional values of algorithms of different such as 1 institute of table
Show.
1 the method for the present invention of table and other video frequency abstract algorithm objective evaluation Comparison of standards results
As can be seen from the table, the F value highests that method of the invention calculates, mean apparent 0.54, it is clear that reach best
Performance.The more key frames of algorithms selection proposed, although precision is relatively low, however, since recall rate is higher, it is proposed by the present invention
The F values of method are not much less, it is proposed by the present invention based on the video summarization method of super-pixel segmentation still better than all
Comparison algorithm.
By the above-mentioned means, a kind of video abstraction generating method based on super-pixel segmentation of the present invention, point of video boundaries
Cut by slightly arrive it is thin in a manner of complete, partitioning boundary tends to Local Minimum moving region, make boundary alignment to be suitable for cutting
The part of position extracts the partitioning boundary of video as the key frame extracted, indicates the key activities quilt between action sequence
Extraction, represents the effective information of video, and the operand and complexity of video can be greatly lowered, for improving video point
The real-time of analysis has a significant impact;To have many characteristics, such as similar grain, color, brightness using the method for super-pixel region merging technique
The image block that adjacent pixel is constituted can obtain the redundancy of image by the similarity degree of feature between pixel by group pixels
Information largely reduces the complexity of subsequent image processing task;Using spatial organization's relationship between pixel into
The Similarity measures of row image generate video frequency abstract to eliminate redundancy key frames, and video frequency abstract effect is preferable.
Claims (8)
1. a kind of video abstraction generating method based on super-pixel segmentation, which is characterized in that be specifically implemented according to the following steps:
Step 1 obtains video image, and using video image as input data, extracts video image frame sequence;
Step 2 carries out initial isometric segmentation to video image frame sequence, is divided into multiple isometric frame sequences;
Meanwhile the light stream of video image frame sequence is calculated, and use the light stream offset estimation conduct between video image frame sequence
The motion amplitude of the function of time;
Step 3 is directed to each isometric frame sequence, and the motion amplitude for passing through the function of time finds initial isometric partitioning boundary frame part
The frame of Grad minimum is as key frame;
Step 4 carries out Local Clustering to the pixel in every width key frame images, using the similarity of feature between pixel to key
Frame image carries out super-pixel segmentation, obtains multiple super-pixel;
Step 5 carries out region merging technique to multiple super-pixel, obtains super-pixel region merging technique image;
Step 6 carries out similarity measurement to adjacent two width super-pixel region merging technique image, judges that adjacent two width super-pixel region is closed
And whether image is similar;
One in step 7, the adjacent similar super-pixel region merging technique image of deletion, by remaining super-pixel region merging technique image group
Video frequency abstract is generated after conjunction.
2. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 2
The light stream offset estimation using between successive frame is as the motion amplitude detailed process of the function of time:By being regarded each
Light stream in all pixels of frequency picture frame in aggregation level and vertical direction calculates the amplitude of video image frame sequence movement,
Calculation formula is as follows:
Wherein, OFx(i, j, t) is the x-component of light stream of the pixel (i, j) between frame t and t-1, OFy(i, j, t) be pixel (i,
J) y-component of the light stream between frame t and t-1;Light stream tracks all the points over time, and summation is the movement between frame
The estimation of amount.
3. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 4
Detailed process be:
Every width key frame formation image is converted to 5 dimensional feature vectors under CIELAB color spaces and XY coordinates by step 4.1,
Then module is constructed to 5 dimensional feature vectors:
Assuming that it is the super-pixel of K identical sizes that image, which has N number of pixel, pre-segmentation, the size of each super-pixel is N/K, choosing
The super-pixel central point of fixed each pre-segmentation is seed point, and the distance of each two seed point is approximately
The seed point, is moved to the position of Grad minimum, together by step 4.2, the window that 3 × 3 are divided centered on seed point
When for each seed distribute an individual label;
Step 4.3, for each pixel, the similarity degree for each seed point that distance is nearest therewith is calculated separately, by most like kind
The label of son point is assigned to the pixel, and the pixel with same label forms a super-pixel.
4. a kind of video abstraction generating method based on super-pixel segmentation according to claim 3, which is characterized in that step
The formula of 4.3 similarity degrees for calculating each seed point that distance is nearest therewith is as follows:
Wherein, dlabColor distortion between pixel, dxySpace length between pixel, DiFor the similarity of two pixels;
S is the spacing of seed point, and m is balance parameters, for weighing color value and proportion of the spatial information in measuring similarity, DiIt takes
Value is bigger, illustrates that two pixels are more similar.
5. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 5
Detailed process is:
Step 5.1 presets region merging technique threshold value, it is assumed that super-pixel has K, using non-directed graph G=(V, E) to super-pixel neighbour
The relationship of connecing is recorded;
Wherein V={ S1, S2..., SKBe all K vertex set, E is the set on all boundaries, in figure G=(V, E), often
One region is represented as a node in figure;
Step 5.2 chooses any two region (Si, Sj) ∈ V, if Si, SjIt is adjacent, then there is one between corresponding node
Side is connected, and each side assigns certain weight, calculates weight and corresponds to the cost value for merging two regions;
The weight of cost value minimum merges in step 5.3, selection adjacent area, constantly updates relevant edge weight, and sentence
Whether the areal after the merging that breaks, which is equal to, presets region merging technique threshold value, when equal to presetting region merging technique threshold value,
Then merge termination, obtains multiple super-pixel block;
The image that step 5.4, multiple super-pixel block are formed is super-pixel region merging technique image.
6. a kind of video abstraction generating method based on super-pixel segmentation according to claim 5, which is characterized in that step
5.2 it is described calculate weights and correspond to merge the calculation formula of cost value in two regions and be:
Wherein, N indicates that region S areas, μ indicate that the spectrum mean value in region, l are the Borders of region i and j, and λ is shape ginseng
Number.
7. a kind of video abstraction generating method based on super-pixel segmentation according to claim 1, which is characterized in that step 6
It is described to be to adjacent two width super-pixel region merging technique image progress similarity measurement detailed process:
Preset difference value threshold value;
Calculate the difference value between each pixel of adjacent two width super-pixel region merging technique image:
Wherein, i indicates that picture number, i+1 indicate the adjacent image of i images, and a indicates pixel, to any pixel in image
The coordinate of point a is Ia(x, y), where super-pixel segmentation merge after region be:Ω a=sp (Ia(x, y)), M indicates the area
The pixel number in domain;
The difference value of super-pixel, calculation formula are as follows where calculating all respective pixel points of two width super-pixel region merging technique images:
Wherein, the dividing value of the pixel difference value value during threshold is indicated per adjacent two super-pixel region merging technique image.
8. a kind of video abstraction generating method based on super-pixel segmentation according to claim 7, which is characterized in that step 6
It is described to judge whether similar process is adjacent two width super-pixel region merging technique image:Judge adjacent two width super-pixel region merging technique figure
Whether the difference value of picture is less than threshold value:
If difference value is less than threshold value, two width super-pixel region merging technique images are similar;
If difference value is not less than threshold value, two width super-pixel region merging technique images are dissimilar.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810456341.8A CN108600865B (en) | 2018-05-14 | 2018-05-14 | A kind of video abstraction generating method based on super-pixel segmentation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810456341.8A CN108600865B (en) | 2018-05-14 | 2018-05-14 | A kind of video abstraction generating method based on super-pixel segmentation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108600865A true CN108600865A (en) | 2018-09-28 |
CN108600865B CN108600865B (en) | 2019-07-23 |
Family
ID=63637370
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810456341.8A Active CN108600865B (en) | 2018-05-14 | 2018-05-14 | A kind of video abstraction generating method based on super-pixel segmentation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108600865B (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109726765A (en) * | 2019-01-02 | 2019-05-07 | 京东方科技集团股份有限公司 | A kind of sample extraction method and device of visual classification problem |
CN110347870A (en) * | 2019-06-19 | 2019-10-18 | 西安理工大学 | The video frequency abstract generation method of view-based access control model conspicuousness detection and hierarchical clustering method |
CN110399847A (en) * | 2019-07-30 | 2019-11-01 | 北京字节跳动网络技术有限公司 | Extraction method of key frame, device and electronic equipment |
CN111050210A (en) * | 2018-10-12 | 2020-04-21 | 奥多比公司 | Video patching via confidence weighted motion estimation |
CN111625683A (en) * | 2020-05-07 | 2020-09-04 | 山东师范大学 | Video abstract automatic generation method and system based on graph structure difference analysis |
CN111931811A (en) * | 2020-06-29 | 2020-11-13 | 南京巨鲨显示科技有限公司 | Calculation method based on super-pixel image similarity |
CN112463385A (en) * | 2020-12-02 | 2021-03-09 | 深圳市互盟科技股份有限公司 | AI data scheduling and distributing method for cloud computing and related products |
CN112583900A (en) * | 2020-12-02 | 2021-03-30 | 深圳市互盟科技股份有限公司 | Data processing method for cloud computing and related product |
WO2023056835A1 (en) * | 2021-10-09 | 2023-04-13 | 北京字节跳动网络技术有限公司 | Video cover generation method and apparatus, and electronic device and readable medium |
CN116805316A (en) * | 2023-08-25 | 2023-09-26 | 深圳市鹏顺兴包装制品有限公司 | Degradable plastic processing quality detection method based on image enhancement |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102289847A (en) * | 2011-08-02 | 2011-12-21 | 浙江大学 | Interaction method for quickly extracting video object |
CN102637253A (en) * | 2011-12-30 | 2012-08-15 | 清华大学 | Video foreground object extracting method based on visual saliency and superpixel division |
CN103065153A (en) * | 2012-12-17 | 2013-04-24 | 西南科技大学 | Video key frame extraction method based on color quantization and clusters |
US8605795B2 (en) * | 2008-09-17 | 2013-12-10 | Intel Corporation | Video editing methods and systems |
CN104185089A (en) * | 2013-05-23 | 2014-12-03 | 三星电子(中国)研发中心 | Video summary generation method, server and client-terminal |
CN106851437A (en) * | 2017-01-17 | 2017-06-13 | 南通同洲电子有限责任公司 | A kind of method for extracting video frequency abstract |
US20170358090A1 (en) * | 2016-06-09 | 2017-12-14 | The Penn State Research Foundation | Systems and methods for detection of significant and attractive components in digital images |
CN107844779A (en) * | 2017-11-21 | 2018-03-27 | 重庆邮电大学 | A kind of video key frame extracting method |
-
2018
- 2018-05-14 CN CN201810456341.8A patent/CN108600865B/en active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8605795B2 (en) * | 2008-09-17 | 2013-12-10 | Intel Corporation | Video editing methods and systems |
CN102289847A (en) * | 2011-08-02 | 2011-12-21 | 浙江大学 | Interaction method for quickly extracting video object |
CN102637253A (en) * | 2011-12-30 | 2012-08-15 | 清华大学 | Video foreground object extracting method based on visual saliency and superpixel division |
CN103065153A (en) * | 2012-12-17 | 2013-04-24 | 西南科技大学 | Video key frame extraction method based on color quantization and clusters |
CN104185089A (en) * | 2013-05-23 | 2014-12-03 | 三星电子(中国)研发中心 | Video summary generation method, server and client-terminal |
US20170358090A1 (en) * | 2016-06-09 | 2017-12-14 | The Penn State Research Foundation | Systems and methods for detection of significant and attractive components in digital images |
CN106851437A (en) * | 2017-01-17 | 2017-06-13 | 南通同洲电子有限责任公司 | A kind of method for extracting video frequency abstract |
CN107844779A (en) * | 2017-11-21 | 2018-03-27 | 重庆邮电大学 | A kind of video key frame extracting method |
Non-Patent Citations (5)
Title |
---|
HANA GHARBI, ET AL.: "Key frames extraction using graph modularity clustering for efficient video summarization", 《 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP)》 * |
冀中,樊帅飞.: "基于超图排序算法的视频摘要", 《电子学报》 * |
尚佳敏.: "基于特征聚类的视频摘要生成技术研究", 《中国硕士研究生优秀毕业论文库》 * |
张玉培.: "基于层次结构的视频摘要", 《中国硕士研究生优秀毕业论文库》 * |
贺宏遵.: "视频摘要技术的研究", 《中国硕士研究生优秀毕业论文库》 * |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111050210A (en) * | 2018-10-12 | 2020-04-21 | 奥多比公司 | Video patching via confidence weighted motion estimation |
CN111050210B (en) * | 2018-10-12 | 2023-01-17 | 奥多比公司 | Method of performing operations, video processing system, and non-transitory computer readable medium |
CN109726765A (en) * | 2019-01-02 | 2019-05-07 | 京东方科技集团股份有限公司 | A kind of sample extraction method and device of visual classification problem |
CN110347870A (en) * | 2019-06-19 | 2019-10-18 | 西安理工大学 | The video frequency abstract generation method of view-based access control model conspicuousness detection and hierarchical clustering method |
CN110399847B (en) * | 2019-07-30 | 2021-11-09 | 北京字节跳动网络技术有限公司 | Key frame extraction method and device and electronic equipment |
CN110399847A (en) * | 2019-07-30 | 2019-11-01 | 北京字节跳动网络技术有限公司 | Extraction method of key frame, device and electronic equipment |
CN111625683A (en) * | 2020-05-07 | 2020-09-04 | 山东师范大学 | Video abstract automatic generation method and system based on graph structure difference analysis |
CN111625683B (en) * | 2020-05-07 | 2023-05-23 | 山东师范大学 | Automatic video abstract generation method and system based on graph structure difference analysis |
WO2022001571A1 (en) * | 2020-06-29 | 2022-01-06 | 南京巨鲨显示科技有限公司 | Computing method based on super-pixel image similarity |
CN111931811A (en) * | 2020-06-29 | 2020-11-13 | 南京巨鲨显示科技有限公司 | Calculation method based on super-pixel image similarity |
CN111931811B (en) * | 2020-06-29 | 2024-03-29 | 南京巨鲨显示科技有限公司 | Calculation method based on super-pixel image similarity |
CN112583900A (en) * | 2020-12-02 | 2021-03-30 | 深圳市互盟科技股份有限公司 | Data processing method for cloud computing and related product |
CN112463385A (en) * | 2020-12-02 | 2021-03-09 | 深圳市互盟科技股份有限公司 | AI data scheduling and distributing method for cloud computing and related products |
WO2023056835A1 (en) * | 2021-10-09 | 2023-04-13 | 北京字节跳动网络技术有限公司 | Video cover generation method and apparatus, and electronic device and readable medium |
CN116805316A (en) * | 2023-08-25 | 2023-09-26 | 深圳市鹏顺兴包装制品有限公司 | Degradable plastic processing quality detection method based on image enhancement |
CN116805316B (en) * | 2023-08-25 | 2023-11-28 | 深圳市鹏顺兴包装制品有限公司 | Degradable plastic processing quality detection method based on image enhancement |
Also Published As
Publication number | Publication date |
---|---|
CN108600865B (en) | 2019-07-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108600865B (en) | A kind of video abstraction generating method based on super-pixel segmentation | |
CN104063883B (en) | A kind of monitor video abstraction generating method being combined based on object and key frame | |
US11861848B2 (en) | System and method for generating trackable video frames from broadcast video | |
JP5686800B2 (en) | Method and apparatus for processing video | |
CN106446015A (en) | Video content access prediction and recommendation method based on user behavior preference | |
WO2012071696A1 (en) | Method and system for pushing individual advertisement based on user interest learning | |
CN109064484A (en) | Crowd movement's Activity recognition method with momentum Fusion Features is divided based on subgroup | |
CN107247919A (en) | The acquisition methods and system of a kind of video feeling content | |
CN107358141B (en) | Data identification method and device | |
Omidyeganeh et al. | Video keyframe analysis using a segment-based statistical metric in a visually sensitive parametric space | |
CN104794446B (en) | Human motion recognition method and system based on synthesis description | |
Yuan et al. | Key frame extraction based on global motion statistics for team-sport videos | |
Martin et al. | Optimal choice of motion estimation methods for fine-grained action classification with 3d convolutional networks | |
Zhu et al. | Action recognition in broadcast tennis video using optical flow and support vector machine | |
CN110188625B (en) | Video fine structuring method based on multi-feature fusion | |
Li et al. | An efficient spatiotemporal attention model and its application to shot matching | |
Sandhu et al. | Summarizing Videos by Key frame extraction using SSIM and other Visual Features | |
CN103971100A (en) | Video-based camouflage and peeping behavior detection method for automated teller machine | |
Mademlis et al. | Exploiting stereoscopic disparity for augmenting human activity recognition performance | |
CN108765384B (en) | Significance detection method for joint manifold sequencing and improved convex hull | |
Turchini et al. | Understanding sport activities from correspondences of clustered trajectories | |
Wang et al. | Group Activity Recognition based on Temporal Semantic Sub-Graph Network | |
Dimou et al. | A user-centric approach for event-driven summarization of surveillance videos | |
Park et al. | Extraction of visual information in basketball broadcasting video for event segmentation system | |
CN111191524A (en) | Sports people counting method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20211231 Address after: 710000 No. 29, Xiangyang North Road, Hongqing street, Baqiao District, Xi'an City, Shaanxi Province Patentee after: Shaanxi Zhisheng Desen Elevator Co.,Ltd. Address before: 710048 No. 5 Jinhua South Road, Shaanxi, Xi'an Patentee before: XI'AN University OF TECHNOLOGY |