CN105869178B - A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature - Google Patents

A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature Download PDF

Info

Publication number
CN105869178B
CN105869178B CN201610261547.6A CN201610261547A CN105869178B CN 105869178 B CN105869178 B CN 105869178B CN 201610261547 A CN201610261547 A CN 201610261547A CN 105869178 B CN105869178 B CN 105869178B
Authority
CN
China
Prior art keywords
pixel
super
energy function
image
spatial domain
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610261547.6A
Other languages
Chinese (zh)
Other versions
CN105869178A (en
Inventor
何自芬
张印辉
伍星
张云生
王森
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Kunming University of Science and Technology
Original Assignee
Kunming University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Kunming University of Science and Technology filed Critical Kunming University of Science and Technology
Priority to CN201610261547.6A priority Critical patent/CN105869178B/en
Publication of CN105869178A publication Critical patent/CN105869178A/en
Application granted granted Critical
Publication of CN105869178B publication Critical patent/CN105869178B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20064Wavelet transform [DWT]

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature that the present invention relates to a kind of, belongs to image procossing and technical field of computer vision.The present invention extracts wavelet field high yardstick target signature first, and calculate movement and spatial domain skirt response, by wavelet field, spatially and temporally edge feature merges to obtain Multiscale combination probability skirt response, then combination edge internal maps weight total variation energy functional model is established, the convex optimization of weight total variation functional model is calculated using alternating direction, and the unitary energy function item of markov random file and binary spatial domain energy function when defining super-pixel scale sky accordingly, obtain the posterior probability segmentation result of image sequence finally by figure cutting single-step iteration reasoning.The present invention can effectively changed complex target be split appearance color is inconsistent and shape from dynamic scene at any time.

Description

A kind of complex target dynamic scene based on the convex optimization of Multiscale combination feature is unsupervised Dividing method
Technical field
The present invention relates to a kind of complex target dynamic scene non-formaldehyde finishing sides based on the convex optimization of Multiscale combination feature Method belongs to image procossing and technical field of computer vision.
Background technology
Segmentation of Image Sequences is the technology of the band of position where semantic foreground target in each frame image of extraction, extensive use In the computer visions such as target identification, tracking, image understanding and vision guided navigation and robot vision field.In reality, by In illumination and the variation of environment and the movement of camera itself, cause that actual background is dynamic change and this variation is Random;In addition, the feature of foreground target also has complexity under normal conditions, even if the appearance color of target is in synchronization May also be inconsistent, the topological structure of target can also change at any time.Therefore, the automatic of complex target is divided under dynamic scene Cutting technology becomes difficult point and hot issue that top scientific research institution of various countries competitively chases.
Segmentation of Image Sequences method main can be divided into two classes, that is, have measure of supervision and unsupervised approaches.Early stage Image partition method needs the foreground target for being artificially labeled in certain frames.These methods very labor intensive, because this needs craft Divide training set to handle image sequence.There is supervision Segmentation of Image Sequences method to usually require the language in user's selection key frame Adopted region.It is main to show although these methods obtain segmentation result to a certain extent, but have some limitations In the selection of the key frame of segmentation result heavy dependence mark.Unsupervised image division method is without using the image that manually marks It is trained, but by the relevant range of Computer Automatic Extraction semanteme foreground target.
Recent image sequence non-formaldehyde finishing method has used dynamic background texture model, by comparing its current value Determine whether a pixel belongs to background with past value, and the value for propagating a background pixel enters background extracting mould Type.Gauss hybrid models are handled to estimate background distributions using Dirichlet, and are used them as to scene changes A continuous newer model learning process input.This method is to be based on a typical hypotheses, i.e. background becomes That changes is very steady and slow, and which limits image segmentation of this kind of algorithm under dynamic scene.
Fast target partition algorithm attempts the dynamic profile model of structure target, and assumes that background changes with time and be Smooth.One advantage of this method is that it is possible to handle the Deja Vu in marking refinement stage on image block, so And the only point on moving boundaries in target is initialized, a large amount of wrong report seed point may be will produce, especially in dynamic field Jing Zhong.Subsequent choosing is done from the variation level of granularity with the method permission for building space-time dividing to combine layering clue by setting It selects.Although this method is effective to processing layering shape clue, still there are one very strong limitations, cause the mistake of scene Segmentation, can not solve the problems, such as the foreground segmentation task of itself.
Another kind of unsupervised Segmentation of Image Sequences method is ranked up using the structuring study in target assumption region.Make With shape and action consistency class target it is interregional assume packet sequencings, introduced in the figure of weighted area in frame and interframe Constraint look for maximum weighted super-pixel.Or cut using figure and realize segmentation with color clue, and region is by being based on one The classification of the form clue of a multivariate model is sorted.It sorts although target assumption can be conducive to image block level, in reality Such methods are all since there are the segmentations of the highly redundant in different images region, even if a moderate image sequence is also subject to choosing It selects the higher storage of target assumption and calculates cost.
The present invention is subsidized by project of national nature science fund project (61461022,61302173) and is studied, and essentially consists in exploration It is whole to estimate robust fusion algorithm with local feature multi-scale coupling mechanism and multiple dimensioned perceptual error, solve segmentation posteriority with Really be distributed inconsistent problem, for vision computing platform under dynamic scene the accurate segmentation of complicated foreground target provide it is theoretical with Method foundation.
Invention content
Based on problem above, the present invention provides a kind of dynamic scene non-formaldehyde finishings based on the convex optimization of Analysis On Multi-scale Features Method, solves that appearance color is inconsistent and the topological complex object changed at any time of shape, when target be in due to External environment change and camera motion caused by dynamic background when, other non-formaldehyde finishing technologies be also all difficult to foreground target into The problem of row is effectively divided.
The technical scheme is that:A kind of complex target dynamic scene based on the convex optimization of Multiscale combination feature is without prison Dividing method is superintended and directed, wavelet field high yardstick target signature is extracted first, and calculate movement and spatial domain skirt response, by wavelet field, sky Domain and time domain edge Fusion Features obtain Multiscale combination probability skirt response, and it is total then to establish combination edge internal maps weight Variation energy functional model calculates the convex optimization of weight total variation energy functional model using alternating direction, and defines super-pixel ruler The unitary energy function item and binary spatial domain energy function of markov random file when degree is empty, finally by figure cutting single-step iteration Reasoning obtains the posterior probability segmentation result of image sequence;
The specific step of the complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature It is rapid as follows:
Step1, wavelet field high yardstick objective contour feature extraction;In image sequenceThe image I at each moment ttIn, Maximal margin response of the foreground wavelet field high yardstick objective contour at each location of pixels (a, b) is verified first, and T is video Totalframes;
In the step Step1, the maximal margin response definition at each pixel of wavelet field high yardstick objective contour is:
Bs t(a, b)=max (Hs t(a,b),Vs t(a,b),Ds t(a,b))
Wherein, Hs t、Vs tAnd Ds tIndicate image after wavelet transformation in the horizontal, vertical of the positions scale s (a, b) respectively With it is diagonally opposed on wavelet coefficient.
Step2, the two images I for calculating adjacent momenttAnd It+1Variation light stream vector fieldPass through gradient Operator calculates the gradient of variation light stream vector field amplitudeLinear normalization is carried out to optical flow gradient to handle to obtain Movement edge
The structuring random forest F that Step3, foundation are made of N number of tree tree maximizes meter by the information gain of standard The bifurcated function on each tree is calculated, N is the number set in random forest;By image ItInput structure random forest F carries out non- Linear class prediction F (It), according to the prediction case p (I of each tree treet|Tree), statistical framework random forest is non-thread Property prediction resultIt is normalized to obtain the spatial domain skirt response F of each pixel of imaget (a,b);
Step4, the response of high yardstick wavelet coefficient maximal margin is projected into pixel dimension, and linear combination fine dimension Movement and spatial domain edge feature, fusion obtain Multiscale combination probability skirt response:
Wherein Bs t(a, b) is the maximal margin response at each pixel of wavelet field high yardstick objective contour;
Step5, by image sequenceOver-segmentation constitutes super-pixel set G, connection space neighborhood (a, b) ∈ εsAnd the time Neighborhood (a, b, t) ∈ εt,t+1, markov random file when to build sky;Wherein εsConnect between all adjacent pixels of spatial domain The set of line, εt,t+1The set of line between time-domain t moment and all adjacent pixels of t+1 moment two continuous frames images;
Step6, the unitary energy letter that markov random file is determined according to the convex optimum results of combination edge internal maps Several Ut, the class label l of a super-pixel i ∈ G is judged by the display model of all picture framesiBe foreground or background can It can property;
In the step Step6, markov random file is determined according to the convex optimum results of combination edge internal maps Unitary energy function item UtThe step of it is as follows:
Step6.1, the projection M according to the internal maps seed point at combined probability edge on a fine scale, establish image Foreground target in sequence maps weight total variation energy functional model:
Wherein first item is weight total variation item, and ▽ u are the gradient of alternately variable u;Section 2 is segment smoothing item, Mv-I Indicate that seed point mapping matrix M subtracts each other with gray value of image again with the dot product for replacing variable v;Relative importance between the two is logical Parameter ρ is crossed to balance;
Step6.2, the weight total variation segmentation problem of Pixel-level fine dimension is converted to constrained minimization problem:U*= argminu,vH (u, v), constraints are u ∈ [0,1], v ∈ [0,1];
Apply linear restriction with internal seeds point mapping matrix M between Step6.3, two alternating variables u and v, by multiplying Sub- alternating direction convex optimized algorithm, iteration, which updates to obtain, successively minimizes variable u*;
Step6.4, according to convex optimum results u*, training gauss hybrid models foreground parameter ΘFWith context parameter ΘB, obtain The unitary energy function item U of Markov random field model after convex optimization is smootht=Π (li|u*);Π is to be similar to outside super-pixel So;
Step6.5, the posterior probability for calculating current super-pixel RGB color vector, it is five to define super-pixel shape likelihood Π The negative log-likelihood function of the color character of ingredient gauss hybrid models:
Wherein Pr is probability density function, ciIndicate super-pixel index, the class label l of pixel ii=0 or li=1, Value is determined according to convex optimum results u*, and negative log-likelihood function is calculated further according to the gauss hybrid models of color clue Πt(li|u*)。
Step7, markov random file binary spatial domain energy function V and binary time domain energy function W are determined, calculated separately In spatial neighborhood (a, b) ∈ εsWith time neighborhood (a, b, t) ∈ εt,t+1Two super-pixel at both ends take class label respectively on line li∈ L and ljThe possibility of ∈ L, L are the set of all class labels;
In the step Step7, markov random file binary spatial domain energy function V and binary time domain energy function are determined The step of W, is as follows:
The consistency that Step7.1, binary item encode is embodied in room and time clue, is adjusted with the comparison of a standard Function defines binary spatial domain energy function V:
Wherein dist indicates two super-pixel ci,cjGeometric centerWithBetween Euclidean distance;Πt(li) It indicates to bear log-likelihood function Πt(li|U* abbreviation), σ representation space weight factors;
Step7.2, temporal target action being consistent property on image sequence is made by binary smooth potential function W, Binary time domain energy function W is encoded using overlapping and appearance information:
Wherein β indicates the time weighting factor;
The specific calculation of Step7.3, percentage φ of two super-pixel with same pixel contacted by light stream For:
Wherein # indicates the number of pixel.
Random field posterior probability classification approximate reasoning when Step8, super-pixel scale sky.By image sequenceOver-segmentation Constitute super-pixel set G, connection space neighborhood (a, b) ∈ εsWith time neighborhood (a, b, t) ∈ εt,t+1.Define super-pixel Markov Random field energy function E (L) solves foreground-background class label optimization class result L;
Step9, likelihood distribution of being classified by the posterior probability of energy function E (L) described in figure syncopation single-step iteration reasoning, Output makes the label L that energy function minimizes for one*
L*=argminLE(L)
Export L*Obtain the final segmentation result of dynamic scene image sequence.
The beneficial effects of the invention are as follows:
1) present invention is merged using multi-scale transform domain, spatial domain and time-domain clue, obtains a kind of completely new target wheel The combined probability edge extracting method of wide feature can effectively extract foreground in dynamic background image sequence on multiple scales The maximal margin of objective contour responds.
2) Markov established using the convex optimization of combination edge internal maps total variation that the method for the invention proposes Random field energy function, can effectively appearance color is inconsistent and shape at any time changed complex target from dynamic It is split in scene.
Description of the drawings
Fig. 1 is the flow diagram of the present invention;
Fig. 2 is 6 frame complex target dynamic scene image sequence datas of the present invention for test;
Fig. 3 is the present invention at the maximization wavelet coefficient edge that scale 1 detects;
Fig. 4 is the invention detects that the simultaneously movement edge after normalized;
Fig. 5 is detected by structuring random forest and the spatial domain edge after normalized;
Fig. 6 is combination edge proposed by the present invention multiple features fusion result;
Fig. 7 is the foreground internal maps that the present invention is calculated by the combination edge feature proposed;
Fig. 8 is the segmentation result that the present invention obtains in panda image sequence data.
Specific implementation mode
Embodiment 1:As shown in figures 1-8, a kind of complex target dynamic scene based on the convex optimization of Multiscale combination feature without Supervised segmentation method extracts wavelet field high yardstick target signature, and calculates movement and spatial domain skirt response first, by wavelet field, Spatially and temporally edge feature merges to obtain Multiscale combination probability skirt response, then establishes combination edge internal maps weight Total variation energy functional model calculates the convex optimization of weight total variation energy functional model using alternating direction, and defines super-pixel The unitary energy function item of markov random file and binary spatial domain energy function, change finally by one step of figure cutting when scale sky The posterior probability segmentation result of image sequence is obtained for reasoning;
The specific step of the complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature It is rapid as follows:
Step1, wavelet field high yardstick objective contour feature extraction;In image sequenceThe image I at each moment ttIn, Maximal margin response of the foreground wavelet field high yardstick objective contour at each location of pixels (a, b) is verified first, and T is video Totalframes;
In the step Step1, the maximal margin response definition at each pixel of wavelet field high yardstick objective contour is:
Bs t(a, b)=max (Hs t(a,b),Vs t(a,b),Ds t(a,b))
Wherein, Hs t、Vs tAnd Ds tIndicate image after wavelet transformation in the horizontal, vertical of the positions scale s (a, b) respectively With it is diagonally opposed on wavelet coefficient.
Step2, the two images I for calculating adjacent momenttAnd It+1Variation light stream vector fieldPass through gradient Operator calculates the gradient of variation light stream vector field amplitudeLinear normalization is carried out to optical flow gradient to handle to obtain Movement edge
The structuring random forest F that Step3, foundation are made of N number of tree tree maximizes meter by the information gain of standard The bifurcated function on each tree is calculated, N is the number set in random forest;By image ItInput structure random forest F carries out non- Linear class prediction F (It), according to the prediction case p (I of each tree treet|Tree), statistical framework random forest is non-thread Property prediction resultIt is normalized to obtain the spatial domain skirt response F of each pixel of imaget (a,b);
Step4, the response of high yardstick wavelet coefficient maximal margin is projected into pixel dimension, and linear combination fine dimension Movement and spatial domain edge feature, fusion obtain Multiscale combination probability skirt response:
Wherein Bs t(a, b) is the maximal margin response at each pixel of wavelet field high yardstick objective contour;
Step5, by image sequenceOver-segmentation constitutes super-pixel set G, connection space neighborhood (a, b) ∈ εsAnd the time Neighborhood (a, b, t) ∈ εt,t+1, markov random file when to build sky;Wherein εsConnect between all adjacent pixels of spatial domain The set of line, εt,t+1The set of line between time-domain t moment and all adjacent pixels of t+1 moment two continuous frames images;
Step6, the unitary energy letter that markov random file is determined according to the convex optimum results of combination edge internal maps Several Ut, the class label l of a super-pixel i ∈ G is judged by the display model of all picture framesiBe foreground or background can It can property;
In the step Step6, markov random file is determined according to the convex optimum results of combination edge internal maps Unitary energy function item UtThe step of it is as follows:
Step6.1, the projection M according to the internal maps seed point at combined probability edge on a fine scale, establish image Foreground target in sequence maps weight total variation energy functional model:
Wherein first item is weight total variation item, and ▽ u are the gradient of alternately variable u;Section 2 is segment smoothing item, Mv-I Indicate that seed point mapping matrix M subtracts each other with gray value of image again with the dot product for replacing variable v;Relative importance between the two is logical Parameter ρ is crossed to balance;
Step6.2, the weight total variation segmentation problem of Pixel-level fine dimension is converted to constrained minimization problem:U*= argminu,vH (u, v), constraints are u ∈ [0,1], v ∈ [0,1];
Apply linear restriction with internal seeds point mapping matrix M between Step6.3, two alternating variables u and v, by multiplying Sub- alternating direction convex optimized algorithm, iteration, which updates to obtain, successively minimizes variable u*;
Step6.4, according to convex optimum results u*, training gauss hybrid models foreground parameter ΘFWith context parameter ΘB, obtain The unitary energy function item U of Markov random field model after convex optimization is smootht=Π (li|u*);Π is to be similar to outside super-pixel So;
Step6.5, the posterior probability for calculating current super-pixel RGB color vector, it is five to define super-pixel shape likelihood Π The negative log-likelihood function of the color character of ingredient gauss hybrid models:
Wherein Pr is probability density function, ciIndicate super-pixel index, the class label l of pixel ii=0 or li=1, Value is determined according to convex optimum results u*, and negative log-likelihood function is calculated further according to the gauss hybrid models of color clue Πt(li|u*)。
Step7, markov random file binary spatial domain energy function V and binary time domain energy function W are determined, calculated separately In spatial neighborhood (a, b) ∈ εsWith time neighborhood (a, b, t) ∈ εt,t+1Two super-pixel at both ends take class label respectively on line li∈ L and ljThe possibility of ∈ L, L are the set of all class labels;
In the step Step7, markov random file binary spatial domain energy function V and binary time domain energy function are determined The step of W, is as follows:
The consistency that Step7.1, binary item encode is embodied in room and time clue, is adjusted with the comparison of a standard Function defines binary spatial domain energy function V:
Wherein dist indicates two super-pixel ci,cjGeometric centerWithBetween Euclidean distance;Πt(li) It indicates to bear log-likelihood function Πt(li|U* abbreviation), σ representation space weight factors;
Step7.2, temporal target action being consistent property on image sequence is made by binary smooth potential function W, Binary time domain energy function W is encoded using overlapping and appearance information:
Wherein β indicates the time weighting factor;
The specific calculation of Step7.3, percentage φ of two super-pixel with same pixel contacted by light stream For:
Wherein # indicates the number of pixel.
Random field posterior probability classification approximate reasoning when Step8, super-pixel scale sky.By image sequenceOver-segmentation Constitute super-pixel set G, connection space neighborhood (a, b) ∈ εsWith time neighborhood (a, b, t) ∈ εt,t+1.Define super-pixel Markov Random field energy function E (L) solves foreground-background class label optimization class result L;
Step9, likelihood distribution of being classified by the posterior probability of energy function E (L) described in figure syncopation single-step iteration reasoning, Output makes the label L that energy function minimizes for one*
L*=argminLE(L)
Export L*Obtain the final segmentation result of dynamic scene image sequence.
Embodiment 2:As shown in figures 1-8, a kind of complex target dynamic scene based on the convex optimization of Multiscale combination feature without Supervised segmentation method inputs the image sequence data containing dynamic scene information residing for target information and target, calculates multiple dimensioned Wavelet Edge feature, structuring random forest spatial edge feature and movement gradient feature, before in dynamic image sequence The combined probability edge of the maximal margin response extraction objective contour feature of scape objective contour;Structure foreground position on a fine scale The mapping convex Optimized model of seed point total variation is set, the objective contour smoothly extracted in high yardstick of alternately more newly arriving by variable is special Combined probability edge internal maps seed point is levied, the unitary of markov random file is established according to the target location mapping after smooth Data capacity function and binary space, time energy function item utilize Ma Er when figure cutting an iteration completion super-pixel grade sky Section husband random field posterior probability reasoning is to realize the Pixel-level segmentation of dynamic scene image sequence.
It is dynamic that the present invention overcomes the deficiencies of the prior art and provide a kind of complex target based on the convex optimization of Multiscale combination feature State scene image sequence non-formaldehyde finishing method.The flow diagram of the present invention is as shown in Figure 1.It is described special based on Multiscale combination The specific implementation step for levying the complex target dynamic scene image sequence non-formaldehyde finishing method of convex optimization is as follows:
Acquired dynamic image sequence data are inputted first, illumination variation, the foreground of scene environment when to obtaining image It without specifically constraint and is required with image-forming conditions such as background contrasts, the movement of background and the movements of camera itself.Fig. 2 is The present invention for test 6 frame complex target dynamic scene image sequence datas, wherein foreground target appearance color it is uneven and And foreground shape changes over time, it is opposing stationary when between foreground and background there are apparent camera motion, when and opposite transport Dynamic, background more spot is miscellaneous.
(1) then complete high yardstick objective contour feature extraction.In image sequencePer frame image ItIn verify foreground High yardstick objective contour is responded in the maximal margin of wavelet field.It is defined according to high yardstick object edge proposed by the present invention:
Bs t(a, b)=max (Hs t(a,b),Vs t(a,b),Ds t(a,b))
Wherein, Hs t、Vs tAnd Ds tIndicate image after wavelet transformation in the horizontal, vertical of the positions scale s (a, b) respectively With it is diagonally opposed on wavelet coefficient.
Extraction foreground high yardstick objective contour is responded in the maximal margin of wavelet field.The method of the invention is examined in scale 1 The maximization wavelet coefficient edge measured is as shown in Figure 3.
(2) the two images I of adjacent moment is calculatedtAnd It+1Variation light stream vector fieldPass through gradient operator Calculate the gradient of variation light stream vector field amplitudeOptical flow gradient progress linear normalization is handled and is moved EdgeThe method of the invention detects that the movement edge after simultaneously normalized is as shown in Figure 4.
(3) the structuring random forest F being made of N number of tree tree is established, is maximized and is calculated by the information gain of standard Bifurcated function on each tree, N are the number set in random forest.By image ItInput structure random forest F carries out non-thread Property class prediction F (It), according to the prediction case p (I of each tree treet|Tree), statistical framework random forest is non-linear Prediction resultIt is normalized to obtain the spatial domain skirt response F of each pixel of imaget(a, b).The method of the invention detects that the spatial domain edge after simultaneously normalized is as shown in Figure 5 by structuring random forest.
(4) response of high yardstick wavelet coefficient maximal margin is projected into pixel dimension, and the fortune of linear combination fine dimension Dynamic and spatial domain edge feature, fusion obtain Multiscale combination probability skirt response:
Wherein Bs t(a, b) is the maximal margin response at each pixel of wavelet field high yardstick objective contour;
The results are shown in Figure 6 for combination edge proposed by the present invention multiple features fusion.
(5) the projection M according to the internal maps seed point at combined probability edge on a fine scale, is established in image sequence Foreground target map weight total variation energy functional model:
Weight total variation item and segment smoothing item weight parameter ρ values between the two are 1.8.
(6) the weight total variation segmentation problem of Pixel-level fine dimension is converted to constrained minimization problem:
U*=argminu,vH (u, v), constraints are u ∈ [0,1], v ∈ [0,1].
Apply linear restriction with internal seeds point mapping matrix M between (7) two alternating variables u and v, is replaced by multiplier Direction convex optimized algorithm carries out the proximal end project between alternately variable successively, and iteration, which updates to obtain, minimizes variable u*.This Invention is described as shown in Figure 7 by combining the foreground internal maps that edge feature is calculated.
(8) according to convex optimum results u*, training gauss hybrid models foreground parameter ΘFWith context parameter ΘB, obtain convex excellent Change the unitary data capacity function item of smooth rear Markov random field modelsΠ is to be similar to outside super-pixel So.
(9) posterior probability of current super-pixel RGB color vector is calculated, definition super-pixel shape likelihood is five ingredient Gausses The negative log-likelihood function of the color character of mixed model:
Wherein Pr is probability density function, ciIndicate super-pixel index, the class label l of pixel ii=0 or li=1, Value is determined according to convex optimum results u*, and negative log-likelihood function is calculated further according to the gauss hybrid models of color clue Πt(li|u*)。
(10) consistency of binary item coding is embodied in room and time clue, with Bayesian segmentation system and is defined Space binary item V, the comparison adjustment item as a standard:
Wherein dist indicates two super-pixel ci,cjGeometric centerWithBetween Euclidean distance, parameter σ takes Value is set as 2.8.
(11) temporal target action being consistent property on image sequence is made by binary smooth potential function W, used Overlapping and appearance information carry out scramble time constraint:
Wherein, parameter beta value is set as 1.6.
(12) the percentage φ that there is same pixel by two super-pixel that light stream contacts is calculated:
Wherein # indicates the number of pixel.
(13) random field posterior probability classification approximate reasoning when super-pixel scale sky.By image sequenceOver-segmentation structure At super-pixel set G, connection space neighborhood (a, b) ∈ εsWith time neighborhood (a, b, t) ∈ εt,t+1.Define super-pixel Markov Random field energy function E (L) solves foreground-background class label optimization class result L.
(14) defeated by the classification likelihood distribution of the posterior probability of energy function E (L) described in figure syncopation single-step iteration reasoning Go out makes the label L that energy function minimizes for one*
L*=argminLE(L)
Export L*Obtain the final segmentation result of dynamic scene image sequence.It is of the present invention to be based on Multiscale combination feature Segmentation result such as Fig. 8 that the complex target dynamic scene non-formaldehyde finishing method of convex optimization obtains in panda image sequence data It is shown.As seen from the figure, in due to dynamic background image caused by camera motion and environmental change, appearance color it is uneven and The foreground panda semantic objects region that shape changes over time can access effective segmentation at all six moment.
The specific implementation mode of the present invention is explained in detail above in conjunction with attached drawing, but the present invention is not limited to above-mentioned Embodiment within the knowledge of a person skilled in the art can also be before not departing from present inventive concept Put that various changes can be made.

Claims (4)

1. a kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature, feature exist In:Wavelet field high yardstick target signature is extracted first, and calculates movement and spatial domain skirt response, by wavelet field, spatially and temporally Edge feature merges to obtain Multiscale combination probability skirt response, then establishes combination edge internal maps weight total variation energy Functional model calculates the convex optimization of weight total variation energy functional model, and horse when defining super-pixel scale sky using alternating direction The unitary energy function item and binary spatial domain energy function of Er Kefu random fields, obtain finally by figure cutting single-step iteration reasoning The posterior probability segmentation result of image sequence;
The specific steps of the complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature are such as Under:
Step1, wavelet field high yardstick objective contour feature extraction;In image sequenceThe image I at each moment ttIn, first Maximal margin response of the foreground wavelet field high yardstick objective contour at each location of pixels (a, b) is verified, T is total frame of video Number;
Step2, the two images I for calculating adjacent momenttAnd It+1Variation light stream vector fieldPass through gradient operator meter Calculate the gradient of variation light stream vector field amplitudeLinear normalization is carried out to optical flow gradient to handle to obtain movement side Edge
The structuring random forest F that Step3, foundation are made of N number of tree tree is maximized by the information gain of standard and is calculated often Bifurcated function on a tree, N are the number set in random forest;By image ItInput structure random forest F carries out non-linear Class prediction F (It), according to the prediction case p (I of each tree treet|Tree), statistical framework random forest is non-linear pre- Survey resultIt is normalized to obtain the spatial domain skirt response F of each pixel of imaget(a,b);
Step4, the response of high yardstick wavelet coefficient maximal margin is projected into pixel dimension, and the movement of linear combination fine dimension With spatial domain edge feature, fusion obtains Multiscale combination probability skirt response:
WhereinFor the maximal margin response at each pixel of wavelet field high yardstick objective contour;
Step5, by image sequenceOver-segmentation constitutes super-pixel set G, connection space neighborhood (a, b) ∈ εsWith time neighborhood (a,b,t)∈εt,t+1, markov random file when to build sky;Wherein εsThe line between all adjacent pixels of spatial domain Set, εt,t+1The set of line between time-domain t moment and all adjacent pixels of t+1 moment two continuous frames images;
Step6, the unitary energy function item that markov random file is determined according to the convex optimum results of combination edge internal maps Ut, the class label l of a super-pixel i ∈ G is judged by the display model of all picture framesiIt is the possibility of foreground or background;
Step7, markov random file binary spatial domain energy function V and binary time domain energy function W are determined, calculated separately in sky Between neighborhood (a, b) ∈ εsWith time neighborhood (a, b, t) ∈ εt,t+1Two super-pixel at both ends take class label l respectively on linei∈L And ljThe possibility of ∈ L, L are the set of all class labels;
Step8, super-pixel markov random file energy function E (L) is defined to solve foreground-background class label optimization class As a result L;
Step9, by formula L*=argminLThe posterior probability classification likelihood of energy function E (L) described in E (L) single-step iteration reasoning, it is defeated Go out L*Obtain final dynamic scene Segmentation of Image Sequences result.
2. the complex target dynamic scene non-formaldehyde finishing according to claim 1 based on the convex optimization of Multiscale combination feature Method, it is characterised in that:In the step Step1, the maximal margin response at each pixel of wavelet field high yardstick objective contour It is defined as:
Bs t(a, b)=max (Hs t(a,b),Vs t(a,b),Ds t(a,b))
Wherein, Hs t、Vs tAnd Ds tIndicate image after wavelet transformation in the horizontal, vertical and right of the positions scale s (a, b) respectively Wavelet coefficient on angular direction.
3. the complex target dynamic scene non-formaldehyde finishing according to claim 1 based on the convex optimization of Multiscale combination feature Method, it is characterised in that:In the step Step6, Markov is determined according to the convex optimum results of combination edge internal maps The unitary energy function item U of random fieldtThe step of it is as follows:
Step6.1, the projection M according to the internal maps seed point at combined probability edge on a fine scale, establish image sequence In foreground target map weight total variation energy functional model:
Wherein first item is weight total variation item,For the gradient of alternately variable u;Section 2 is segment smoothing item, and Mv-I is indicated Seed point mapping matrix M subtracts each other with gray value of image again with the dot product for replacing variable v;Relative importance between the two passes through ginseng ρ is counted to balance;
Step6.2, the weight total variation segmentation problem of Pixel-level fine dimension is converted to constrained minimization problem:U*= argminu,vH (u, v), constraints are u ∈ [0,1], v ∈ [0,1];
Apply linear restriction with internal seeds point mapping matrix M between Step6.3, two alternating variables u and v, is handed over by multiplier For direction convex optimized algorithm, iteration, which updates to obtain, successively minimizes variable u*;
Step6.4, according to convex optimum results u*, training gauss hybrid models foreground parameter ΘFWith context parameter ΘB, obtain convex excellent Change the unitary energy function item U of smooth rear Markov random field modelt=Π (li|u*);Π is super-pixel shape likelihood;
Step6.5, the posterior probability for calculating current super-pixel RGB color vector, it is five ingredients to define super-pixel shape likelihood Π The negative log-likelihood function of the color character of gauss hybrid models:
Wherein Pr is probability density function, ciIndicate super-pixel index, the class label l of pixel ii=0 or li=1, value It is determined according to convex optimum results u*, negative log-likelihood function Π is calculated further according to the gauss hybrid models of color cluet(li| u*)。
4. the complex target dynamic scene non-formaldehyde finishing according to claim 3 based on the convex optimization of Multiscale combination feature Method, it is characterised in that:In the step Step7, markov random file binary spatial domain energy function V and binary time domain are determined The step of energy function W, is as follows:
The consistency that Step7.1, binary item encode is embodied in room and time clue, with the comparison adjustment function of a standard Define binary spatial domain energy function V:
Wherein dist indicates two super-pixel ci,cjGeometric centerWithBetween Euclidean distance;Πt(li) indicate negative Log-likelihood function Πt(li|u*) abbreviation, σ representation space weight factors;
Step7.2, temporal target action being consistent property on image sequence is made by binary smooth potential function W, used It is overlapped and encodes binary time domain energy function W with appearance information:
Wherein β indicates the time weighting factor;
Step7.3, it is by the specific calculation of percentage φ of two super-pixel with same pixel of light stream contact:
Wherein # indicates the number of pixel.
CN201610261547.6A 2016-04-26 2016-04-26 A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature Active CN105869178B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610261547.6A CN105869178B (en) 2016-04-26 2016-04-26 A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610261547.6A CN105869178B (en) 2016-04-26 2016-04-26 A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature

Publications (2)

Publication Number Publication Date
CN105869178A CN105869178A (en) 2016-08-17
CN105869178B true CN105869178B (en) 2018-10-23

Family

ID=56629177

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610261547.6A Active CN105869178B (en) 2016-04-26 2016-04-26 A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature

Country Status (1)

Country Link
CN (1) CN105869178B (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106846362B (en) * 2016-12-26 2020-07-24 歌尔科技有限公司 Target detection tracking method and device
JP7115846B2 (en) * 2016-12-30 2022-08-09 ダッソー システムズ Generate Segmented Image Using Markov Random Field Optimization
CN108875756B (en) * 2017-05-08 2021-01-05 深圳荆虹科技有限公司 Behavior category acquisition method and device for video
CN107918759A (en) * 2017-10-09 2018-04-17 大圣科技股份有限公司 Automatic segmentation recognition method, electronic equipment and the storage medium of indoor object
CN108647649B (en) * 2018-05-14 2021-10-01 中国科学技术大学 Method for detecting abnormal behaviors in video
CN109063577B (en) * 2018-07-05 2021-08-31 浙江大学 Satellite image segmentation optimal segmentation scale determination method based on information gain rate
CN109165565A (en) * 2018-08-01 2019-01-08 西安交通大学 A kind of video object discovery and dividing method based on Coupled Dynamic Markov Network
CN109492530B (en) * 2018-10-10 2022-03-04 重庆大学 Robust visual object tracking method based on depth multi-scale space-time characteristics
CN110264482B (en) * 2019-05-10 2022-09-09 河南科技大学 Active contour segmentation method based on transformation matrix factorization of noose set
CN110232663B (en) * 2019-05-16 2021-04-30 福建自贸试验区厦门片区Manteia数据科技有限公司 Method and device for correcting automatic sketching model of organs at risk
CN110827318A (en) * 2019-10-18 2020-02-21 天津大学 Target tracking method based on fusion of multilayer semantic features and multi-response graph
CN111127479A (en) * 2019-12-17 2020-05-08 昆明理工大学 Level set image segmentation method based on curve area
CN111340852B (en) * 2020-03-10 2022-09-27 南昌航空大学 Image sequence optical flow calculation method based on optimized semantic segmentation
CN111950517A (en) * 2020-08-26 2020-11-17 司马大大(北京)智能系统有限公司 Target detection method, model training method, electronic device and storage medium
CN112164092B (en) * 2020-10-13 2022-09-27 南昌航空大学 Generalized Markov dense optical flow determination method and system
CN113420179B (en) * 2021-06-24 2022-03-22 杭州电子科技大学 Semantic reconstruction video description method based on time sequence Gaussian mixture hole convolution

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101540047A (en) * 2009-04-30 2009-09-23 西安电子科技大学 Texture image segmentation method based on independent Gaussian hybrid model

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101540047A (en) * 2009-04-30 2009-09-23 西安电子科技大学 Texture image segmentation method based on independent Gaussian hybrid model

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Bag of Textons for Image Segmentation via Soft Clustering and Convex Shift;Zhiding Yu et al.;《2012 IEEE Conference on Computer Vision and Pattern Recognition》;20120616;第781-788页 *
基于多尺度结构张量的多类无监督彩色纹理图像分割方法;杨勇 等;《计算机辅助设计与图形学学报》;20140531;第26卷(第5期);第812-825页 *
高分辨率SAR图像分割与分类方法研究;冯籍澜;《中国博文学位论文全文数据库(信息科技辑)》;20160315(第3期);正文第6-7页,第11页,第42页,第51页,第53页,第57页 *

Also Published As

Publication number Publication date
CN105869178A (en) 2016-08-17

Similar Documents

Publication Publication Date Title
CN105869178B (en) A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature
CN110175566B (en) Hand posture estimation system and method based on RGBD fusion network
CN106897670B (en) Express violence sorting identification method based on computer vision
CN108197587B (en) Method for performing multi-mode face recognition through face depth prediction
CN106997597B (en) It is a kind of based on have supervision conspicuousness detection method for tracking target
CN106682598B (en) Multi-pose face feature point detection method based on cascade regression
CN104599275B (en) The RGB-D scene understanding methods of imparametrization based on probability graph model
CN106909902B (en) Remote sensing target detection method based on improved hierarchical significant model
CN109800794B (en) Cross-camera re-identification fusion method and system for appearance similar targets
CN112784736B (en) Character interaction behavior recognition method based on multi-modal feature fusion
CN109657612B (en) Quality sorting system based on facial image features and application method thereof
CN105528794A (en) Moving object detection method based on Gaussian mixture model and superpixel segmentation
CN111882586B (en) Multi-actor target tracking method oriented to theater environment
CN108021889A (en) A kind of binary channels infrared behavior recognition methods based on posture shape and movable information
CN107066969A (en) A kind of face identification method
CN114758288A (en) Power distribution network engineering safety control detection method and device
CN110176016B (en) Virtual fitting method based on human body contour segmentation and skeleton recognition
CN103927511A (en) Image identification method based on difference feature description
CN104102904B (en) A kind of static gesture identification method
CN105279769A (en) Hierarchical particle filtering tracking method combined with multiple features
CN103886619A (en) Multi-scale superpixel-fused target tracking method
CN105574545B (en) The semantic cutting method of street environment image various visual angles and device
CN106570887A (en) Adaptive Mean Shift target tracking method based on LBP features
CN110334656A (en) Multi-source Remote Sensing Images Clean water withdraw method and device based on information source probability weight
CN104751111A (en) Method and system for recognizing human action in video

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: He Zifen

Inventor after: Zhang Yinhui

Inventor after: Wu Xing

Inventor after: Zhang Yunsheng

Inventor after: Wang Sen

Inventor before: Zhang Yinhui

Inventor before: He Zifen

Inventor before: Wu Xing

Inventor before: Zhang Yunsheng

Inventor before: Wang Sen

CB03 Change of inventor or designer information
GR01 Patent grant
GR01 Patent grant