CN105590327A - Motion estimation method and apparatus - Google Patents

Motion estimation method and apparatus Download PDF

Info

Publication number
CN105590327A
CN105590327A CN201410578173.1A CN201410578173A CN105590327A CN 105590327 A CN105590327 A CN 105590327A CN 201410578173 A CN201410578173 A CN 201410578173A CN 105590327 A CN105590327 A CN 105590327A
Authority
CN
China
Prior art keywords
motion
cutting apart
frame
model
apart
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410578173.1A
Other languages
Chinese (zh)
Inventor
陈子冲
刘余钱
章国锋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN201410578173.1A priority Critical patent/CN105590327A/en
Publication of CN105590327A publication Critical patent/CN105590327A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention, which belongs to the technical field of the digital signal processing, discloses a motion estimation method and apparatus. The method comprises: an initial light stream of each frame of image in an inputted image sequence is obtained; according to color information, each frame of image is divided into at least one segment; initial motion models of all segments in each frame of image are obtained; motion segmentation time-sequence matching is carried out on initial motion models of all segments in each frame of image, so that the same object in the image sequence has the same mark in each frame in the image sequence; and according to the time-sequence matching result, a motion segment and an output light stream are obtained. According to the invention, the motion segment and motion estimation of a long image sequence can be processed and the motion segment and motion estimation results with good time-space coherence can be obtained, so that the time-space coherence of the motion segments can be improved effectively.

Description

Method for estimating and device
Technical field
The present invention relates to digital signal processing technique field, particularly a kind of method for estimating and device.
Background technology
Light stream (English: the OpticalFlow) estimation technique, i.e. estimation in sequence of video images, is motion diagramA kind of important method in picture analysis, it has a wide range of applications in computer vision field, for example:To image tracing, video interpolation and three-dimensional reconstruction etc. Wherein, light stream refers to the apparent fortune of brightness of image patternMoving, express the variation of image, due to the information that it has comprised target travel, therefore can observed person be used forDetermine the motion conditions of target. Existing light stream method of estimation conventionally need to be to optical flow field (English: OpticalFlowField) analyze, the set of all light streams is called optical flow field, concrete, and optical flow field refers in imageOne two dimension (the English: 2-Dimensions that all pixels form; Be called for short: 2D) instantaneous velocity, whereinTwo-dimension speed vector is that the three dimensional velocity vectors of visible point in scenery is in the projection of imaging surface. So light stream is notOnly comprise the movable information of observed object, but also included the abundant letter about object three-dimensional structureBreath. In recent years, carrying out based on optical flow field the method that motion segmentation improves estimation accuracy causes graduallyThe concern of academia. How further to improve the accuracy rate that light stream is estimated, and obtaining object of which movement estimationIn time, obtains motion segmentation accurately and becomes gradually the main flow direction of motion estimation techniques development.
Comparatively conventional a kind of method for estimating has proposed a kind of fortune of combining of blocking processing that shows at presentThe moving unified variate model of estimating and cutting apart, represents optical flow field with multiple labeling, and each mark representsA kind of parameterized motor pattern, also represent a motion segmentation simultaneously.
Although can carrying out motion segmentation and the light stream of multiframe, existing method for estimating estimates to obtain comparativelyMotion segmentation accurately still, due to the restriction of space-time cost, can only be carried out a small amount of a few frame figure conventionally simultaneouslyThe processing of picture, is difficult to ensure the space-time consistency of motion segmentation, and wherein, the space-time consistency of motion segmentation isRefer to the estimation of video image and be segmented in time domain to there is uniformity, in spatial domain, there is continuity.
Summary of the invention
In order to improve the space-time consistency of motion segmentation, the embodiment of the present invention provides a kind of method for estimatingAnd device. Described technical scheme is as follows:
First aspect, provides a kind of method for estimating, and described method comprises:
Obtain the initial light stream of each two field picture in the image sequence of input;
According to colouring information, described each two field picture being divided into at least one cuts apart;
Obtain all initial motion model of cutting apart in described each two field picture;
All initial motion model of cutting apart in described each two field picture are carried out to the sequential coupling of motion segmentation,To make thering is identical mark in same target in described image sequence each frame in described image sequenceNote;
Obtain motion segmentation and output light stream according to the result of sequential coupling.
In the possible embodiment of the first of first aspect, described according to all points of each two field pictureBefore the initial motion model of cutting is carried out the sequential coupling of motion segmentation, described method also comprises:
According to all initial motion model of cutting apart in described each two field picture, same movement form will be there isInitial motion model is carried out cluster optimization; And/or
Merge consistent the cutting apart of moving in described each two field picture.
In conjunction with the possible embodiment of the first of first aspect or first aspect, in the possible enforcement of the secondIn mode, described by all initial motion model of cutting apart in described each two field picture carry out motion segmentation timeOrder coupling comprises:
Using the first two field picture in the image sequence of input as initial matching masterplate, its each cutting apart as oneIndividual data sectional, for data sectional described in each distributes unique mark;
For cutting apart in t frameLight stream according to t frame to t-1 frame, follows the tracks ofIn each pixel xtCorresponding points x in t-1 framet-1, wherein, xt-1In t-1 frame, be designated as cutting apart of place?StructureThe set becomingExpression is cut apartCorresponding likely cutting apart in t-1 frame;
ForIn each cutting apartCalculateWithMatching rate:
r t - > t - 1 k - > k ′ = | x t | x t ∈ s t k , x t - 1 ∈ s t - 1 k ′ | | s t - 1 k ′ |
Wherein, ︱ xt︱ is for cutting apartIn with cut apartThe number of corresponding pixel, denominator represents to cut apartThe number of mid point;
CalculateWithMatching rate:
r t - 1 - > t k ′ - > k = c | s t k |
Wherein, rightIn each cutting apartLight stream according to t-1 frame to t frame, obtainsIn eachPoint xt-1Corresponding points x in tt, statistics xtBelong toThe number of point be designated as c,For cutting apartInThe number of point;
CalculateWithFinal matching rate:
r t , t - 1 k , k ′ = min ( r t - > t - 1 k - > k ′ , r t - 1 - > t k ′ - > k ) ;
WhenTime, determineWithCoupling, willFlag update beMark, and willJoinIn the data sectional at place, wherein, σmFor default matching threshold;
WhenWhile not existing coupling to cut apart in t-1 frame, willWith mating cutting apart in t-2 frame;
In the time that being greater than default coupling step-length, matching times finishes coupling, for not existing coupling to cut apartDistributeA new mark, willJoin in new data sectional.
In conjunction with the possible embodiment of the second of first aspect, in the third possible embodiment, instituteState according to the result of sequential coupling and obtain motion segmentation and export light stream comprising: divide according to the result of sequential couplingDo not carry out the space-time consistency optimization of motion segmentation and the space-time consistency optimization of estimation, obtain space-time oneThe motion segmentation causing and output light stream.
In conjunction with the third possible embodiment of first aspect, in the 4th kind of possible embodiment, instituteThe space-time consistency optimization of stating motion segmentation comprises:
The dividing mark of fixing each two field picture is constant;
According to light stream, follow the tracks of each pixel x in t frametAt adjacent NtCorresponding points in frame, are waitedSelect tag set P (xt);
According to described candidate's tag set P (xt) obtain the mark that occurrence number is maximum;
Work as Cs<Nt/ 2 or CPt<σPtTime, wherein, CsFor the number of cutting apart comprising in each data sectionalAmount, CPtFor the number of all cut-points in described data sectional, σPtFor predetermined threshold value, for described dataEach cutting apart in segmentation, according to the mark situation of the corresponding points in other frames of the point in described data sectional,Select the data sectional of an optimum Match to force to merge;
Upgrade the mark of remaining data sectional.
In conjunction with the third possible embodiment of first aspect, in the 5th kind of possible embodiment, instituteThe space-time consistency optimization of stating estimation comprises:
According to current motion, obtain data sectional that each mark l represents corresponding cut apart imitative in each framePenetrate transformation model
For t frame, according to affine Transform ModelCalculate light stream, to make each in t framePixel xtCorresponding points x in other frames t 't'In the Euclidean distance sum minimum of color space;
Upgrade the light stream of described t frame;
Carry out the detection of occlusion area;
The motion of occlusion area is shown to correction.
In conjunction with the possible embodiment of the first of first aspect or first aspect, the 6th kind of possible enforcementIn mode, described initial motion model comprises affine model;
Described in described obtaining, in each two field picture, all initial motion model of cutting apart comprise:
Calculate the parameter of affine transformation, obtain the affine model of matching initial motion;
Obtain the affine motion field of each two field picture according to described affine model;
Described affine motion field and initial optical flow field are merged;
Carry out the detection of occlusion area according to the uniformity of motion;
The affine motion that occlusion area is cut apart according to the place rectification of explicitly moving.
In conjunction with the 6th kind of possible embodiment of first aspect, in the 7th kind of possible embodiment, instituteState according to all initial motion model of cutting apart in described each two field picture, by have same movement form at the beginning ofBeginning motion model carries out cluster optimization and comprises:
To sorting all the cutting apart according to size in a two field picture, successively descending according to areaScreening affine model;
When cutting apart siArea be greater than preset area SizesOr the affine model collection { quantity of affine model in A}Be less than predetermined number NumATime, will cut apart siAffine model AiWith { A} compares, if { exist in A}A same A of affine model A'iEuclidean distance be less than predeterminable range ThA, give up current affine model Ai,Otherwise by AiJoin that { A} obtains least affine Models Sets { A}.
In conjunction with the 6th kind of possible embodiment of first aspect, in the 8th kind of possible embodiment, instituteState and merge consistent the cutting apart of moving in described each two field picture and comprise:
For each s of cutting apart, calculate the color distortion of cutting apart interior all pixels under the current affine model of cutting apartSummation;
Computed segmentation SiThe adjacent set { S of cutting apart of candidateNei, wherein, cut apart s ' completely for each candidate is adjacentFoot 2|s'| > | s|;
For { SNeiIn each adjacent s ' of cutting apart, calculate respectively the uniformity of moving on the same s ' of s bordersIn the uniformity of affine model of the same s ' of current motion of pixelPixel in s is at the affine model of s 'The ratio C of the color distortion of the lower colour-difference similarities and differences self affine models,s'
From { SNeiIn select and meet Ss,s'>0.5,Fs,s'>0.5,Cs,s'The adjacent of > 0.5 cut apart, by Cs,s'MaximumAdjacent cutting apart as segmentation candidates, be designated as s*, work as Cs,s*>θwTime, wherein, θwFor default ratio, willS merges in s*;
Each s of cutting apart is merged, if the last merged sum of cutting apart has exceeded one of initial segmentation numberHalf, re-start the estimation of affine model.
In the 9th kind of possible embodiment of first aspect, described initial light stream comprise forward direction light stream and/orOppositely light stream.
Second aspect, provides a kind of movement estimation apparatus, and described device comprises:
Acquiring unit, for obtaining the initial light stream of each two field picture of image sequence of input;
Cutting unit, cuts apart for described each two field picture being divided into at least one according to colouring information;
Motion model estimation unit, for obtaining the described all initial motion model of cutting apart of each two field picture;
Sequential matching unit, for transporting all described each two field picture initial motion model of cutting apartThe moving sequential coupling of cutting apart, to make every in described image sequence of same target in described image sequenceIn one frame, there is identical mark;
Processing unit, for obtaining motion segmentation and output light stream according to the result of sequential coupling.
In the possible embodiment of the first of second aspect, described device also comprises:
Cluster cell, for according to all initial motion model of cutting apart of described each two field picture, will haveThe initial motion model of same movement form is carried out cluster optimization; And/or
Merge cells, for merging consistent the cutting apart of moving of described each two field picture.
In conjunction with the possible embodiment of the first of second aspect or second aspect, in the possible enforcement of the secondIn mode, described sequential matching unit specifically for:
Using the first two field picture in the image sequence of input as initial matching masterplate, its each cutting apart as oneIndividual data sectional, for data sectional described in each distributes unique mark;
For cutting apart in t frameLight stream according to t frame to t-1 frame, follows the tracks ofIn each pixel xtCorresponding points x in t-1 framet-1, wherein, xt-1In t-1 frame, be designated as cutting apart of place?StructureThe set becomingExpression is cut apartCorresponding likely cutting apart in t-1 frame;
ForIn each cutting apartCalculateWithMatching rate:
r t - > t - 1 k - > k ′ = | x t | x t ∈ s t k , x t - 1 ∈ s t - 1 k ′ | | s t - 1 k ′ |
Wherein, ︱ xt︱ is for cutting apartIn with cut apartThe number of corresponding pixel, denominator represents to cut apartThe number of mid point;
CalculateWithMatching rate:
r t - 1 - > t k ′ - > k = c | s t k |
Wherein, rightIn each cutting apartLight stream according to t-1 frame to t frame, obtainsIn eachPoint xt-1Corresponding points x in tt, statistics xtBelong toThe number of point be designated as c,For cutting apartInThe number of point;
CalculateWithFinal matching rate:
r t , t - 1 k , k ′ = min ( r t - > t - 1 k - > k ′ , r t - 1 - > t k ′ - > k ) ;
WhenTime, determineWithCoupling, willFlag update beMark, and willJoinIn the data sectional at place, wherein, σmFor default matching threshold;
WhenWhile not existing coupling to cut apart in t-1 frame, willWith mating cutting apart in t-2 frame;
In the time that being greater than default coupling step-length, matching times finishes coupling, for not existing coupling to cut apartDistributeA new mark, willJoin in new data sectional.
In conjunction with the possible embodiment of the second of second aspect, in the third possible embodiment, instituteStating processing unit comprises: motion segmentation optimizes module and estimation is optimized module;
Described motion segmentation is optimized module, for carry out the space-time one of motion segmentation according to the result of sequential couplingThe optimization of causing property;
Described estimation is optimized module, for carry out the space-time one of estimation according to the result of sequential couplingThe optimization of causing property.
In conjunction with the third possible embodiment of second aspect, in the 4th kind of possible embodiment, instituteState fortune motion segmentation optimize module specifically for:
The dividing mark of fixing each two field picture is constant;
According to light stream, follow the tracks of each pixel x in t frametAt adjacent NtCorresponding points in frame, are waitedSelect tag set P (xt);
According to described candidate's tag set P (xt) obtain the mark that occurrence number is maximum;
Work as Cs<Nt/ 2 or CPt<σPtTime, wherein, CsFor the number of cutting apart comprising in each data sectionalAmount, CPtFor the number of all cut-points in described data sectional, σPtFor predetermined threshold value, for described dataEach cutting apart in segmentation, according to the mark situation of the corresponding points in other frames of the point in described data sectional,Select the data sectional of an optimum Match to force to merge;
Upgrade the mark of remaining data sectional.
In conjunction with the third possible embodiment of second aspect, in the 5th kind of possible embodiment, instituteState estimation optimize module specifically for:
According to current motion, obtain data sectional that each mark l represents corresponding cut apart imitative in each framePenetrate transformation model
For t frame, according to affine Transform ModelCalculate light stream, to make each in t framePixel xtCorresponding points x in other frames t 't'In the Euclidean distance sum minimum of color space;
Upgrade the light stream of described t frame;
Carry out the detection of occlusion area;
The motion of occlusion area is shown to correction.
In conjunction with the possible embodiment of the first of second aspect or second aspect, the 6th kind of possible enforcementIn mode, described initial motion model comprises affine model;
Described motion model estimation unit specifically for:
Calculate the parameter of affine transformation, obtain the affine model of matching initial motion;
Obtain the affine motion field of each two field picture according to described affine model;
Described affine motion field and initial optical flow field are merged;
Carry out the detection of occlusion area according to the uniformity of motion;
The affine motion that occlusion area is cut apart according to the place rectification of explicitly moving.
In conjunction with the 6th kind of possible embodiment of second aspect, in the 7th kind of possible embodiment, instituteState cluster cell specifically for:
To sorting all the cutting apart according to size in a two field picture, successively descending according to areaScreening affine model;
When cutting apart siArea be greater than preset area SizesOr the affine model collection { quantity of affine model in A}Be less than predetermined number NumATime, will cut apart siAffine model AiWith { A} compares, if { exist in A}A same A of affine model A'iEuclidean distance be less than predeterminable range ThA, give up current affine model Ai,Otherwise by AiJoin that { A} obtains least affine Models Sets { A}.
In conjunction with the 6th kind of possible embodiment of second aspect, in the 8th kind of possible embodiment, instituteState merge cells specifically for:
For each s of cutting apart, calculate the color distortion of cutting apart interior all pixels under the current affine model of cutting apartSummation;
Computed segmentation SiThe adjacent set { S of cutting apart of candidateNei, wherein, cut apart s ' completely for each candidate is adjacentFoot 2|s'| > | s|;
For { SNeiIn each adjacent s ' of cutting apart, calculate respectively the uniformity of moving on the same s ' of s bordersIn the uniformity of affine model of the same s ' of current motion of pixelPixel in s is at the affine model of s 'The ratio C of the color distortion of the lower colour-difference similarities and differences self affine models,s'
From { SNeiIn select and meet Ss,s'>0.5,Fs,s'>0.5,Cs,s'The adjacent of > 0.5 cut apart, by Cs,s'MaximumAdjacent cutting apart as segmentation candidates, be designated as s*, work as Cs,s*>θwTime, wherein, θwFor default ratio, willS merges in s*;
Each s of cutting apart is merged, if the last merged sum of cutting apart has exceeded one of initial segmentation numberHalf, re-start the estimation of affine model.
In the 9th kind of possible embodiment of second aspect, described initial light stream comprise forward direction light stream and/orOppositely light stream.
The beneficial effect that the technical scheme that the embodiment of the present invention provides is brought is:
The invention provides a kind of method for estimating and device, by by each two field picture in image sequenceIn all initial motion model of cutting apart carry out the sequential coupling of motion segmentation, to make same in image sequenceIn one object each frame in this image sequence, there is identical mark, and obtain according to the result of sequential couplingTo motion segmentation and output light stream. Adopt a kind of like this method for estimating can process longer image sequenceMotion segmentation and estimation, obtain space-time consistency good motion segmentation and motion estimation result, fromAnd can effectively improve the space-time consistency of motion segmentation.
Brief description of the drawings
In order to be illustrated more clearly in the technical scheme in the embodiment of the present invention, institute in describing embodiment belowNeed the accompanying drawing using to be briefly described, apparently, the accompanying drawing in the following describes is only the present inventionSome embodiment, for those of ordinary skill in the art, do not paying under the prerequisite of creative work,Can also obtain according to these accompanying drawings other accompanying drawing.
Fig. 1 is the method flow diagram of a kind of method for estimating of providing of the embodiment of the present invention;
Fig. 2 is the original sequence schematic diagram that the embodiment of the present invention provides;
Fig. 3 is the initial light stream schematic diagram of the image sequence that provides of the embodiment of the present invention;
Fig. 4 is that the image of the image sequence that provides of the embodiment of the present invention is cut apart schematic diagram;
Fig. 5 be the embodiment of the present invention provide obtain all initial motion model of cutting apart in each two field pictureThe method flow diagram of concrete steps;
Fig. 6 is the figure of the initial light stream that provides of the embodiment of the present invention through affine model parameter Estimation and after mergingThe light stream schematic diagram of picture sequence;
Fig. 7 is the schematic diagram that blocks of the image sequence that provides of the embodiment of the present invention;
Fig. 8 is the method flow diagram of another method for estimating of providing of the embodiment of the present invention;
Fig. 9 is the method flow diagram of the concrete steps optimized of the cluster that provides of the embodiment of the present invention;
Figure 10 is the method flow diagram of the concrete steps of the combination and segmentation that provides of the embodiment of the present invention;
Figure 11 be the embodiment of the present invention provide complete the image order of motion after the consistent merging of cutting apartList intention;
Figure 12 is the method flow diagram of the concrete steps of the motion segmentation sequential coupling that provides of the embodiment of the present invention;
Figure 13 is the method for the concrete steps optimized of the space-time consistency of the motion segmentation that provides of the embodiment of the present inventionFlow chart;
Figure 14 is the method for the concrete steps optimized of the space-time consistency of the estimation that provides of the embodiment of the present inventionFlow chart;
Figure 15 is the structural representation of a kind of movement estimation apparatus of providing of the embodiment of the present invention;
Figure 16 is the structural representation of another movement estimation apparatus of providing of the embodiment of the present invention.
Detailed description of the invention
For making the object, technical solutions and advantages of the present invention clearer, below in conjunction with accompanying drawing to the present inventionEmbodiment is described in further detail.
The method for estimating that the embodiment of the present invention provides can be for moving and estimate sequence of video imagesMeter, the motion estimation result of acquisition can destination object be cut apart, identify, be followed the tracks of for realizing andThe various application such as shape information recovery. Certainly more than, also only illustrate, carry out the fortune of sequence of video imagesThe result that moving estimation obtains can be widely used in various computer vision fields, the embodiment of the present invention pairThis does not limit.
The method for estimating that the embodiment of the present invention provides, as shown in Figure 1, the method flow process comprises:
Step 101, movement estimation apparatus obtain the initial light stream of each two field picture in the image sequence of input.
Step 102, movement estimation apparatus are divided at least one according to colouring information by each two field picture and cut apart.
Step 103, movement estimation apparatus obtain all initial motion model of cutting apart in each two field picture.
Step 106, movement estimation apparatus move all initial motion model of cutting apart in each two field pictureThe sequential coupling of cutting apart, has in same target in image sequence each frame in image sequence makingIdentical mark.
Step 107, movement estimation apparatus obtain motion segmentation and output light stream according to the result of sequential coupling.
The invention provides a kind of method for estimating, by owning in each two field picture in image sequenceThe initial motion model of cutting apart is carried out the sequential coupling of motion segmentation, to make the same target in image sequenceIn each frame in this image sequence, there is identical mark, and moved according to the result of sequential couplingCut apart and export light stream. Adopt a kind of like this method for estimating can process the motion of longer image sequenceCut apart and estimation, obtain good motion segmentation and the motion estimation result of space-time consistency, thus canEffectively improve the space-time consistency of motion segmentation.
Concrete, in actual process of carrying out estimation, for the input of obtaining described in step 101Image sequence in the process of initial light stream of each two field picture, can adopt existing light in prior artStream method of estimation is estimated to obtain to the image in image sequence. For example, can be in image sequence everyTwo frame adjacent images carry out estimation, owing to needing to carry out phase in follow-up space-time consistency optimizing processThe tracking of adjacent frame therefore can be carried out forward direction light stream (English: forwardflow) and/or reverse light stream (English simultaneouslyLiterary composition: estimation backwardflow). Taking the image sequence shown in Fig. 2 as example, initial forward direction light stream canWith as shown in Figure 3.
Described in step 102, according to colouring information, described each two field picture is divided into at least one cuts apartProcess can adopt various existing image partition methods equally, and the embodiment of the present invention does not limit this.Because this step is cut apart based on color of image information, the segmentation result therefore obtaining can keep and thingThe uniformity of body real border, same taking the image sequence shown in Fig. 2 as example, according to colouring information to oftenThe image that one frame carries out is cut apart can be as shown in Figure 4. Wherein, Fig. 4 is for adopting conventional images dividing methodThe figure obtaining is cut apart figure, and in the figure each is cut apart interior image and had identical colouring information, everyUnder individual pixel cut apart can mark with set { siIn a figure denote.
In step 103, can suppose that the motion of the pixel in each cutting apart is consistent, and meet certainForms of motion, the pixel within each cutting apart has motion uniformity. Concrete, in the invention processIn example, suppose that each forms of motion of cutting apart is affine transformation, regard each cutting apart as a plane of movement,The initial motion model of cutting apart interior point is affine model. Obtain all initial motions of cutting apart in each two field pictureThe concrete steps of model as shown in Figure 5, comprise the steps 1031 to 1035:
Step 1031, calculate the parameter of affine transformation, obtain the affine model of matching initial motion.
For example, can suppose that pixel X in the t frame homogeneous coordinates in image coordinate system are X=(x, y, 1),In t two field picture, pixel X is expressed as to the affine transformation matrix of t ' frame
A = a 1 a 2 b 1 a 3 a 4 b 2 0 0 1
In t, the position of pixel X in t ' is X '=AX, note X '=(x ', y ', 1), have
x'=a1·x+a2·y+b1
y'=a3·x+a4·y+b2
Note X is w=(u, v) to the motion of X ', u=x '-x, and v=y '-y, that is:
u=a1·x+a2·y+b1-x
v=a3·x+a4·y+b2-y
The light stream u obtaining according to initial calculation, v, can set up system of linear equations:
a1·x+a2·y+b1=u+x
a3·x+a4·y+b2=v+y
Owing to containing 6 variablees in A, but the motion of horizontal and vertical direction is independent, therefore needs 3 pointsCould estimate A. For the mistake that reduces to estimate, therefore counting of actual samples is greater than 3 points.In the time having n data point, the system of linear equations of acquisition is as follows:
a 1 · x 1 + a 2 · y 1 + b 1 = u 1 + x 1 a 3 · x 1 + a 4 · y 1 + b 2 = v 1 + y 1 a 1 · x 2 + a 2 · y 2 + b 1 = u 2 + x 2 a 3 · x 2 + a 4 · y 2 + b 2 = v 2 + y 2 . . . a 1 · x n + a 2 · y n + b 1 = u n + x n a 3 · x n + a 4 · y n + b 2 = v n + y n
Wherein, n represents the number of the sample point for carrying out parameter fitting, this system of linear equations solve employingLeast square method is carried out.
But it should be noted that, carry out parameter fitting if interior all points are cut apart in use, will increaseAdd time cost, can obviously not improve the accuracy of affine transformation simultaneously. Therefore can be to cutting apart inner pointFirst carry out sampling uniformly to obtain data point set, then adopt the consistent (English: Random of random sampleSampleConsensus, is called for short: mode RANSAC) is carried out the estimation of affine model.
Concrete, in the time that initial movement accuracy is relatively good, only need to choose a small amount of data just can be accurateAffine model is estimated, but in the time that initial movement accuracy is poor, just need to choose more pointEstimate. Therefore, the size of data point set is relevant with the accuracy of initial light stream, can be by initiallyThe accuracy of motion is estimated, and then the sample rate of specified data point set. First, calculate according to initial lightThe present frame t mid point x that stream obtainstWith its corresponding points x in consecutive frame t 't’Between color distortion d, thenCalculate in present frame average color difference a little, i.e. avg_d, and then count color distortion and be less thanThe number of the point of average avg_d, thus the accuracy rate of initial light stream obtainedAs r > 0.75 time, think that initial light stream is more accurately, sample rate is taken as 0.25, i.e. every four pixels choosingSelect one as data point; When 0.75>r>0.5 time, sample rate is taken as 0.5; In the time of r<0.5, sample rate is1.0。
After having determined sample rate, for the each pixel that may elect data point as, by this pixel place relativelyD compare with avg_d, determine want this pixel to be chosen to be data point. If its d is greater than in advanceIf threshold value, gets rid of this pixel, otherwise adds in set of data points. Wherein, predetermined threshold value canChoose according to actual needs, when the initial motion at pixel place is inaccurate or have while blocking its d valueBe greater than this predetermined threshold value.
The data point sampling in cutting apart is assigned to during corresponding this cut apart, thereby obtains each S of cutting aparti'sData point set DSi, next to each S of cutting apartiCarry out the estimation of affine model by RANSAC method.
Step 1032, obtain the affine motion field of each two field picture according to affine model.
Step 1033, affine motion field and initial optical flow field are merged.
For example, can use the pseudo-boolean of secondary to optimize (English: Quadratic with initial optical flow field affine motion fieldPseudo-BooleanOptimization, is called for short: QPBO) algorithm merges. Fig. 6 is initial light stream warpCross the light stream example after affine model parameter Estimation and fusion, after merging with affine motion, energyEnough initial light stream excessively level and smooth phenomenons on Moving Objects border of effectively improving.
Step 1034, according to motion uniformity carry out the detection of occlusion area.
The rectification of explicitly moving of step 1035, the affine motion that occlusion area is cut apart according to place.
In light stream is estimated, in the time that being blocked, the specific objective object in image sequence will cause this target pairThe tracking failure of elephant, occlusion issue is a relatively stubborn problem, existing most of optical flow approach are allImplicitly adopt robust function to process occlusion area, the result that this processing mode obtains is also inadequateGood. In embodiments of the present invention, after the detection of blocking according to the uniformity of motion, can be to blockingThe affine motion that region is cut apart according to the place rectification of explicitly moving.
For example,, for the some pixel x in t two field picturet, according to xtMotion, can obtain its underCorresponding points x ' in one frame (x '=xt+v(xt). For an x ', the motion according to its place frame to present frame,Can obtain the corresponding points x of x ' in present framet+1→t(xt+1→t=x '+v (x ')), if || xt-xt+1→t||>θfAnd s (xt)≠s(xt+1→t)(θfFor threshold values, represent the largest motion deviation of permission, as be made as 1 pixel),Think xtIn next frame, be blocked, be designated as Ot+1→t(xt)=1, otherwise Ot+1→t(xt)=0, representsxtBe not blocked. In like manner, if xtIn former frame, be blocked, correspondingly had Ot→t-1(xt)=1。
As shown in Figure 7, wherein white portion represents the part being blocked to the Occlusion Map O that detection obtains. According toThe Occlusion Map O that detection obtains explicitly blocks processing, if xtBe blocked, i.e. O (xt)=1,Make v (xt)=Axt-xt, use affine motion as xtMotion.
Optionally, as shown in Figure 8, before step 106, the estimation side that the embodiment of the present invention providesMethod can also comprise the steps at least one item in 104 and 105:
Step 104, movement estimation apparatus are according to all initial motion model of cutting apart in each two field picture, by toolThere is the initial motion model of same movement form to carry out cluster optimization.
Can find that by observing, in reality scene, forms of motion often has repeatability and unicity. MotionThe repeatability of form refers to that the forms of motion of different objects in Same Scene may be identical, forms of motionForms of motion that unicity refers to a large amount of objects in Same Scene minority often, although object in sceneMotion may be complicated, but same target is inner or the motion of several objects but meets same fortuneMoving form. Therefore, can extract the main movement form in scene, and then according to these main movement shapesFormula is carried out cluster optimization to the initial motion model of object in scene.
Concrete, after estimating through affine model, obtain a series of affine model, these affine modelsConventionally there is very large redundancy, be therefore necessary to carry out the cluster of model, obtain a least affine modelCollection { A}. Concrete, as shown in Figure 9, step 104 comprises the steps 1041 to 1042:
Step 1041, to sorting all the cutting apart according to size in a two field picture, according to area byLarge to the little affine model that screens successively.
Step 1042, when cutting apart SiArea be greater than preset area SizesOr affine model collection is { affine in A}The quantity of model is less than predetermined number NumATime, will cut apart SiAffine model AiWith A} compares, if{ in A}, there is a same A of affine model A 'iEuclidean distance be less than predeterminable range ThA, give up current imitativePenetrate model Ai, otherwise by AiJoin that { A} obtains least affine Models Sets { A}.
Wherein, preset area SizesRepresent one and cut apart big or small threshold value, this value can be entered according to actual needsRow is chosen, as Sizes=W*H*0.01. Predetermined number NumARepresent the lower limit of affine model number, according toConcrete scene is set, as is made as 10. Predeterminable range ThAFor default Euclidean distance, as be made as 0.01. ThisBright embodiment does not limit the value of preset value.
Further, by { in A}, the same mark value l of each affine model is corresponding, adds a volume simultaneouslyOuter mark is used for representing initial motion, and therefore the number of mark is Nl=| { next A}|+1 solves about fortuneMoving multiple labeling problem, as l < Nl-1 o'clock, obtain the motion of each point according to affine model, work as l=Nl-1Time, obtain the motion of each point according to initial light stream.
Target equation is:
E ( x , l ) = &Sigma; x ( &lambda; Data w O ( x ) d 2 d 2 + &sigma; C 2 + &lambda; Smooth &Sigma; y &Element; N ( x ) w s ( x , y ) [ l ( x ) &NotEqual; l ( y ) ] )
Wherein, wo(x) represent to block weights, w O ( x ) = 0.01 O ( x ) = 1 1 O ( x ) = 0 ; D is the Europe of rgb spaceFormula distance, d=||I'(x')-I (x) ||; ws(x, y) is level and smooth coefficient, cuts apart acquisition according to current, w s ( x , y ) = 1 s ( x ) = s ( y ) 0 Otherwise , N (x) represents the neighborhood space of x, λDataSmoothBe respectively data item peaceThe weights of sliding.
Use image cutting (English: Graphcut) method to solve above-mentioned energy equation minimization problem, and rootAccording to the solution obtaining, optical flow field is upgraded.
Through the combined optimization of too much affine model, and move after correction according to blocking estimated result, canEffectively correct the motion of occlusion area.
Consider the motion uniformity of cutting apart, because the initial segmentation obtaining according to clustering algorithm is in sequentialThe poor over-segmentation (English: oversegmentation) of uniformity, and be to obtain according to color and spaceCut apart, instead of according to motion. Therefore, be necessary initial segmentation to close according to the consistent situation of motionAnd, thereby obtain consistent the cutting apart of moving. On the one hand, cut apart through can effectively reducing after mergingNumber, also reduced the number of mark, reduced space-time in follow-up space-time consistency optimizing process multipleAssorted degree; Secondly,, after the consistent merging of motion, it is certain that the motion segmentation obtaining has kept in sequentialUniformity, the convergence rate of the follow-up optimization of quickening that can be effectively.
And/or before step 106, the method for estimating that the embodiment of the present invention provides can also comprise:
Step 105, movement estimation apparatus merge consistent the cutting apart of moving in each two field picture.
In order to improve the uniformity of initial segmentation in sequential, be necessary initial segmentation according to the consistent feelings of motionCondition merges, thereby obtains consistent the cutting apart of moving. On the one hand, after merging, can effectively subtractThe number of cutting apart less, also reduced the number of mark, reduced in follow-up space-time consistency optimizing processSpace-time Complexity; On the other hand, after the merging of cutting apart through motion uniformity, the motion obtainingBe segmented in and in sequential, kept certain uniformity, thus the convergence speed of the follow-up optimization of quickening that can be effectivelyDegree.
Concrete, as shown in figure 10, step 105 comprises the steps 1051 to 1055:
Step 1051, for each s of cutting apart, calculate and under the current affine model of cutting apart, cut apart interior all pixelsColor distortion summation.
Step 1052, computed segmentation SiThe adjacent set { S of cutting apart of candidateNei, wherein, for each candidate's phaseNeighbour is cut apart s ' and is met 2|s'| > | s|.
Step 1053, for { SNeiIn each adjacent s ' of cutting apart, calculate respectively on the same s ' of s border, move oneCausing property Ss,s', the uniformity F of the affine model of the same s ' of current motion of the pixel in ss,s', the pixel in s is at s 'Affine model under the ratio C of color distortion of the colour-difference similarities and differences self affine models,s'
Wherein, SS,S’,FS,S’,CS,S’Account form is respectively:
S s , s &prime; = &Sigma; x &Element; s , y &Element; s &prime; , | x - y | = 1 ( | | v ( x ) - v ( y ) | | < Th Smooth ) &Sigma; x &Element; s , y &Element; s &prime; , | x - y | = 1 1
F s , s &prime; = &Sigma; x &Element; s ( | | v ( x ) - ( A s &prime; x - x ) | | < Th Flow ) &Sigma; x &Element; s ( 1 )
C s , s &prime; = &Sigma; x &Element; s | | I &prime; ( A s x ) - I ( x ) | | &Sigma; x &Element; s | | I &prime; ( A s &prime; x ) - I ( x ) | |
Step 1054, from { SNeiIn select and meet Ss,s'>0.5,Fs,s'>0.5,Cs,s'The adjacent of > 0.5 cut apart,By Cs,s'Maximum adjacent cutting apart as segmentation candidates, be designated as s*, whenTime, wherein, θwFor defaultRatio, merges to s by s*In.
Wherein, θwCan choose according to actual conditions, if value is 0.9, the embodiment of the present invention is not done thisLimit.
Step 1055, each s of cutting apart is merged, if at the beginning of the last merged sum of cutting apart exceededThe half of beginning Segmentation Number, re-starts the estimation of affine model.
By iteration above-mentioned steps, can be merged until do not cut apart, complete motion and there is uniformityThe merging of cutting apart. Example results after merging as shown in figure 11. As can be seen from the results, through mergingAfterwards, be incorporated in together multiple the cutting apart that belongs to same forms of motion.
Further, cutting apart while optimizing with the space-time consistency of estimation, in order to ensure sameIn the different frame of object on video sequence, have same mark, need to be mated cutting apart of different frame,As shown in figure 12, step 106 specifically comprises the steps 1061 to 1068:
Step 1061, using input image sequence in the first two field picture as initial matching masterplate, it is eachCut apart as a data sectional (English: volumesegment), for data sectional described in each distributesUnique mark.
Step 1062, for cutting apart in t frameLight stream according to t frame to t-1 frame, follows the tracks ofInEach pixel xtCorresponding points x in t-1 framet-1, wherein, xt-1In t-1 frame, be designated as cutting apart of place?The set formingExpression is cut apartCorresponding likely cutting apart in t-1 frame.
Step 1063, forIn each cutting apartCalculateWithMatching rate:
r t - > t - 1 k - > k &prime; = | x t | x t &Element; s t k , x t - 1 &Element; s t - 1 k &prime; | | s t - 1 k &prime; |
Wherein, | xt| for cutting apartIn with cut apartThe number of corresponding pixel, denominatorExpression is cut apartInThe number of point.
Step 1064, calculatingWithMatching rate:
r t - 1 - > t k &prime; - > k = c | s t k |
Wherein, rightIn each cutting apartLight stream according to t-1 frame to t frame, obtainsIn eachPoint xt-1Corresponding points x in tt, statistics xtBelong toThe number of point be designated as c,For cutting apartMid pointNumber.
Step 1065, calculatingWithFinal matching rate:
r t , t - 1 k , k &prime; = min ( r t - > t - 1 k - > k &prime; , r t - 1 - > t k &prime; - > k ) .
Step 1066, whenTime, determineWithCoupling, willFlag update beMarkNote, and willJoinIn the data sectional at place, wherein, σmFor default matching threshold.
Concrete, ifσmFor predefined coupling threshold values, as being made as 0.5, recognizeForWithMate, or also can beIn find matching rateMaximum cuts apartWithMate, now maximum matching rateAll generally to be greater than σm, and matching rate is greater than σmPointCutIn general be all unique, existCan only find at the most one cuts apartWithCouplingRate is greater than σm, willFlag update beMark, and willJoinThe data sectional at placeIn.
Step 1067, whenWhile not existing coupling to cut apart in t-1 frame, willWith carrying out cutting apart in t-2 frameCoupling.
, for not there is not coupling point in step 1068, the coupling that finishes in the time that matching times is greater than default coupling step-lengthCutDistribute a new mark, willJoin in new data sectional.
Wherein, owing to following the tracks of the constantly accumulation of time stream error, therefore default coupling step-length is commonGet a less value, if value is 5, the embodiment of the present invention does not limit this.
After overmatching, obtain some data sectionals, be designated as S={Sk| k=1,2..., K}, each data are dividedSection SkIn corresponding different frame, belong to cutting apart of same object, and use unique mark l to represent, therefore asPixel x in fruit t frametCorresponding pixel x in same t ' framet’Belong to same data sectional, so xtAnd xt’Have identical mark, i.e. S (xt)=S(xt')。
After the sequential coupling of over-segmentation, obtain some data sectionals, being divided in each data sectionalConsistent cutting apart in sequential in each frame, meanwhile, it is consistent that these motions in cutting apart also roughly meet space-timeProperty. Next can carry out respectively according to the result of sequential coupling space-time consistency optimization and the fortune of motion segmentationThe moving space-time consistency optimization of estimating, obtains the consistent motion segmentation of space-time and output light stream.
Wherein, for the space-time consistency optimization of motion segmentation, concrete, based on cutting apart and base of single framesWhat obtain in sequential consistent boundary cuts apart, and due to some accidentalia, causes the segmentation result obtaining at space-timeOn still consistent not. For in a certain frame t in video sequence, occur two different objects due to color similarity alsoAnd space is adjacent and lead to errors and cut apart this situation by cluster to same, apart from far away its of this frameIn he frame t ', due to the difference of object motion, object position is spatially no longer adjacent, thereby in t ',These two objects can be correctly divided. But these two objects are still very near the frame part of this frameMay be cut apart mistakenly, if therefore just the processing of the multiframe based on local still cannot correctly correctThis segmentation errors.
In the space-time consistency Optimum Operation of motion segmentation, can utilize in sequential correctly carve information to goCorrect the erroneous segmentation in those local frames, correctly carve information is promoted globally, instead of entersThe multi-frame processing that row is local, thereby can obtain upper consistent the cutting apart of space-time of the overall situation. For in video sequenceEach pixel x in each frame tt, followed the tracks of in other frames according to light stream, obtain corresponding points xt’. If pointIt is correct cutting, so xtWith xt’Should there is identical data sectional mark, i.e. S (xt)=S(xt’). Therefore, systemCount all corresponding points xt’Data sectional mark, obtain a tag set P (xt), x sotMark shouldWhen being set P (xt) the middle maximum mark l of occurrence number. Said process can form be turned to following energy sideJourney:
E ( l ) = &Sigma; t = 1 n ( &Sigma; x t ( E d ( x t , l ) + &Sigma; y t &Element; N ( x t ) E s ( l ( x t ) , l ( y t ) ) ) )
Wherein, Ed(l,xt) be data item constraint, Es(l(xt),l(yt)) represent level and smooth item constraint, N (xt) expression xt'sNeighbor. Ed(S(xt)) be defined as follows:
Ed(l,xt)=-wplogLp(l(xt),xt)-wclogLc(l(xt),xt)-wslogLs(l(xt),xt)+wflog(1+Df(l(xt),xt))
Wherein, Lp(l(xt),xt) be xtBelong to the marking probability of l, Lc(l(xt),xt) expression xtWith data corresponding to lThe color similarity of segmentation mid point, Ls(l(xt),xt) expression xtWith being correlated with of the locus of cutting apart mid point corresponding to lProperty, Df(l(xt),xt) represent the uniformity of the affine model that current motion cuts apart with l correspondence, wp,wc,ws,wfBe respectively marking probability, color similarity, spatial coherence, affine conforming coefficient.
L p ( l , x t ) = 1 n p ( &Sigma; t &prime; [ S ( x t &prime; ) = l ] + [ S ( x t ) = 1 ] ) , l &Element; P ( x t )
Wherein, npFor the number of the available point that traces into (comprises xt). Known, Lp(l,xt) larger, show xtThe probability that belongs to the data sectional that mark l is corresponding is larger.
Suppose that distribution of color and the spatial distribution cut apart meet Gaussian distribution N (), therefore color similarity Lc(l(xt),xt) be defined as:
L c ( l ( x t ) , x t ) = N ( I ( x t ) | &mu; S l c , &Sigma; S l c )
WithRepresent respectively the data sectional S that l is correspondinglThe average of middle distribution of color and covariance. SimilarGround, Ls(l(xt),xt) be defined as:
L s ( l ( x t ) , x t ) = 1 | f ( S l ) | &Sigma; t &prime; &Element; f ( S l ) N ( x t | &eta; s l , &Delta; s l )
Wherein, f (Sl) expression SlInterior all frame index that are segmented in video sequence, therefore Ls(l(xt),xt) meterThat calculate is xtAnd corresponding points xt’At SlSpatial coherence in inside cutting apart separately average. SlRepresent dividing bodyThe S of unitlCorresponding cutting apart in frame t ',WithRepresent respectively SlAverage and the covariance of interior space of points position.
Affine uniformity Df(l(xt),xt) be defined as:
D f ( l , x t ) = | | ( A l t x t - x t ) - v x t | |
From defining above, marking probability is larger, color similarity is larger, and spatial coherence is higher, imitativePenetrate uniformity better, the value at cost of data item is less. Coefficient wpWith respect to wcAnd ws, value is larger,And wcAnd wsValue is less, wfAccording to the accuracy setting of initial motion, therefore approximately intrafascicular of data item is mainStill the marking probability constraint of effect of contraction, color similarity and spatial coherence are just used for carrying out complementaryConstraint.
Es(l(xt),l(yt)) be used for carrying out regularization (smoothly) constraint, guarantee that on space, the mark of neighbor is consistent.Es(l(xt),l(yt)) be defined as anisotropy, as follows:
E s ( l ( x t ) , l ( y t ) ) = ( &lambda; l &epsiv; l p l ( x t , y t ) + &epsiv; l + &lambda; c &epsiv; c | | I ( x t ) - I ( y t ) | | + &epsiv; c ) [ l ( x t ) &NotEqual; l ( y t ) ]
Wherein,For the weights of anisotropic level and smooth, pl(xt,yt)Represent xtAnd ytThe inconsistent probability of mark, || I (xt)-I(yt)) || represent xtAnd ytIn the European distance of color spaceFrom, λlAnd λcRepresent coefficient, εlAnd εcFor sensitivity coefficient, control the span of weights. pl(xt,yt) definitionFor:
p l ( x t , y t ) = 1 n l ( &Sigma; t &prime; ( [ s ( x t &prime; ) &NotEqual; s ( y t &prime; ) ] ) + [ s ( x t ) &NotEqual; s ( y t ) ] )
Known, work as xtAnd ytDo not cut apart or when color distortion is larger, the weights of level and smooth are less same;Otherwise weights are larger.
Due to the restriction of time and memory headroom cost, directly E (l) is solved very difficult, can iterationGround solves E (t, l) successively, simultaneously in order to improve the efficiency of processing, can only choose the N of present frame ttIndividualConsecutive frame is followed the tracks of, for each pixel xt, equally also choose the N of cost minimizationlIndividual mark is optimized.
Concrete, as shown in figure 13, the space-time consistency optimization method of motion segmentation is as follows, comprises following stepRapid 10711 to 10715:
Step 10711, the dividing mark of fixing each two field picture are constant.
Wherein, first calculate average and the covariance and each of the Gaussian distribution of color in each data sectionalIn data sectional, each is segmented in average and the covariance of the Gaussian distribution of the locus in frame separately, connectsFrom t=1 frame, carry out the fixing of dividing mark for each two field picture.
Step 10712, according to light stream, follow the tracks of each pixel x in t frametAt adjacent NtCorrespondence in framePoint, obtains candidate's tag set P (xt)。
Step 10713, according to candidate's tag set P (xt) obtain the mark that occurrence number is maximum.
Concrete, calculating P (xt) the middle maximum N of occurrence numberlAfter individual mark, can calculate this NlThe marking probability of individual mark, and xtEach mark corresponding cut apart in spatial similarity and space correlationProperty, thereby obtain data item cost.
Further, according to the definition of level and smooth item constraint, calculate the weights of the level and smooth item constraint of each some xt.
Use the efficient belief propagation of early vision (English: EfficientBeliefPropagationforEarlyVision, is called for short: BP) method solves energy equation E (t, l). Upgrade present frame t's according to the result obtainingDividing mark.
After one takes turns and optimized, add up the quantity C of cutting apart comprising in each data sectionalsAnd this numberAccording to the number C of all cut-points in segmentationpt
Step 10714, work as Cs<Nt/ 2 or CPt<σPtTime, wherein, CsFor comprising in each data sectionalThe quantity of cutting apart, CPtFor the number of all cut-points in data sectional, σPtFor predetermined threshold value, for numberAccording to each the cutting apart in segmentation, according to the mark situation of the corresponding points in other frames of the point in data sectional, choosingSelecting the data sectional of an optimum Match forces to merge.
Through after aforesaid operations, the data sectional that some mark l are corresponding has not existed.
Step 10715, upgrade the mark of remaining data sectional.
Similar with the space-time consistency optimization of motion segmentation, for the space-time consistency optimization of estimation,For each mark l, in corresponding each frame, belong to each affine transformation mould of cutting apart of same data sectionalTypeBecause cutting apart in same data sectionalMotion might not be constant in sequential, stillIn each frameInterior some forms of motion is consistent substantially,The motion of interior point all meetsRightEach pixel x in t frametIf, xtMotion meet affine Transform Model at present frame tSoxtCorresponding points x in other frames t 't’Motion should meet affine Transform ModelCan use xtWith its correspondencePoint xt’Measure the accuracy of motion at the Euclidean distance of color space, therefore, target is exactly to be each pixelChoose a best mark l*, make xtL coexists*Corresponding affine transformationUnder all corresponding points of obtainingxt’Between Euclidean distance and minimum, keep the mark of Local Phase neighbouring region pixel consistent simultaneously, formalization is above-mentionedDescription obtains about the energy equation of motion optimization as follows:
E ( l ) = &Sigma; t = 1 n ( &Sigma; x t ( &sigma; c * E c ( x t , l ) + &sigma; l * E l ( x t , l ) + &sigma; f * E f ( x t , f ) + &sigma; s * E s ( x t , l ) ) )
Wherein, EcRepresent xtThe corresponding points that obtain under mark l of coexisting are at the Euclidean distance of rgb space, ElRightxtThe marking probability of the corresponding points that obtain under mark l is punished, EfRepresent xtIn present frame under mark lAffine motion with the difference of current motion, EsFor smoothness constraint, σc,σl,σf,σsFor the power of corresponding entryValue.
E c ( x t , l ) = &Sigma; t &prime; ( w O ( x t ) d 2 d 2 + &theta; C 2 ) , d = | | I t &prime; ( x t &prime; ) - I t ( x t ) | |
Wherein, xt’For xtCorresponding points under affine transformation corresponding to mark l in other frames t ', by continuousAffine transformation obtains,t1,t2,tkFor t is to the intermediate frame of t '.
E l ( x t , l ) = - log ( &Sigma; t &prime; ( S t &prime; ( x t &prime; ) = = l ) N )
Wherein, St'(xt') expression xtCorresponding points x in t ' framet’The dividing mark at place.
E f ( x t , l ) = | | ( A t l x t - x t ) - v ( x t ) | |
For current some xtThe motion obtaining under affine model corresponding to mark l, v (xt) expression xt'sCurrent motion.
E s ( x t , l ) = &Sigma; y t &Element; N ( x t ) w s ( x , y ) ( l ( x t ) = = l ( y t ) ? 0 : &theta; Smooth )
ws(x, y) is xt,ytBetween the weights of level and smooth item constraint, cut apart acquisition according to current, be used for strengthening motionThe uniformity of continuum, allows noncoherent boundary mark value difference. θSmoothFor consecutive points mark inconsistentPenalty value.
Space-time consistency optimization with motion segmentation is the same, and this energy equation of direct solution is very difficult, because ofThis carries out iterative, chooses Nt consecutive frame t ' simultaneously.
Concrete, as shown in figure 14, the space-time consistency optimization of estimation comprises, comprises the steps 10721To 10725:
Step 10721, according to current motion, obtain the data sectional that each mark l represents right in each frameThe affine Transform Model of cutting apart of answering
Step 10722, for t frame, according to affine Transform ModelCalculate light stream, to make tEach pixel x in frametCorresponding points x in other frames t 't’In the Euclidean distance sum minimum of color space.
Concrete, from t=1 frame, for each frame t, the data item E of calculating energy equationc、El、Ef,Further, calculate the coefficient w of level and smooth item constraints(x, y), carries out ENERGY METHOD by image cutting methodSolve.
The light stream of step 10723, renewal t frame.
Step 10724, carry out the detection of occlusion area.
Step 10725, the motion of occlusion area is shown to correction.
The invention provides a kind of method for estimating, by owning in each two field picture in image sequenceThe initial motion model of cutting apart is carried out the sequential coupling of motion segmentation, to make the same target in image sequenceIn each frame in this image sequence, there is identical mark, and moved according to the result of sequential couplingCut apart and export light stream. Adopt a kind of like this method for estimating can process the motion of longer image sequenceCut apart and estimation, obtain good motion segmentation and the motion estimation result of space-time consistency, thus canEffectively improve the space-time consistency of motion segmentation.
The method for estimating that adopts the embodiment of the present invention to provide, can improve color phase in image effectivelyLike, block and without the motion estimation result of texture region; In addition, sports ground keeps good structure, is dividingThe inside of cutting, motion is level and smooth and consistent, on the border of cutting apart, is also often the border of object in scene,Motion has kept good motion discontinuity.
It should be noted that, the core of technical solution of the present invention is to have in video sequence based on objectCut apart and the feature of the space-time consistency of estimation, effectively utilize cutting apart in different frame in video sequenceWith movable information correct in a certain frame causing due to accidentalia cut apart and move occur wrongSituation. Similarly, can use this technical scheme to carry out the depth recovery of video sequence, to obtain space-time oneThe picture depth information causing. In addition, apply method proposed by the invention, carry out the motion of video sequence and estimateThe result that obtains of meter can be widely applied to video interpolation, to the computer vision such as image tracing, three-dimensional reconstructionIn field, differ and one illustrate at this, but all should include protection scope of the present invention in.
The embodiment of the present invention also provides a kind of movement estimation apparatus, and as shown in figure 15, this device comprises:
Acquiring unit 1, for obtaining the initial light stream of each two field picture of image sequence of input.
Cutting unit 2, cuts apart for described each two field picture being divided into at least one according to colouring information.
Motion model estimation unit 3, for obtaining all initial motion model of cutting apart of each two field picture.
Sequential matching unit 6, for moving all each two field picture initial motion model of cutting apart pointThe sequential coupling of cutting, to make having phase in same target in image sequence each frame in image sequenceSame mark.
Processing unit 7, for obtaining motion segmentation and output light stream according to the result of sequential coupling.
The invention provides a kind of movement estimation apparatus, by owning in each two field picture in image sequenceThe initial motion model of cutting apart is carried out the sequential coupling of motion segmentation, to make the same target in image sequenceIn each frame in this image sequence, there is identical mark, and moved according to the result of sequential couplingCut apart and export light stream. Adopt a kind of like this method for estimating can process the motion of longer image sequenceCut apart and estimation, obtain good motion segmentation and the motion estimation result of space-time consistency, thus canEffectively improve the space-time consistency of motion segmentation.
Wherein, motion model estimation unit 3 specifically can be for:
Calculate the parameter of affine transformation, obtain the affine model of matching initial motion.
Obtain the affine motion field of each two field picture according to affine model.
Affine motion field and initial optical flow field are merged.
Carry out the detection of occlusion area according to the uniformity of motion.
The affine motion that occlusion area is cut apart according to the place rectification of explicitly moving.
Optionally, as shown in figure 16, described device can also comprise:
Cluster cell 4, for according to all initial motion model of cutting apart of each two field picture, will have identicalThe initial motion model of forms of motion is carried out cluster optimization.
Concrete, cluster cell 6 can be for:
To sorting all the cutting apart according to size in a two field picture, successively descending according to areaScreening affine model.
When cutting apart siArea be greater than preset area SizesOr the affine model collection { quantity of affine model in A}Be less than predetermined number NumATime, will cut apart siAffine model AiWith { A} compares, if { exist in A}A same A of affine model A'iEuclidean distance be less than predeterminable range ThA, give up current affine model Ai,Otherwise by AiJoin that { A} obtains least affine Models Sets { A}.
And/or as shown in figure 16, described device can also comprise:
Merge cells 5, for merging consistent the cutting apart of moving of each two field picture.
Concrete, merge cells 5 can be for:
For each s of cutting apart, calculate the color distortion of cutting apart interior all pixels under the current affine model of cutting apartSummation.
Computed segmentation SiThe adjacent set { S of cutting apart of candidateNei, wherein, cut apart s ' completely for each candidate is adjacentFoot 2|s'| > | s|.
For { SNeiIn each adjacent s ' of cutting apart, calculate respectively the uniformity of moving on the same s ' of s bordersIn the uniformity of affine model of the same s ' of current motion of pixelPixel in s is at the affine model of s 'The ratio C of the color distortion of the lower colour-difference similarities and differences self affine models,s'
Wherein, SS,S’,FS,S’,CS,S’Account form is respectively:
S s , s &prime; = &Sigma; x &Element; s , y &Element; s &prime; , | x - y | = 1 ( | | v ( x ) - v ( y ) | | < Th Smooth ) &Sigma; x &Element; s , y &Element; s &prime; , | x - y | = 1 1
F s , s &prime; = &Sigma; x &Element; s ( | | v ( x ) - ( A s &prime; x - x ) | | < Th Flow ) &Sigma; x &Element; s ( 1 )
C s , s &prime; = &Sigma; x &Element; s | | I &prime; ( A s x ) - I ( x ) | | &Sigma; x &Element; s | | I &prime; ( A s &prime; x ) - I ( x ) | |
From { SNeiIn select and meet Ss,s'>0.5,Fs,s'>0.5,Cs,s'The adjacent of > 0.5 cut apart, by Cs,s'MaximumAdjacent cutting apart as segmentation candidates, be designated as s*, work as Cs,s*>θwTime, wherein, θwFor default ratio, willS merges in s*.
Wherein, θwCan choose according to actual conditions, if value is 0.9, the embodiment of the present invention is not done thisLimit.
Each s of cutting apart is merged, if the last merged sum of cutting apart has exceeded one of initial segmentation numberHalf, re-start the estimation of affine model.
Further, cutting apart while optimizing with the space-time consistency of estimation, in order to ensure sameIn the different frame of object on video sequence, have same mark, need to be mated cutting apart of different frame,Sequential matching unit 6 specifically can be for:
Using the first two field picture in the image sequence of input as initial matching masterplate, its each cutting apart as oneIndividual data sectional, for data sectional described in each distributes unique mark.
For cutting apart in t frameLight stream according to t frame to t-1 frame, follows the tracks ofIn each pixel xtCorresponding points x in t-1 framet-1, wherein, xt-1In t-1 frame, be designated as cutting apart of place?StructureThe set becomingExpression is cut apartCorresponding likely cutting apart in t-1 frame.
ForIn each cutting apartCalculateWithMatching rate:
r t - > t - 1 k - > k &prime; = | x t | x t &Element; s t k , x t - 1 &Element; s t - 1 k &prime; | | s t - 1 k &prime; |
Wherein, ︱ xt︱ is for cutting apartIn with cut apartThe number of corresponding pixel, denominator represents to cut apartThe number of mid point.
CalculateWithMatching rate:
r t - 1 - > t k &prime; - > k = c | s t k |
Wherein, rightIn each cutting apartLight stream according to t-1 frame to t frame, obtainsIn eachPoint xt-1Corresponding points x in tt, statistics xtBelong toThe number of point be designated as c,For cutting apartInThe number of point.
CalculateWithFinal matching rate:
r t , t - 1 k , k &prime; = min ( r t - > t - 1 k - > k &prime; , r t - 1 - > t k &prime; - > k ) .
WhenTime, determineWithCoupling, willFlag update beMark, and willJoinIn the data sectional at place, wherein, σmFor default matching threshold.
Concrete, ifσmFor predefined coupling threshold values, as being made as 0.5, recognizeForWithMate, or also can beIn find matching rateMaximum cuts apartWithMate, now maximum matching rateAll generally to be greater than σm, and matching rate is greater than σmPointCutIn general be all unique, existCan only find at the most one cuts apartWithCouplingRate is greater than σm, willFlag update beMark, and willJoinThe data sectional at placeIn.
WhenWhile not existing coupling to cut apart in t-1 frame, willWith mating cutting apart in t-2 frame.
In the time that being greater than default coupling step-length, matching times finishes coupling, for not existing coupling to cut apartDistributeA new mark, willJoin in new data sectional.
Wherein, owing to following the tracks of the constantly accumulation of time stream error, therefore default coupling step-length is commonGet a less value, if value is 5, the embodiment of the present invention does not limit this.
After overmatching, obtain some data sectionals, be designated as S={Sk| k=1,2..., K}, each data are dividedSection SkIn corresponding different frame, belong to cutting apart of same object, and use unique mark l to represent, therefore asPixel x in fruit t frametCorresponding pixel x in same t ' framet’Belong to same data sectional, so xtAnd xt’Have identical mark, i.e. S (xt)=S(xt')。
After the sequential coupling of over-segmentation, obtain some data sectionals, being divided in each data sectionalConsistent cutting apart in sequential in each frame, meanwhile, it is consistent that these motions in cutting apart also roughly meet space-timeProperty. Next can carry out respectively according to the result of sequential coupling space-time consistency optimization and the fortune of motion segmentationThe moving space-time consistency optimization of estimating, obtains the consistent motion segmentation of space-time and output light stream.
Concrete, as shown in figure 16, processing unit 7 comprises: motion segmentation is optimized module 71 and motion is estimatedMeter is optimized module 72.
Motion segmentation is optimized module 71, and the space-time that carries out motion segmentation for the result of mating according to sequential is consistentProperty is optimized.
Concrete, fortune motion segmentation is optimized module 71 can be for:
The dividing mark of fixing each two field picture is constant.
According to light stream, follow the tracks of each pixel x in t frametAt adjacent NtCorresponding points in frame, are waitedSelect tag set P (xt)。
According to candidate's tag set P (xt) obtain the mark that occurrence number is maximum.
Work as Cs<Nt/ 2 or CPt<σPtTime, wherein, CsFor the number of cutting apart comprising in each data sectionalAmount, CPtFor the number of all cut-points in described data sectional, σPtFor predetermined threshold value, for described dataEach cutting apart in segmentation, according to the mark situation of the corresponding points in other frames of the point in described data sectional,Select the data sectional of an optimum Match to force to merge.
Upgrade the mark of remaining data sectional.
Estimation is optimized module 72, and the space-time that carries out estimation for the result of mating according to sequential is consistentProperty is optimized.
Concrete, estimation is optimized module 72 can be for:
According to current motion, obtain data sectional that each mark l represents corresponding cut apart imitative in each framePenetrate transformation model
For t frame, according to affine Transform ModelCalculate light stream, to make each in t framePixel xtCorresponding points x in other frames t 't’In the Euclidean distance sum minimum of color space.
Upgrade the light stream of t frame.
Carry out the detection of occlusion area.
The motion of occlusion area is shown to correction.
The invention provides a kind of movement estimation apparatus, by owning in each two field picture in image sequenceThe initial motion model of cutting apart is carried out the sequential coupling of motion segmentation, to make the same target in image sequenceIn each frame in this image sequence, there is identical mark, and moved according to the result of sequential couplingCut apart and export light stream. Adopt a kind of like this method for estimating can process the motion of longer image sequenceCut apart and estimation, obtain good motion segmentation and the motion estimation result of space-time consistency, thus canEffectively improve the space-time consistency of motion segmentation.
The movement estimation apparatus that adopts the embodiment of the present invention to provide, can improve color phase in image effectivelyLike, block and without the motion estimation result of texture region; In addition, sports ground keeps good structure, is dividingThe inside of cutting, motion is level and smooth and consistent, on the border of cutting apart, is also often the border of object in scene,Motion has kept good motion discontinuity.
It should be noted that, the core of technical solution of the present invention is to have in video sequence based on objectCut apart and the feature of the space-time consistency of estimation, effectively utilize cutting apart in different frame in video sequenceWith movable information correct in a certain frame causing due to accidentalia cut apart and move occur wrongSituation. Similarly, can use this technical scheme to carry out the depth recovery of video sequence, to obtain space-time oneThe picture depth information causing. In addition, apply method proposed by the invention, carry out the motion of video sequence and estimateThe result that obtains of meter can be widely applied to video interpolation, to the computer vision such as image tracing, three-dimensional reconstructionIn field, differ and one illustrate at this, but all should include protection scope of the present invention in.
The movement estimation apparatus that above-described embodiment provides is in the time carrying out estimation, only with above-mentioned each functional moduleDivision be illustrated, in practical application, can as required above-mentioned functions be distributed by differentFunctional module completes, and is divided into different functional modules by the internal structure of device, to complete above descriptionAll or part of function. In addition, the movement estimation apparatus that above-described embodiment provides and previous embodiment instituteThe method for estimating embodiment stating belongs to same design, and its specific implementation process refers to embodiment of the method, thisIn repeat no more.
One of ordinary skill in the art will appreciate that all or part of step that realizes above-described embodiment can pass throughHardware completes, and also can carry out the hardware that instruction is relevant by program and complete, and described program can be stored inIn a kind of computer-readable recording medium, the above-mentioned storage medium of mentioning can be read-only storage, disk orCD etc.
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, all of the present inventionWithin spirit and principle, any amendment of doing, be equal to replacement, improvement etc., all should be included in of the present inventionWithin protection domain.

Claims (20)

1. a method for estimating, is characterized in that, described method comprises:
Obtain the initial light stream of each two field picture in the image sequence of input;
According to colouring information, described each two field picture being divided into at least one cuts apart;
Obtain all initial motion model of cutting apart in described each two field picture;
All initial motion model of cutting apart in described each two field picture are carried out to the sequential coupling of motion segmentation,To make thering is identical mark in same target in described image sequence each frame in described image sequenceNote;
Obtain motion segmentation and output light stream according to the result of sequential coupling.
2. method according to claim 1, is characterized in that, described according to the institute of each two field pictureBefore the sequential that has the initial motion model of cutting apart to carry out motion segmentation is mated, described method also comprises:
According to all initial motion model of cutting apart in described each two field picture, same movement form will be there isInitial motion model is carried out cluster optimization; And/or
Merge consistent the cutting apart of moving in described each two field picture.
3. method according to claim 1 and 2, is characterized in that, described by described each two field pictureIn all initial motion model of cutting apart sequential coupling of carrying out motion segmentation comprise:
Using the first two field picture in the image sequence of input as initial matching masterplate, its each cutting apart as oneIndividual data sectional, for data sectional described in each distributes unique mark;
For cutting apart in t frameLight stream according to t frame to t-1 frame, follows the tracks ofIn each pixel xtCorresponding points x in t-1 framet-1, wherein, xt-1In t-1 frame, be designated as cutting apart of place?StructureThe set becomingExpression is cut apartCorresponding likely cutting apart in t-1 frame;
ForIn each cutting apartCalculateWithMatching rate:
r t - > t - 1 k - > k &prime; = | x t | x t &Element; s t k , x t - 1 &Element; s t - 1 k &prime; | | s t - 1 k &prime; |
Wherein, ︱ xt︱ is for cutting apartIn with cut apartThe number of corresponding pixel, denominator represents to cut apartThe number of mid point;
CalculateWithMatching rate:
r t - 1 - > t k &prime; - > k = c | s t k |
Wherein, rightIn each cutting apartLight stream according to t-1 frame to t frame, obtainsIn eachPoint xt-1Corresponding points x in tt, statistics xtBelong toThe number of point be designated as c,For cutting apartInThe number of point;
CalculateWithFinal matching rate:
r t , t - 1 k , k &prime; = min ( r t - > t - 1 k - > k &prime; , r t - 1 - > t k &prime; - > k ) ;
WhenTime, determineWithCoupling, willFlag update beMark, and willJoinIn the data sectional at place, wherein, σmFor default matching threshold;
WhenWhile not existing coupling to cut apart in t-1 frame, willWith mating cutting apart in t-2 frame;
In the time that being greater than default coupling step-length, matching times finishes coupling, for not existing coupling to cut apartDistributeA new mark, willJoin in new data sectional.
4. method according to claim 3, is characterized in that, the described result according to sequential coupling obtainsComprise to motion segmentation and output light stream: the space-time one that carries out respectively motion segmentation according to the result of sequential couplingThe space-time consistency optimization of the optimization of causing property and estimation, obtains the consistent motion segmentation of space-time and output light stream.
5. method according to claim 4, is characterized in that, the space-time consistency of described motion segmentationOptimization comprises:
The dividing mark of fixing each two field picture is constant;
According to light stream, follow the tracks of each pixel x in t frametAt adjacent NtCorresponding points in frame, are waitedSelect tag set P (xt);
According to described candidate's tag set P (xt) obtain the mark that occurrence number is maximum;
Work as Cs<Nt2 or CPt<σPtTime, wherein, CsFor the number of cutting apart comprising in each data sectionalAmount, CPtFor the number of all cut-points in described data sectional, σPtFor predetermined threshold value, for described dataEach cutting apart in segmentation, according to the mark situation of the corresponding points in other frames of the point in described data sectional,Select the data sectional of an optimum Match to force to merge;
Upgrade the mark of remaining data sectional.
6. method according to claim 4, is characterized in that, the space-time consistency of described estimationOptimization comprises:
According to current motion, obtain data sectional that each mark l represents corresponding cut apart imitative in each framePenetrate transformation model
For t frame, according to affine Transform ModelCalculate light stream, to make each in t framePixel xtCorresponding points x in other frames t 't'In the Euclidean distance sum minimum of color space;
Upgrade the light stream of described t frame;
Carry out the detection of occlusion area;
The motion of occlusion area is shown to correction.
7. method according to claim 1 and 2, is characterized in that, described initial motion model comprisesAffine model;
Described in described obtaining, in each two field picture, all initial motion model of cutting apart comprise:
Calculate the parameter of affine transformation, obtain the affine model of matching initial motion;
Obtain the affine motion field of each two field picture according to described affine model;
Described affine motion field and initial optical flow field are merged;
Carry out the detection of occlusion area according to the uniformity of motion;
The affine motion that occlusion area is cut apart according to the place rectification of explicitly moving.
8. method according to claim 7, is characterized in that, described in described basis in each two field pictureAll initial motion model of cutting apart, carry out cluster optimization by the initial motion model with same movement formComprise:
To sorting all the cutting apart according to size in a two field picture, successively descending according to areaScreening affine model;
When cutting apart siArea be greater than preset area SizesOr the affine model collection { quantity of affine model in A}Be less than predetermined number NumATime, will cut apart siAffine model AiWith { A} compares, if { exist in A}A same A of affine model A'iEuclidean distance be less than predeterminable range ThA, give up current affine model Ai,Otherwise by AiJoin that { A} obtains least affine Models Sets { A}.
9. method according to claim 7, is characterized in that, described in described merging in each two field pictureConsistent the cutting apart of moving comprises:
For each s of cutting apart, calculate the color distortion of cutting apart interior all pixels under the current affine model of cutting apartSummation;
Computed segmentation SiThe adjacent set { S of cutting apart of candidateNei, wherein, cut apart s ' completely for each candidate is adjacentFoot 2|s'| > | s|;
For { SNeiIn each adjacent s ' of cutting apart, calculate respectively the uniformity of moving on the same s ' of s bordersIn the uniformity of affine model of the same s ' of current motion of pixelPixel in s is at the affine model of s 'The ratio of the color distortion of the lower colour-difference similarities and differences self affine model
From { SNeiIn select and meet Ss,s'>0.5,Fs,s'>0.5,Cs,s'The adjacent of > 0.5 cut apart, by Cs,s'MaximumAdjacent cutting apart as segmentation candidates, be designated as s*, whenTime, wherein, θwFor default ratio, willS merges to s*In;
Each s of cutting apart is merged, if the last merged sum of cutting apart has exceeded one of initial segmentation numberHalf, re-start the estimation of affine model.
10. method according to claim 1, is characterized in that, described initial light stream comprises forward lightStream and/or oppositely light stream.
11. 1 kinds of movement estimation apparatus, is characterized in that, described device comprises:
Acquiring unit, for obtaining the initial light stream of each two field picture of image sequence of input;
Cutting unit, cuts apart for described each two field picture being divided into at least one according to colouring information;
Motion model estimation unit, for obtaining the described all initial motion model of cutting apart of each two field picture;
Sequential matching unit, for transporting all described each two field picture initial motion model of cutting apartThe moving sequential coupling of cutting apart, to make every in described image sequence of same target in described image sequenceIn one frame, there is identical mark;
Processing unit, for obtaining motion segmentation and output light stream according to the result of sequential coupling.
12. devices according to claim 11, is characterized in that, described device also comprises:
Cluster cell, for according to all initial motion model of cutting apart of described each two field picture, will haveThe initial motion model of same movement form is carried out cluster optimization; And/or
Merge cells, for merging consistent the cutting apart of moving of described each two field picture.
13. according to the device described in claim 11 or 12, it is characterized in that, described sequential matching unitSpecifically for:
Using the first two field picture in the image sequence of input as initial matching masterplate, its each cutting apart as oneIndividual data sectional, for data sectional described in each distributes unique mark;
For cutting apart in t frameLight stream according to t frame to t-1 frame, follows the tracks ofIn each pixel xtCorresponding points x in t-1 framet-1, wherein, xt-1In t-1 frame, be designated as cutting apart of place?StructureThe set becomingExpression is cut apartCorresponding likely cutting apart in t-1 frame;
ForIn each cutting apartCalculateWithMatching rate:
r t - > t - 1 k - > k &prime; = | x t | x t &Element; s t k , x t - 1 &Element; s t - 1 k &prime; | | s t - 1 k &prime; |
Wherein, ︱ xt︱ is for cutting apartIn with cut apartThe number of corresponding pixel, denominator represents to cut apartThe number of mid point;
CalculateWithMatching rate:
r t - 1 - > t k &prime; - > k = c | s t k |
Wherein, rightIn each cutting apartLight stream according to t-1 frame to t frame, obtainsIn eachPoint xt-1Corresponding points x in tt, statistics xtBelong toThe number of point be designated as c,For cutting apartInThe number of point;
CalculateWithFinal matching rate:
r t , t - 1 k , k &prime; = min ( r t - > t - 1 k - > k &prime; , r t - 1 - > t k &prime; - > k ) ;
WhenTime, determineWithCoupling, willFlag update beMark, and willJoinIn the data sectional at place, wherein, σmFor default matching threshold;
WhenWhile not existing coupling to cut apart in t-1 frame, willWith mating cutting apart in t-2 frame;
In the time that being greater than default coupling step-length, matching times finishes coupling, for not existing coupling to cut apartDistributeA new mark, willJoin in new data sectional.
14. devices according to claim 13, is characterized in that, described processing unit comprises: motionCut apart and optimize module and estimation optimization module;
Described motion segmentation is optimized module, for carry out the space-time one of motion segmentation according to the result of sequential couplingThe optimization of causing property;
Described estimation is optimized module, for carry out the space-time one of estimation according to the result of sequential couplingThe optimization of causing property.
15. devices according to claim 14, is characterized in that, described fortune motion segmentation is optimized moduleSpecifically for:
The dividing mark of fixing each two field picture is constant;
According to light stream, follow the tracks of each pixel x in t frametAt adjacent NtCorresponding points in frame, are waitedSelect tag set P (xt);
According to described candidate's tag set P (xt) obtain the mark that occurrence number is maximum;
Work as Cs<Nt2 or CPt<σPtTime, wherein, CsFor the number of cutting apart comprising in each data sectionalAmount, CPtFor the number of all cut-points in described data sectional, σPtFor predetermined threshold value, for described dataEach cutting apart in segmentation, according to the mark situation of the corresponding points in other frames of the point in described data sectional,Select the data sectional of an optimum Match to force to merge;
Upgrade the mark of remaining data sectional.
16. devices according to claim 14, is characterized in that, described estimation is optimized module toolBody is used for:
According to current motion, obtain data sectional that each mark l represents corresponding cut apart imitative in each framePenetrate transformation model
For t frame, according to affine Transform ModelCalculate light stream, to make each in t framePixel xtCorresponding points x in other frames t 't'In the Euclidean distance sum minimum of color space;
Upgrade the light stream of described t frame;
Carry out the detection of occlusion area;
The motion of occlusion area is shown to correction.
17. according to the device described in claim 11 or 12, it is characterized in that described initial motion modelComprise affine model;
Described motion model estimation unit specifically for:
Calculate the parameter of affine transformation, obtain the affine model of matching initial motion;
Obtain the affine motion field of each two field picture according to described affine model;
Described affine motion field and initial optical flow field are merged;
Carry out the detection of occlusion area according to the uniformity of motion;
The affine motion that occlusion area is cut apart according to the place rectification of explicitly moving.
18. devices according to claim 17, is characterized in that, described cluster cell specifically for:
To sorting all the cutting apart according to size in a two field picture, successively descending according to areaScreening affine model;
When cutting apart siArea be greater than preset area SizesOr the affine model collection { quantity of affine model in A}Be less than predetermined number NumATime, will cut apart siAffine model AiWith { A} compares, if { exist in A}A same A of affine model A'iEuclidean distance be less than predeterminable range ThA, give up current affine model Ai,Otherwise by AiJoin that { A} obtains least affine Models Sets { A}.
19. devices according to claim 17, is characterized in that, described merge cells specifically for:
For each s of cutting apart, calculate the color distortion of cutting apart interior all pixels under the current affine model of cutting apartSummation;
Computed segmentation SiThe adjacent set { S of cutting apart of candidateNei, wherein, cut apart s ' completely for each candidate is adjacentFoot 2|s'| > | s|;
For { SNeiIn each adjacent s ' of cutting apart, calculate respectively the uniformity of moving on the same s ' of s bordersIn the uniformity of affine model of the same s ' of current motion of pixelPixel in s is at the affine model of s 'The ratio C of the color distortion of the lower colour-difference similarities and differences self affine models,s'
From { SNeiIn select and meet Ss,s'>0.5,Fs,s'>0.5,Cs,s'The adjacent of > 0.5 cut apart, by Cs,s'MaximumAdjacent cutting apart as segmentation candidates, be designated as s*, whenTime, wherein, θwFor default ratio, willS merges to s*In;
Each s of cutting apart is merged, if the last merged sum of cutting apart has exceeded one of initial segmentation numberHalf, re-start the estimation of affine model.
20. devices according to claim 11, is characterized in that, described initial light stream comprises forward lightStream and/or oppositely light stream.
CN201410578173.1A 2014-10-24 2014-10-24 Motion estimation method and apparatus Pending CN105590327A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410578173.1A CN105590327A (en) 2014-10-24 2014-10-24 Motion estimation method and apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410578173.1A CN105590327A (en) 2014-10-24 2014-10-24 Motion estimation method and apparatus

Publications (1)

Publication Number Publication Date
CN105590327A true CN105590327A (en) 2016-05-18

Family

ID=55929882

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410578173.1A Pending CN105590327A (en) 2014-10-24 2014-10-24 Motion estimation method and apparatus

Country Status (1)

Country Link
CN (1) CN105590327A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106210449A (en) * 2016-08-11 2016-12-07 上海交通大学 The frame rate up-conversion method for estimating of a kind of Multi-information acquisition and system
CN107465911A (en) * 2016-06-01 2017-12-12 东南大学 A kind of extraction of depth information method and device
CN108230305A (en) * 2017-12-27 2018-06-29 浙江新再灵科技股份有限公司 Method based on the detection of video analysis staircase abnormal operating condition
CN108492323A (en) * 2018-01-18 2018-09-04 天津大学 Merge the Submerged moving body detection and recognition methods of machine vision and the sense of hearing
CN110163888A (en) * 2019-05-30 2019-08-23 闽江学院 A kind of novel motion segmentation model quantity detection method
CN110798630A (en) * 2019-10-30 2020-02-14 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5999651A (en) * 1997-06-06 1999-12-07 Matsushita Electric Industrial Co., Ltd. Apparatus and method for tracking deformable objects
CN101765022A (en) * 2010-01-22 2010-06-30 浙江大学 Depth representing method based on light stream and image segmentation
CN103237228A (en) * 2013-04-28 2013-08-07 清华大学 Time-space consistency segmentation method for binocular stereoscopic video

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5999651A (en) * 1997-06-06 1999-12-07 Matsushita Electric Industrial Co., Ltd. Apparatus and method for tracking deformable objects
CN101765022A (en) * 2010-01-22 2010-06-30 浙江大学 Depth representing method based on light stream and image segmentation
CN103237228A (en) * 2013-04-28 2013-08-07 清华大学 Time-space consistency segmentation method for binocular stereoscopic video

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
GUOFENG ZHANG 等: "Robust Bilayer Segmentation and Motion/Depth Estimation with a Handheld Camera", <IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE> *
MICHAEL J. BLACK 等: "Estimating Multiple Independent Motions in Segmented Images using Parametric Models with Local Deformations", 《MOTION OF NON-RIGID AND ARTICULATED OBJECTS》 *
卢汉清 等: "基于内容的视频信号与图像库检索中的图像技术", 《自动化学报》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107465911A (en) * 2016-06-01 2017-12-12 东南大学 A kind of extraction of depth information method and device
CN107465911B (en) * 2016-06-01 2019-03-15 东南大学 A kind of extraction of depth information method and device
CN106210449A (en) * 2016-08-11 2016-12-07 上海交通大学 The frame rate up-conversion method for estimating of a kind of Multi-information acquisition and system
CN106210449B (en) * 2016-08-11 2020-01-07 上海交通大学 Multi-information fusion frame rate up-conversion motion estimation method and system
CN108230305A (en) * 2017-12-27 2018-06-29 浙江新再灵科技股份有限公司 Method based on the detection of video analysis staircase abnormal operating condition
CN108492323A (en) * 2018-01-18 2018-09-04 天津大学 Merge the Submerged moving body detection and recognition methods of machine vision and the sense of hearing
CN110163888A (en) * 2019-05-30 2019-08-23 闽江学院 A kind of novel motion segmentation model quantity detection method
CN110163888B (en) * 2019-05-30 2021-03-02 闽江学院 Novel motion segmentation model quantity detection method
CN110798630A (en) * 2019-10-30 2020-02-14 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
Zhang et al. Learning signed distance field for multi-view surface reconstruction
US11763433B2 (en) Depth image generation method and device
US10803546B2 (en) Systems and methods for unsupervised learning of geometry from images using depth-normal consistency
CN109003325B (en) Three-dimensional reconstruction method, medium, device and computing equipment
CN105993034B (en) Contour completion for enhanced surface reconstruction
CN110688905B (en) Three-dimensional object detection and tracking method based on key frame
CN105590327A (en) Motion estimation method and apparatus
US20130129190A1 (en) Model-Based Stereo Matching
CN114424250A (en) Structural modeling
CN109255833A (en) Based on semantic priori and the wide baseline densification method for reconstructing three-dimensional scene of gradual optimization
CN104200523A (en) Large-scale scene three-dimensional reconstruction method for fusion of additional information
EP3293700B1 (en) 3d reconstruction for vehicle
Zhang et al. Robust stereo matching with surface normal prediction
Liu et al. Deep learning based multi-view stereo matching and 3D scene reconstruction from oblique aerial images
CN107610148B (en) Foreground segmentation method based on binocular stereo vision system
Hu et al. Geometric feature enhanced line segment extraction from large-scale point clouds with hierarchical topological optimization
Yan et al. Efficient implicit neural reconstruction using lidar
Xu et al. Video-object segmentation and 3D-trajectory estimation for monocular video sequences
CN117315169A (en) Live-action three-dimensional model reconstruction method and system based on deep learning multi-view dense matching
Tian et al. HPM-TDP: An efficient hierarchical PatchMatch depth estimation approach using tree dynamic programming
Shi et al. Improved event-based dense depth estimation via optical flow compensation
CN113763474B (en) Indoor monocular depth estimation method based on scene geometric constraint
Nobis et al. Exploring the capabilities and limits of 3d monocular object detection-a study on simulation and real world data
CN113920254A (en) Monocular RGB (Red Green blue) -based indoor three-dimensional reconstruction method and system thereof
Gao et al. Multi-target 3d reconstruction from rgb-d data

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160518