CN102034247A - Motion capture method for binocular vision image based on background modeling - Google Patents
Motion capture method for binocular vision image based on background modeling Download PDFInfo
- Publication number
- CN102034247A CN102034247A CN 201010602544 CN201010602544A CN102034247A CN 102034247 A CN102034247 A CN 102034247A CN 201010602544 CN201010602544 CN 201010602544 CN 201010602544 A CN201010602544 A CN 201010602544A CN 102034247 A CN102034247 A CN 102034247A
- Authority
- CN
- China
- Prior art keywords
- background
- image
- binocular vision
- pixel
- binocular
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 44
- 230000033001 locomotion Effects 0.000 title claims abstract description 24
- 230000008569 process Effects 0.000 claims abstract description 17
- 230000000694 effects Effects 0.000 claims abstract description 8
- 230000011218 segmentation Effects 0.000 claims description 7
- 230000006870 function Effects 0.000 claims description 6
- 230000005484 gravity Effects 0.000 claims description 6
- 230000008859 change Effects 0.000 claims description 3
- 230000004927 fusion Effects 0.000 claims description 2
- 238000004321 preservation Methods 0.000 claims description 2
- 230000001105 regulatory effect Effects 0.000 claims description 2
- 238000000605 extraction Methods 0.000 claims 1
- 230000003993 interaction Effects 0.000 abstract description 6
- 238000007670 refining Methods 0.000 abstract 1
- 238000005516 engineering process Methods 0.000 description 10
- 230000002452 interceptive effect Effects 0.000 description 4
- 230000013011 mating Effects 0.000 description 4
- 238000011161 development Methods 0.000 description 3
- 230000018109 developmental process Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 239000004744 fabric Substances 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 230000000903 blocking effect Effects 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000010168 coupling process Methods 0.000 description 2
- 238000005859 coupling reaction Methods 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 238000009499 grossing Methods 0.000 description 2
- 238000007654 immersion Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004040 coloring Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000000977 initiatory effect Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012797 qualification Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 239000011800 void material Substances 0.000 description 1
Images
Abstract
The invention discloses a motion capture method for a binocular vision image based on background modeling. By the method, a human body serving as a foreground can be divided and the upper torso of the human body is subjected to motion capture simultaneously so as to achieve the effect of man-machine interaction. The method comprises the following steps of: on the basis of background modeling, establishing a Gaussian model for a clean background acquired by a camera, comparing an acquired video with a background model, providing a probability value which belongs to the foreground or the background for each pixel of a scene through depth information acquired by a binocular camera, and dividing the foreground and the background of the scene through an image cutting algorithm. When the divided foreground is the upper torso of the human body, a basic skeleton model of the human body is acquired by refining a foreground profile, de-noising and determining a key point so as to complete the motion capture process.
Description
Technical field
The invention belongs to computer vision technique and interactive digital entertainment field, relate to a kind of background segment that binocular camera shooting head and background modeling technology finish and motion-captured process utilized.
Background technology
Movement capturing technology refers to computer vision or other means utilized, and can capture the motion process of human body in real time, exactly.Along with the development of computer software and hardware and the raising of computer user's demand, movement capturing technology is more obvious in the effect of the inside, fields such as digital entertainment, video monitoring, motion analysis.
Yet the development of movement capturing technology also is subjected to the restriction of various conditions and many limitation occur.Such as problems such as blocking of variation, complicated background and the motion process of light.These factors make motion-captured process become more difficult.Yet, carry out the result of background segment by the method for utilizing binocular vision, the prospect in scene has only under the prerequisite of human body, and motion-captured problem will be converted into the prospect profile problem of analyzing scene, make calculated amount simplify greatly.Simultaneously, in the interactive digital entertainment field, movement capturing technology also is the research focus of man-machine interaction in playing in recent years as a kind of video interactive technology.And camera become the general outfit of PC, and man-machine interaction mode general, immersion more and more becomes the focus of digital entertainment research.So, have the research prospect of application fields based on the binocular vision movement capturing technology of background segment technology.
Summary of the invention
Cutting apart of prospect that the objective of the invention is to utilize the binocular camera shooting head to obtain scene and background, finish motion-captured process simultaneously on this basis.This method is at first trained clean background, gathers the background picture of certain frame number, finishes the foundation of background model.On this basis, utilize the new image of gathering, finish the foundation that figure cuts network chart with the color distortion of background model and the depth information of binocular vision, and the method for utilizing dynamic figure to cut, scene prospect and background are cut apart.On the basis of cutting apart, the human body of prospect is carried out the analysis of structure simultaneously, obtain the location of upper body trunk various piece, thereby finish motion-captured process.
For achieving the above object, the invention provides based on background modeling the motion capture method of binocular vision image comprised that step is as follows:
Step S1: with binocular camera shooting head stationkeeping, close white balance, obtain the binocular vision image;
Step S2: to the binocular vision image that obtains, under the clean background image of setting frame number, carry out background modeling, obtain background model;
Step S3: the binocular depth information that utilizes computer binocular vision to obtain, calculating pixel belongs to the probability of prospect and background;
Step S4: utilize binocular depth information and background modeling data and dynamic figure to cut algorithm, binocular vision display foreground and background are cut apart, and extracted prospect profile;
Step S5: prospect profile carries out refinement, determines the human body key point, finishes motion-captured.
Good effect of the present invention:
The present invention utilizes computer vision and image processing techniques, isolates the human body of prospect naturally from scene, and finishes the motion-captured of upper body trunk, thereby realizes the man-machine interaction of nature.The characteristics of traditional interactive mode are based on the hand contact, as mouse, keyboard etc.Development along with computer vision technique, increasing system is by the process of having finished man-machine interaction of the method nature of camera, the user can experience the enjoyment of man-machine interaction more easily by the mode of vision, simultaneously, interface as recreation makes the game player obtain more feeling of immersion.
In addition, the present invention has utilized the collection of binocular vision and the foundation of background model.The employing of binocular vision mainly is to have made full use of depth information, and the prospect of considering often belongs to from the nearer zone of camera, the problem of avoided by shade simultaneously, blocking the segmentation errors that causes.In addition, setting up background model can be so that the cost of cutting apart better obtains calculating, the method for utilizing dynamic figure to cut simultaneously, make cut apart quicker.
Description of drawings
Figure 1A is overall flow figure of the present invention;
Fig. 1 is a binocular vision image of the present invention;
Fig. 2 is left figure and right figure and the parallax that the present invention utilizes binocular vision to obtain;
Fig. 3 is that figure of the present invention cuts the max-flow of algorithm or the network flow graph of minimal cut;
Fig. 4 is a process flow diagram of the present invention;
Fig. 5 is one group of design sketch that video background is cut apart of the present invention;
Fig. 6 is a background segment result's of the present invention edge smoothing synoptic diagram;
Fig. 7 is the refinement of profile of the present invention and the result that key position extracts.
Embodiment
Below in conjunction with accompanying drawing the present invention is described in detail, described embodiment only is intended to be convenient to the understanding of the present invention, and it is not played any qualification effect.
Further specify a kind of operating process of the motion capture method based on background modeling below by example.
All codes of this example are C++ and write, and move under Microsoft visual studio 2005 environment, can also adopt other software and hardware conditions, do not repeat them here.
Figure 1A illustrates and the present invention is based on the overall flow figure of background modeling to the motion capture method of binocular vision image.
The present invention is based on the motion capture method of background modeling to the binocular vision image, based on binocular vision and background segment, its method comprises that step is as follows:
Step S1: with binocular camera shooting head stationkeeping, close white balance, obtain the binocular vision image;
Step S2: to the binocular vision image that obtains, under the clean background image of setting frame number, carry out background modeling, obtain background model;
Step S3: the binocular depth information that utilizes computer binocular vision to obtain, calculating pixel belongs to the probability of prospect and background;
Step S4: utilize binocular depth information and background modeling data and dynamic figure to cut algorithm, binocular vision display foreground and background are cut apart, and extracted prospect profile;
Step S5: prospect profile carries out refinement, determines the human body key point, finishes motion-captured.
Comprise according to obtaining the binocular vision image step described in the step S2:
Step S211: guarantee the stationkeeping of camera, do not have tangible light and shade to change in the scene;
Step S212: close the Automatic white balance of camera, in the hardware parameter of camera, the function of automatic exposure parameter and Automatic white balance is arranged generally, so that when scene light changes, realize regulating automatically the function of picture quality; In background modeling, need to set white balance parameter and fix;
Step S213: gather the fixedly clean background image of frame number (100 frame), be stored in the internal memory.
Comprise as follows according to the step of under the clean background image of setting frame number, carrying out background modeling described in the step S2:
Step S221: utilize the Gaussian Background model to gather the coloured image of each frame in the binocular vision image, use R respectively, G, B represent red, and green and blue three-channel value, span are 0~255;
Step S221: obtained N image in the background modeling process, each image comprises 320 * 240 pixels, calculate the brightness I of each pixel and colourity (r, g).Wherein, r=R/ (R+G+B), g=G/ (R+G+B), R, G, B represent the redness in the Color Channel respectively, the value of green and blue component;
Step S221: the fusion background model of setting up pixel scale; Calculate the brightness and average and the variance of colourity in N image of each pixel, and deposit internal memory in;
Step S221: set up feature background model at brightness space, chrominance space set up based on the colourity model.Deposit the colourity obtained and the background model in the brightness space in internal memory.
According to the depth data cost of each pixel in the binocular vision image described in the step S2 is calculated, obtain the degree of depth cost of each pixel, thereby the binocular depth information is introduced, concrete steps comprise:
Step 231: gather and preservation binocular vision image, be designated as left image and right image respectively;
Step 232: set a depth value for each pixel of left image, described depth value is represented with the parallax of left image and right image;
Step 233:, calculate the difference cost of left image and right image at each depth value;
Step 234: add up the cost value in the left image, and the cost value in the left image is divided into four groups according to the size of described cost value;
Step 235: the cost value of each group is upgraded the prospect of this pixel and the cost of background, and the cost that wherein belongs to prospect reduces according to the exponential relationship to parallax, and the cost of background increases according to the exponential relationship of parallax.
According to utilizing binocular depth information and background modeling data and dynamic figure to cut algorithm described in the step S4, binocular vision display foreground and background to be cut apart, and extracted prospect profile, concrete steps comprise:
Step S41: background modeling reads in the binocular vision image that newly reads after finishing, and described binocular vision image comprises left image and right image;
Step S42: utilize the result of binocular vision data cost acquisition, obtain the data cost of binocular information;
Step S43: utilize background model, compare, obtain to utilize figure to cut algorithm basic principle, set up the network flow of max-flow or minimal cut based on the color cost value with the pixel of left figure;
Step S44: two data cost value utilizing step S42 and step S43 to obtain obtain figure and cut data cost value in the algorithm;
Step S45: utilize the relationship of contrast between the left pixel, the level and smooth item that figure is cut in the algorithm carries out assignment;
Step S46: utilize dynamic figure to cut algorithm, will cut apart based on the video flowing of pixel aspect, segmentation result is divided into two parts, and a part is a prospect, and a part is a background in addition.
Step S47: the prospect background that will cut apart is stored in the picture of identical size according to 0 or 1, and 0 or 1 prospect background picture is obtained edge contour;
Step S48: the mode of utilizing High frequency filter makes the edge more level and smooth the edge denoising;
Step S49: utilize the cut zone of the data error of former frames to proofread and correct.
Utilize picture denoising, refinement mode according to step S5, obtain the key point of trunk, thereby realize that motion-captured effect step comprises:
Step S51: will carry out convergent-divergent through the human body contour outline of aftertreatment;
Step S52: the human body contour outline of convergent-divergent is carried out refinement;
Step S53: the human body contour outline of refinement is enlarged, expand original size to;
Step S54: once more profile is carried out refinement;
Step S55: find neighborhood territory pixel greater than 2 node, and get its center-of-gravity value, be set at the gravity center of human body;
Step S56: search for up and down along center of gravity, find node, be set at head and waist;
Step S57: along about center of gravity, search for, find left arm and right arm, and proportionally determine ancon and shoulder with eccentricity;
Step S58: 9 key points will determining compare with former frames, obtain comparatively stable and trunk position accurately.
The first step as shown in Figure 1 is an images acquired.This method adopts the input of binocular vision video.Among the figure, (z) expression is the coordinate of world coordinate system for x, y; (x
L, y
L) and (x
R, y
R) pixel coordinate of the same object of expression in left figure and right figure.
(1) mostly the information of Digital Image Processing is two-dimensional signal, and the process information amount is very big.The piece image here with two-dimensional function f (x, y) expression, x wherein, y is a two-dimensional coordinate, f (x, y) expression point (x, colouring information y).Camera is gathered all optical information in the camera lens from the space, these information enter after the computing machine, is converted to the color model that meets computer standard, carries out Digital Image Processing with the program of entering, and guarantees the continuity and the real-time of video.From the image of gathering, each pixel is handled 76800 pixels of 320 * 240 pixels altogether.The initial effect of video of gathering as shown in Figure 1.Project all operations and computing 320 * 240 pixels that all are based on this each frame subsequently.In the binocular vision, same pixel about the position difference of imaging among two figure, and the size of position difference, reflection be the degree of depth of image.Relatively moving of two pixels can be calculated by the coupling of pixel.Method of the present invention is utilized these information, auxiliary finishing cutting apart of prospect and background.As shown in Figure 2, the utilization of binocular information is that the cost that two width of cloth figure mate about usefulness realizes.What wherein P represented is the position of certain pixel in left figure, and P+d represents the position of this pixel in right figure, and what d represented is exactly the parallax (Display) of this pixel.
(2) the present invention is made up of two parts in the process of utilizing the binocular depth information.
Step 1: the coupling cost in that pixel xi calculates is divided into four groups (maximal value of parallax d is set at 32) according to different parallax value:
A group: pixel x
iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d>16, represent that this pixel belongs to prospect very much;
B group: pixel x
iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d≤16and d>12, represent that this pixel has the very big prospect that may belong to;
C group: pixel x
iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d≤12and d>5, represent that this pixel has the very big background that may belong to;
D group: pixel x
iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d≤5, represent that this pixel belongs to background very much.
Under such hypothesis, the present invention needs less time that pixel is divided into four groups, rather than each pixel is carried out 32 possible parallaxes suppose.
Step 2: set the suitable data cost value for figure cuts algorithm.Data item of the present invention comprises that respectively this pixel belongs to the cost of prospect or background, uses D respectively
i(B) and D
i(F) expression.The parallax value of pixel is big more, so it to belong to the possibility of prospect big more, so D
i(F) value correspondence reduces D
i(B) value is corresponding to be increased.By such corresponding relation, the present invention proposes a corresponding scheme, express with following formula:
For all t=A, B, C, D, λ
t>0.Wherein
What represent is the background model data item that incorporates binocular information, belongs to t=A respectively, B, C, four groups of D.D
i(B) expression is the background segment data item of monocular vision.λ
tBe the parameter of binocular data cost, what i represented is pixel coordinate.
What represent is the foreground model data item that incorporates binocular information.The parallax value that d represents (Disparity).c
tWhat represent is the parameter of control d.
As shown in Figure 3, figure cut algorithm max-flow or the network flow graph of minimal cut.P wherein, what q represented is two adjacent pixels.Shown in Figure 4 is the process flow diagram that figure cuts algorithm, comprises the assignment of front end and the partitioning portion of rear end.
(3) figure cuts the important component part that algorithm is a background segment, and it to the effect that utilizes the principle of max-flow or minimal cut, the pixel in the image is cut apart according to certain path, and which is calculated belong to prospect or and background respectively.
The segmentation problem of prospect or background in the image can be considered as the binary identified problems in the computer vision field.If pixel i belongs to prospect, the label f of this pixel of mark then
i=F, F refers to prospect.In like manner, if this pixel belongs to background, then be labeled as f
i=B.Correspond to two-value label problem, label set only comprises two labels.Figure cuts the weighted graph that algorithm constructs and comprises two corresponding with it summit s and t.As shown in Figure 3, among the figure, left figure is the weighted graph G that provides by 3 * 3 original image structure, G=<V, ε 〉, wherein V is a vertex set, is to be called source node S respectively and terminal node T is dimerous by ordinary node and two.Wherein S and the T two-value label of representing prospect and background respectively for summit ε representative be the limit that is connected the summit, the weights size on limit is represented with the thickness of simplification in last figure.
Flow process such as Fig. 4 that dynamic figure cuts.Comprise data item and level and smooth in the energy function, being provided with of they directly affects figure and cuts the final segmentation result of algorithm.That Fig. 5 represents is several groups of Video Segmentation results of the present invention, and wherein left side 3 width of cloth figure are left figure images of doing video in the input video, and right side 3 width of cloth figure are the results after cutting apart.
(4) low-pass filter that designed in the frequency domain of the present invention comes smooth boundary.Along boundary curve C, the boundary curve of the process of edge smoothing of the present invention as shown in Figure 6, the picture left above is represented the input source image, top right plot is represented the result cut apart; The flat prospect of spending that lower-left figure represents or the edge of background, bottom-right graph is represented is result after level and smooth.Point sequence z (i)=[x (i), y (i)] its complex representation form that obtains of sampling at certain intervals is:
z(i)=x(i)+jy(i)
The Fourier transform of discrete z (i) is:
In the formula, j, u, K represent complex symbol respectively, frequency and constant term, f (u) is the Fourier transform of z (i), the Fourier that is called the border is described son, is the expression of boundary point sequence in frequency domain.By the Fourier transform theory as can be known, high fdrequency component comprises details, low frequency component decision global shape.Curve is because jagged ability is rough, and high fdrequency component is contained in these rough zones.The HFS of f (u) is carried out filtering just can obtain smooth curve.The present invention defines the high-frequency energy of low frequency energy ratio and filtering 5%:
Wherein || be modulo operation.Getting the minimum l value that r (l)>0.95 is set up is the cutoff frequency of low-pass filter.Utilize the character of fourier coefficient
(
Be the conjugate complex number of f).In coefficient f (u), the radio-frequency component of cancellation in from l to the K-1-l scope.Carry out inverse fourier transform again, the part of curve sudden change has obtained smoothly.
Being motion-captured result of the present invention as shown in Figure 7, wherein is two two field pictures of left figure in the video in the left hand view, and the right side is key point and the skeleton that segmentation result has extracted.Key point represents with circle that skeleton is represented with line.
(5) the present invention motion-captured on the basis of cutting apart comprises three steps,
Step 1: the result that will cut apart carries out aftertreatment, obtains level and smooth and stable relatively contour area, cuts apart owing to relate to profile, so the border does not need accurate calculating.Under will not situation, can finish the skeleton motion tracking effect that this paper needs preferably than macroscopic-void.
Step 2: the profile that will cut apart positions, and determines the basic comprising of nine points.Comprising A
1, A
2, A
3Be representative group, A herein
1, A
2, A
3, A
4, A
5, A
6, A
7, A
8, A
9Nine points.A
1, A
2, A
3Represent three points of head and trunk respectively, A
4, A
5, A
6And A
7, A
8, A
9Represent three points of left arm and right arm respectively.
Step 3: the order of nine points being installed the skeleton profile connects, and finishes motion-captured.
The above; only be the embodiment among the present invention, but protection scope of the present invention is not limited thereto, anyly is familiar with the people of this technology in the disclosed technical scope of the present invention; can understand conversion or the replacement expected, all should be encompassed within the protection domain of claims of the present invention.
Claims (6)
1. one kind based on the motion capture method of background modeling to the binocular vision image, is based on the method for binocular vision and background segment, it is characterized in that, comprises that step is as follows:
Step S1: with binocular camera shooting head stationkeeping, close white balance, obtain the binocular vision image;
Step S2: to the binocular vision image that obtains, under the clean background image of setting frame number, carry out background modeling, obtain background model;
Step S3: the binocular depth information that utilizes computer binocular vision to obtain, calculating pixel belongs to the probability of prospect and background;
Step S4: utilize binocular depth information and background modeling data and dynamic figure to cut algorithm, binocular vision display foreground and background are cut apart, and extracted prospect profile;
Step S5: prospect profile carries out refinement, determines the human body key point, finishes motion-captured.
2. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the step of obtaining the binocular vision image described in the step S2 comprises as follows:
Step S211: guarantee the stationkeeping of camera, do not have tangible light and shade to change in the scene;
Step S212: close the Automatic white balance of camera, in the hardware parameter of camera, the function of automatic exposure parameter and Automatic white balance is arranged generally, so that when scene light changes, realize regulating automatically the function of picture quality; In background modeling, need to set white balance parameter and fix;
Step S213: gather fixedly frame number clean background image, be stored in the internal memory.
3. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the step of carrying out background modeling described in the step S2 under the clean background image of setting frame number comprises as follows:
Step S221: utilize the Gaussian Background model to gather the coloured image of each frame in the binocular vision image, use R respectively, G, B represent red, and green and blue three-channel value, span are 0~255;
Step S221: in the background modeling process, obtained N image, each image comprises 320 * 240 pixels, calculate the brightness I of each pixel and colourity (r, g), wherein, r=R/ (R+G+B), g=G/ (R+G+B), R, G, B represents the redness in the Color Channel respectively, the value of green and blue component;
Step S221: the fusion background model of setting up pixel scale; Calculate the brightness and average and the variance of colourity in N image of each pixel, and deposit internal memory in;
Step S221: set up feature background model at brightness space, chrominance space set up based on the colourity model, deposit the colourity obtained and the background model in the brightness space in internal memory.
4. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the depth data cost to each pixel in the binocular vision image described in the step S2 is calculated, obtain the degree of depth cost of each pixel, thereby the binocular depth information is introduced, and concrete steps comprise as follows:
Step 231: gather and preservation binocular vision image, be designated as left image and right image respectively;
Step 232: set a depth value for each pixel of left image, described depth value is represented with the parallax of left image and right image;
Step 233:, calculate the difference cost of left image and right image at each depth value;
Step 234: add up the cost value in the left image, and the cost value in the left image is divided into four groups according to the size of described cost value;
Step 235: the cost value of each group is upgraded the prospect of this pixel and the cost of background, and the cost that wherein belongs to prospect reduces according to the exponential relationship to parallax, and the cost of background increases according to the exponential relationship of parallax.
5. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the described binocular depth information and background modeling data and dynamic figure of utilizing of step S4 cut algorithm, binocular vision display foreground and background are cut apart, and the extraction prospect profile, concrete steps comprise as follows:
Step S41: background modeling reads in the binocular vision image that newly reads after finishing, and described binocular vision image comprises left image and right image;
Step S42: utilize the result of binocular vision data cost acquisition, obtain the data cost of binocular information;
Step S43: utilize background model, compare, obtain to utilize figure to cut algorithm basic principle, set up the network flow of max-flow or minimal cut based on the color cost value with the pixel of left figure;
Step S44: two data cost value utilizing step S42 and step S43 to obtain obtain figure and cut data cost value in the algorithm;
Step S45: utilize the relationship of contrast between the left pixel, the level and smooth item that figure is cut in the algorithm carries out assignment;
Step S46: utilize dynamic figure to cut algorithm, will cut apart based on the video flowing of pixel aspect, segmentation result is divided into two parts, and a part is a prospect, and a part is a background in addition;
Step S47: the prospect background that will cut apart is stored in the picture of identical size according to 0 or 1, and 0 or 1 prospect background picture is obtained edge contour;
Step S48: the mode of utilizing High frequency filter makes the edge more level and smooth the edge denoising;
Step S49: utilize the cut zone of the data error of former frames to proofread and correct.
6. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: utilize picture denoising, refinement mode, obtain the key point of trunk, thereby realize that motion-captured effect step comprises as follows:
Step S51: will carry out convergent-divergent through the human body contour outline of aftertreatment;
Step S52: the human body contour outline of convergent-divergent is carried out refinement;
Step S53: the human body contour outline of refinement is enlarged, expand original size to;
Step S54: once more profile is carried out refinement;
Step S55: find neighborhood territory pixel greater than 2 node, and get its center-of-gravity value, be set at the gravity center of human body;
Step S56: search for up and down along center of gravity, find node, be set at head and waist;
Step S57: along about center of gravity, search for, find left arm and right arm, and proportionally determine ancon and shoulder with eccentricity;
Step S58: 9 key points will determining compare with former frames, obtain comparatively stable and trunk position accurately.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201010602544 CN102034247B (en) | 2010-12-23 | 2010-12-23 | Motion capture method for binocular vision image based on background modeling |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN 201010602544 CN102034247B (en) | 2010-12-23 | 2010-12-23 | Motion capture method for binocular vision image based on background modeling |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102034247A true CN102034247A (en) | 2011-04-27 |
CN102034247B CN102034247B (en) | 2013-01-02 |
Family
ID=43887100
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN 201010602544 Expired - Fee Related CN102034247B (en) | 2010-12-23 | 2010-12-23 | Motion capture method for binocular vision image based on background modeling |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102034247B (en) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102184008A (en) * | 2011-05-03 | 2011-09-14 | 北京天盛世纪科技发展有限公司 | Interactive projection system and method |
CN102927652A (en) * | 2012-10-09 | 2013-02-13 | 清华大学 | Intelligent air conditioner control method based on positions of indoor persons and objects |
CN103826071A (en) * | 2014-03-11 | 2014-05-28 | 深圳市中安视科技有限公司 | Three-dimensional camera shooting method for three-dimensional identification and continuous tracking |
CN104243951A (en) * | 2013-06-07 | 2014-12-24 | 索尼电脑娱乐公司 | Image processing device, image processing system and image processing method |
CN105374043A (en) * | 2015-12-02 | 2016-03-02 | 福州华鹰重工机械有限公司 | Method and device of background filtering of visual odometry |
CN105516579A (en) * | 2014-09-25 | 2016-04-20 | 联想(北京)有限公司 | Image processing method and device and electronic equipment |
CN106056056A (en) * | 2016-05-23 | 2016-10-26 | 浙江大学 | Long-distance non-contact luggage volume detection system and method thereof |
CN109064511A (en) * | 2018-08-22 | 2018-12-21 | 广东工业大学 | A kind of gravity center of human body's height measurement method, device and relevant device |
CN109214996A (en) * | 2018-08-29 | 2019-01-15 | 深圳市元征科技股份有限公司 | A kind of image processing method and device |
CN110490877A (en) * | 2019-07-04 | 2019-11-22 | 西安理工大学 | Binocular stereo image based on Graph Cuts is to Target Segmentation method |
CN111567036A (en) * | 2017-12-07 | 2020-08-21 | 微软技术许可有限责任公司 | Video capture system and method |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040125207A1 (en) * | 2002-08-01 | 2004-07-01 | Anurag Mittal | Robust stereo-driven video-based surveillance |
US20070031037A1 (en) * | 2005-08-02 | 2007-02-08 | Microsoft Corporation | Stereo image segmentation |
CN101344965A (en) * | 2008-09-04 | 2009-01-14 | 上海交通大学 | Tracking system based on binocular camera shooting |
CN101389004A (en) * | 2007-09-13 | 2009-03-18 | 中国科学院自动化研究所 | Moving target classification method based on on-line study |
-
2010
- 2010-12-23 CN CN 201010602544 patent/CN102034247B/en not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040125207A1 (en) * | 2002-08-01 | 2004-07-01 | Anurag Mittal | Robust stereo-driven video-based surveillance |
US20070031037A1 (en) * | 2005-08-02 | 2007-02-08 | Microsoft Corporation | Stereo image segmentation |
CN101389004A (en) * | 2007-09-13 | 2009-03-18 | 中国科学院自动化研究所 | Moving target classification method based on on-line study |
CN101344965A (en) * | 2008-09-04 | 2009-01-14 | 上海交通大学 | Tracking system based on binocular camera shooting |
Non-Patent Citations (2)
Title |
---|
《IEEE Transactions on Pattern Analysis and Machine Intelligence》 20060930 Vladimir Kolmogorov et al. Probabilistic Fusion of Stereo with Color and Contrast for Bilayer Segmentation 全文 1-6 第28卷, 第9期 2 * |
《LNCS》 20091231 Xiaoyu Wu et al. Video Background Segmentation Using Adaptive Background Models 全文 1-6 第5716卷, 2 * |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102184008A (en) * | 2011-05-03 | 2011-09-14 | 北京天盛世纪科技发展有限公司 | Interactive projection system and method |
CN102927652B (en) * | 2012-10-09 | 2015-06-24 | 清华大学 | Intelligent air conditioner control method based on positions of indoor persons and objects |
CN102927652A (en) * | 2012-10-09 | 2013-02-13 | 清华大学 | Intelligent air conditioner control method based on positions of indoor persons and objects |
CN104243951B (en) * | 2013-06-07 | 2017-01-11 | 索尼电脑娱乐公司 | Image processing device, image processing system and image processing method |
CN104243951A (en) * | 2013-06-07 | 2014-12-24 | 索尼电脑娱乐公司 | Image processing device, image processing system and image processing method |
US10293252B2 (en) | 2013-06-07 | 2019-05-21 | Sony Interactive Entertainment Inc. | Image processing device, system and method based on position detection |
CN103826071A (en) * | 2014-03-11 | 2014-05-28 | 深圳市中安视科技有限公司 | Three-dimensional camera shooting method for three-dimensional identification and continuous tracking |
CN105516579A (en) * | 2014-09-25 | 2016-04-20 | 联想(北京)有限公司 | Image processing method and device and electronic equipment |
CN105374043A (en) * | 2015-12-02 | 2016-03-02 | 福州华鹰重工机械有限公司 | Method and device of background filtering of visual odometry |
CN106056056A (en) * | 2016-05-23 | 2016-10-26 | 浙江大学 | Long-distance non-contact luggage volume detection system and method thereof |
CN111567036A (en) * | 2017-12-07 | 2020-08-21 | 微软技术许可有限责任公司 | Video capture system and method |
CN109064511A (en) * | 2018-08-22 | 2018-12-21 | 广东工业大学 | A kind of gravity center of human body's height measurement method, device and relevant device |
CN109064511B (en) * | 2018-08-22 | 2022-02-15 | 广东工业大学 | Method and device for measuring height of center of gravity of human body and related equipment |
CN109214996A (en) * | 2018-08-29 | 2019-01-15 | 深圳市元征科技股份有限公司 | A kind of image processing method and device |
CN109214996B (en) * | 2018-08-29 | 2021-11-12 | 深圳市元征科技股份有限公司 | Image processing method and device |
CN110490877A (en) * | 2019-07-04 | 2019-11-22 | 西安理工大学 | Binocular stereo image based on Graph Cuts is to Target Segmentation method |
Also Published As
Publication number | Publication date |
---|---|
CN102034247B (en) | 2013-01-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102034247B (en) | Motion capture method for binocular vision image based on background modeling | |
CN111797716B (en) | Single target tracking method based on Siamese network | |
CN107204010B (en) | A kind of monocular image depth estimation method and system | |
CN107423698B (en) | A kind of gesture estimation method based on convolutional neural networks in parallel | |
CN102567727B (en) | Method and device for replacing background target | |
CN105160310A (en) | 3D (three-dimensional) convolutional neural network based human body behavior recognition method | |
CN110008832A (en) | Based on deep learning character image automatic division method, information data processing terminal | |
CN109146948A (en) | The quantization of crop growing state phenotypic parameter and the correlation with yield analysis method of view-based access control model | |
CN103455984A (en) | Method and device for acquiring Kinect depth image | |
CN110398720A (en) | A kind of anti-unmanned plane detection tracking interference system and photoelectric follow-up working method | |
CN102184551A (en) | Automatic target tracking method and system by combining multi-characteristic matching and particle filtering | |
CN105869178A (en) | Method for unsupervised segmentation of complex targets from dynamic scene based on multi-scale combination feature convex optimization | |
CN104966286A (en) | 3D video saliency detection method | |
CN113240691A (en) | Medical image segmentation method based on U-shaped network | |
CN112464847B (en) | Human body action segmentation method and device in video | |
CN111179189B (en) | Image processing method and device based on generation of countermeasure network GAN, electronic equipment and storage medium | |
CN111462027B (en) | Multi-focus image fusion method based on multi-scale gradient and matting | |
CN112733950A (en) | Power equipment fault diagnosis method based on combination of image fusion and target detection | |
CN109712247B (en) | Live-action training system based on mixed reality technology | |
CN110263768A (en) | A kind of face identification method based on depth residual error network | |
CN103413323B (en) | Based on the object tracking methods of component-level apparent model | |
CN103440662A (en) | Kinect depth image acquisition method and device | |
CN104063871B (en) | The image sequence Scene Segmentation of wearable device | |
CN111382613A (en) | Image processing method, apparatus, device and medium | |
CN107194948B (en) | Video significance detection method based on integrated prediction and time-space domain propagation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20130102 Termination date: 20151223 |
|
EXPY | Termination of patent right or utility model |