CN102034247A - Motion capture method for binocular vision image based on background modeling - Google Patents

Motion capture method for binocular vision image based on background modeling Download PDF

Info

Publication number
CN102034247A
CN102034247A CN 201010602544 CN201010602544A CN102034247A CN 102034247 A CN102034247 A CN 102034247A CN 201010602544 CN201010602544 CN 201010602544 CN 201010602544 A CN201010602544 A CN 201010602544A CN 102034247 A CN102034247 A CN 102034247A
Authority
CN
China
Prior art keywords
background
image
binocular vision
pixel
binocular
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN 201010602544
Other languages
Chinese (zh)
Other versions
CN102034247B (en
Inventor
王阳生
时岭
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN 201010602544 priority Critical patent/CN102034247B/en
Publication of CN102034247A publication Critical patent/CN102034247A/en
Application granted granted Critical
Publication of CN102034247B publication Critical patent/CN102034247B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

The invention discloses a motion capture method for a binocular vision image based on background modeling. By the method, a human body serving as a foreground can be divided and the upper torso of the human body is subjected to motion capture simultaneously so as to achieve the effect of man-machine interaction. The method comprises the following steps of: on the basis of background modeling, establishing a Gaussian model for a clean background acquired by a camera, comparing an acquired video with a background model, providing a probability value which belongs to the foreground or the background for each pixel of a scene through depth information acquired by a binocular camera, and dividing the foreground and the background of the scene through an image cutting algorithm. When the divided foreground is the upper torso of the human body, a basic skeleton model of the human body is acquired by refining a foreground profile, de-noising and determining a key point so as to complete the motion capture process.

Description

A kind of based on the motion capture method of background modeling to the binocular vision image
Technical field
The invention belongs to computer vision technique and interactive digital entertainment field, relate to a kind of background segment that binocular camera shooting head and background modeling technology finish and motion-captured process utilized.
Background technology
Movement capturing technology refers to computer vision or other means utilized, and can capture the motion process of human body in real time, exactly.Along with the development of computer software and hardware and the raising of computer user's demand, movement capturing technology is more obvious in the effect of the inside, fields such as digital entertainment, video monitoring, motion analysis.
Yet the development of movement capturing technology also is subjected to the restriction of various conditions and many limitation occur.Such as problems such as blocking of variation, complicated background and the motion process of light.These factors make motion-captured process become more difficult.Yet, carry out the result of background segment by the method for utilizing binocular vision, the prospect in scene has only under the prerequisite of human body, and motion-captured problem will be converted into the prospect profile problem of analyzing scene, make calculated amount simplify greatly.Simultaneously, in the interactive digital entertainment field, movement capturing technology also is the research focus of man-machine interaction in playing in recent years as a kind of video interactive technology.And camera become the general outfit of PC, and man-machine interaction mode general, immersion more and more becomes the focus of digital entertainment research.So, have the research prospect of application fields based on the binocular vision movement capturing technology of background segment technology.
Summary of the invention
Cutting apart of prospect that the objective of the invention is to utilize the binocular camera shooting head to obtain scene and background, finish motion-captured process simultaneously on this basis.This method is at first trained clean background, gathers the background picture of certain frame number, finishes the foundation of background model.On this basis, utilize the new image of gathering, finish the foundation that figure cuts network chart with the color distortion of background model and the depth information of binocular vision, and the method for utilizing dynamic figure to cut, scene prospect and background are cut apart.On the basis of cutting apart, the human body of prospect is carried out the analysis of structure simultaneously, obtain the location of upper body trunk various piece, thereby finish motion-captured process.
For achieving the above object, the invention provides based on background modeling the motion capture method of binocular vision image comprised that step is as follows:
Step S1: with binocular camera shooting head stationkeeping, close white balance, obtain the binocular vision image;
Step S2: to the binocular vision image that obtains, under the clean background image of setting frame number, carry out background modeling, obtain background model;
Step S3: the binocular depth information that utilizes computer binocular vision to obtain, calculating pixel belongs to the probability of prospect and background;
Step S4: utilize binocular depth information and background modeling data and dynamic figure to cut algorithm, binocular vision display foreground and background are cut apart, and extracted prospect profile;
Step S5: prospect profile carries out refinement, determines the human body key point, finishes motion-captured.
Good effect of the present invention:
The present invention utilizes computer vision and image processing techniques, isolates the human body of prospect naturally from scene, and finishes the motion-captured of upper body trunk, thereby realizes the man-machine interaction of nature.The characteristics of traditional interactive mode are based on the hand contact, as mouse, keyboard etc.Development along with computer vision technique, increasing system is by the process of having finished man-machine interaction of the method nature of camera, the user can experience the enjoyment of man-machine interaction more easily by the mode of vision, simultaneously, interface as recreation makes the game player obtain more feeling of immersion.
In addition, the present invention has utilized the collection of binocular vision and the foundation of background model.The employing of binocular vision mainly is to have made full use of depth information, and the prospect of considering often belongs to from the nearer zone of camera, the problem of avoided by shade simultaneously, blocking the segmentation errors that causes.In addition, setting up background model can be so that the cost of cutting apart better obtains calculating, the method for utilizing dynamic figure to cut simultaneously, make cut apart quicker.
Description of drawings
Figure 1A is overall flow figure of the present invention;
Fig. 1 is a binocular vision image of the present invention;
Fig. 2 is left figure and right figure and the parallax that the present invention utilizes binocular vision to obtain;
Fig. 3 is that figure of the present invention cuts the max-flow of algorithm or the network flow graph of minimal cut;
Fig. 4 is a process flow diagram of the present invention;
Fig. 5 is one group of design sketch that video background is cut apart of the present invention;
Fig. 6 is a background segment result's of the present invention edge smoothing synoptic diagram;
Fig. 7 is the refinement of profile of the present invention and the result that key position extracts.
Embodiment
Below in conjunction with accompanying drawing the present invention is described in detail, described embodiment only is intended to be convenient to the understanding of the present invention, and it is not played any qualification effect.
Further specify a kind of operating process of the motion capture method based on background modeling below by example.
All codes of this example are C++ and write, and move under Microsoft visual studio 2005 environment, can also adopt other software and hardware conditions, do not repeat them here.
Figure 1A illustrates and the present invention is based on the overall flow figure of background modeling to the motion capture method of binocular vision image.
The present invention is based on the motion capture method of background modeling to the binocular vision image, based on binocular vision and background segment, its method comprises that step is as follows:
Step S1: with binocular camera shooting head stationkeeping, close white balance, obtain the binocular vision image;
Step S2: to the binocular vision image that obtains, under the clean background image of setting frame number, carry out background modeling, obtain background model;
Step S3: the binocular depth information that utilizes computer binocular vision to obtain, calculating pixel belongs to the probability of prospect and background;
Step S4: utilize binocular depth information and background modeling data and dynamic figure to cut algorithm, binocular vision display foreground and background are cut apart, and extracted prospect profile;
Step S5: prospect profile carries out refinement, determines the human body key point, finishes motion-captured.
Comprise according to obtaining the binocular vision image step described in the step S2:
Step S211: guarantee the stationkeeping of camera, do not have tangible light and shade to change in the scene;
Step S212: close the Automatic white balance of camera, in the hardware parameter of camera, the function of automatic exposure parameter and Automatic white balance is arranged generally, so that when scene light changes, realize regulating automatically the function of picture quality; In background modeling, need to set white balance parameter and fix;
Step S213: gather the fixedly clean background image of frame number (100 frame), be stored in the internal memory.
Comprise as follows according to the step of under the clean background image of setting frame number, carrying out background modeling described in the step S2:
Step S221: utilize the Gaussian Background model to gather the coloured image of each frame in the binocular vision image, use R respectively, G, B represent red, and green and blue three-channel value, span are 0~255;
Step S221: obtained N image in the background modeling process, each image comprises 320 * 240 pixels, calculate the brightness I of each pixel and colourity (r, g).Wherein, r=R/ (R+G+B), g=G/ (R+G+B), R, G, B represent the redness in the Color Channel respectively, the value of green and blue component;
Step S221: the fusion background model of setting up pixel scale; Calculate the brightness and average and the variance of colourity in N image of each pixel, and deposit internal memory in;
Step S221: set up feature background model at brightness space, chrominance space set up based on the colourity model.Deposit the colourity obtained and the background model in the brightness space in internal memory.
According to the depth data cost of each pixel in the binocular vision image described in the step S2 is calculated, obtain the degree of depth cost of each pixel, thereby the binocular depth information is introduced, concrete steps comprise:
Step 231: gather and preservation binocular vision image, be designated as left image and right image respectively;
Step 232: set a depth value for each pixel of left image, described depth value is represented with the parallax of left image and right image;
Step 233:, calculate the difference cost of left image and right image at each depth value;
Step 234: add up the cost value in the left image, and the cost value in the left image is divided into four groups according to the size of described cost value;
Step 235: the cost value of each group is upgraded the prospect of this pixel and the cost of background, and the cost that wherein belongs to prospect reduces according to the exponential relationship to parallax, and the cost of background increases according to the exponential relationship of parallax.
According to utilizing binocular depth information and background modeling data and dynamic figure to cut algorithm described in the step S4, binocular vision display foreground and background to be cut apart, and extracted prospect profile, concrete steps comprise:
Step S41: background modeling reads in the binocular vision image that newly reads after finishing, and described binocular vision image comprises left image and right image;
Step S42: utilize the result of binocular vision data cost acquisition, obtain the data cost of binocular information;
Step S43: utilize background model, compare, obtain to utilize figure to cut algorithm basic principle, set up the network flow of max-flow or minimal cut based on the color cost value with the pixel of left figure;
Step S44: two data cost value utilizing step S42 and step S43 to obtain obtain figure and cut data cost value in the algorithm;
Step S45: utilize the relationship of contrast between the left pixel, the level and smooth item that figure is cut in the algorithm carries out assignment;
Step S46: utilize dynamic figure to cut algorithm, will cut apart based on the video flowing of pixel aspect, segmentation result is divided into two parts, and a part is a prospect, and a part is a background in addition.
Step S47: the prospect background that will cut apart is stored in the picture of identical size according to 0 or 1, and 0 or 1 prospect background picture is obtained edge contour;
Step S48: the mode of utilizing High frequency filter makes the edge more level and smooth the edge denoising;
Step S49: utilize the cut zone of the data error of former frames to proofread and correct.
Utilize picture denoising, refinement mode according to step S5, obtain the key point of trunk, thereby realize that motion-captured effect step comprises:
Step S51: will carry out convergent-divergent through the human body contour outline of aftertreatment;
Step S52: the human body contour outline of convergent-divergent is carried out refinement;
Step S53: the human body contour outline of refinement is enlarged, expand original size to;
Step S54: once more profile is carried out refinement;
Step S55: find neighborhood territory pixel greater than 2 node, and get its center-of-gravity value, be set at the gravity center of human body;
Step S56: search for up and down along center of gravity, find node, be set at head and waist;
Step S57: along about center of gravity, search for, find left arm and right arm, and proportionally determine ancon and shoulder with eccentricity;
Step S58: 9 key points will determining compare with former frames, obtain comparatively stable and trunk position accurately.
The first step as shown in Figure 1 is an images acquired.This method adopts the input of binocular vision video.Among the figure, (z) expression is the coordinate of world coordinate system for x, y; (x L, y L) and (x R, y R) pixel coordinate of the same object of expression in left figure and right figure.
(1) mostly the information of Digital Image Processing is two-dimensional signal, and the process information amount is very big.The piece image here with two-dimensional function f (x, y) expression, x wherein, y is a two-dimensional coordinate, f (x, y) expression point (x, colouring information y).Camera is gathered all optical information in the camera lens from the space, these information enter after the computing machine, is converted to the color model that meets computer standard, carries out Digital Image Processing with the program of entering, and guarantees the continuity and the real-time of video.From the image of gathering, each pixel is handled 76800 pixels of 320 * 240 pixels altogether.The initial effect of video of gathering as shown in Figure 1.Project all operations and computing 320 * 240 pixels that all are based on this each frame subsequently.In the binocular vision, same pixel about the position difference of imaging among two figure, and the size of position difference, reflection be the degree of depth of image.Relatively moving of two pixels can be calculated by the coupling of pixel.Method of the present invention is utilized these information, auxiliary finishing cutting apart of prospect and background.As shown in Figure 2, the utilization of binocular information is that the cost that two width of cloth figure mate about usefulness realizes.What wherein P represented is the position of certain pixel in left figure, and P+d represents the position of this pixel in right figure, and what d represented is exactly the parallax (Display) of this pixel.
(2) the present invention is made up of two parts in the process of utilizing the binocular depth information.
Step 1: the coupling cost in that pixel xi calculates is divided into four groups (maximal value of parallax d is set at 32) according to different parallax value:
A group: pixel x iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d>16, represent that this pixel belongs to prospect very much;
B group: pixel x iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d≤16and d>12, represent that this pixel has the very big prospect that may belong to;
C group: pixel x iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d≤12and d>5, represent that this pixel has the very big background that may belong to;
D group: pixel x iThe parallax that mates is most arranged, promptly optimum parallax (Disparity), the degree of mating most) d≤5, represent that this pixel belongs to background very much.
Under such hypothesis, the present invention needs less time that pixel is divided into four groups, rather than each pixel is carried out 32 possible parallaxes suppose.
Step 2: set the suitable data cost value for figure cuts algorithm.Data item of the present invention comprises that respectively this pixel belongs to the cost of prospect or background, uses D respectively i(B) and D i(F) expression.The parallax value of pixel is big more, so it to belong to the possibility of prospect big more, so D i(F) value correspondence reduces D i(B) value is corresponding to be increased.By such corresponding relation, the present invention proposes a corresponding scheme, express with following formula:
D i , t s ( B ) = D i ( B ) + λ t e - d c t , D i , t s ( F ) = D i ( F ) - λ t e - d c t c t
For all t=A, B, C, D, λ t>0.Wherein
Figure BDA0000040221370000073
What represent is the background model data item that incorporates binocular information, belongs to t=A respectively, B, C, four groups of D.D i(B) expression is the background segment data item of monocular vision.λ tBe the parameter of binocular data cost, what i represented is pixel coordinate.
Figure BDA0000040221370000074
What represent is the foreground model data item that incorporates binocular information.The parallax value that d represents (Disparity).c tWhat represent is the parameter of control d.
As shown in Figure 3, figure cut algorithm max-flow or the network flow graph of minimal cut.P wherein, what q represented is two adjacent pixels.Shown in Figure 4 is the process flow diagram that figure cuts algorithm, comprises the assignment of front end and the partitioning portion of rear end.
(3) figure cuts the important component part that algorithm is a background segment, and it to the effect that utilizes the principle of max-flow or minimal cut, the pixel in the image is cut apart according to certain path, and which is calculated belong to prospect or and background respectively.
The segmentation problem of prospect or background in the image can be considered as the binary identified problems in the computer vision field.If pixel i belongs to prospect, the label f of this pixel of mark then i=F, F refers to prospect.In like manner, if this pixel belongs to background, then be labeled as f i=B.Correspond to two-value label problem, label set only comprises two labels.Figure cuts the weighted graph that algorithm constructs and comprises two corresponding with it summit s and t.As shown in Figure 3, among the figure, left figure is the weighted graph G that provides by 3 * 3 original image structure, G=<V, ε 〉, wherein V is a vertex set, is to be called source node S respectively and terminal node T is dimerous by ordinary node and two.Wherein S and the T two-value label of representing prospect and background respectively for summit ε representative be the limit that is connected the summit, the weights size on limit is represented with the thickness of simplification in last figure.
Flow process such as Fig. 4 that dynamic figure cuts.Comprise data item and level and smooth in the energy function, being provided with of they directly affects figure and cuts the final segmentation result of algorithm.That Fig. 5 represents is several groups of Video Segmentation results of the present invention, and wherein left side 3 width of cloth figure are left figure images of doing video in the input video, and right side 3 width of cloth figure are the results after cutting apart.
(4) low-pass filter that designed in the frequency domain of the present invention comes smooth boundary.Along boundary curve C, the boundary curve of the process of edge smoothing of the present invention as shown in Figure 6, the picture left above is represented the input source image, top right plot is represented the result cut apart; The flat prospect of spending that lower-left figure represents or the edge of background, bottom-right graph is represented is result after level and smooth.Point sequence z (i)=[x (i), y (i)] its complex representation form that obtains of sampling at certain intervals is:
z(i)=x(i)+jy(i)
The Fourier transform of discrete z (i) is:
f ( u ) = 1 K Σ i = 0 K - 1 z ( i ) e - j 2 πui / K
In the formula, j, u, K represent complex symbol respectively, frequency and constant term, f (u) is the Fourier transform of z (i), the Fourier that is called the border is described son, is the expression of boundary point sequence in frequency domain.By the Fourier transform theory as can be known, high fdrequency component comprises details, low frequency component decision global shape.Curve is because jagged ability is rough, and high fdrequency component is contained in these rough zones.The HFS of f (u) is carried out filtering just can obtain smooth curve.The present invention defines the high-frequency energy of low frequency energy ratio and filtering 5%:
r ( l ) = Σ u = 0 l | f ( u ) | 2 / Σ k = 0 K - 1 | f ( u ) | 2
Wherein || be modulo operation.Getting the minimum l value that r (l)>0.95 is set up is the cutoff frequency of low-pass filter.Utilize the character of fourier coefficient (
Figure BDA0000040221370000084
Be the conjugate complex number of f).In coefficient f (u), the radio-frequency component of cancellation in from l to the K-1-l scope.Carry out inverse fourier transform again, the part of curve sudden change has obtained smoothly.
Being motion-captured result of the present invention as shown in Figure 7, wherein is two two field pictures of left figure in the video in the left hand view, and the right side is key point and the skeleton that segmentation result has extracted.Key point represents with circle that skeleton is represented with line.
(5) the present invention motion-captured on the basis of cutting apart comprises three steps,
Step 1: the result that will cut apart carries out aftertreatment, obtains level and smooth and stable relatively contour area, cuts apart owing to relate to profile, so the border does not need accurate calculating.Under will not situation, can finish the skeleton motion tracking effect that this paper needs preferably than macroscopic-void.
Step 2: the profile that will cut apart positions, and determines the basic comprising of nine points.Comprising A 1, A 2, A 3Be representative group, A herein 1, A 2, A 3, A 4, A 5, A 6, A 7, A 8, A 9Nine points.A 1, A 2, A 3Represent three points of head and trunk respectively, A 4, A 5, A 6And A 7, A 8, A 9Represent three points of left arm and right arm respectively.
Step 3: the order of nine points being installed the skeleton profile connects, and finishes motion-captured.
The above; only be the embodiment among the present invention, but protection scope of the present invention is not limited thereto, anyly is familiar with the people of this technology in the disclosed technical scope of the present invention; can understand conversion or the replacement expected, all should be encompassed within the protection domain of claims of the present invention.

Claims (6)

1. one kind based on the motion capture method of background modeling to the binocular vision image, is based on the method for binocular vision and background segment, it is characterized in that, comprises that step is as follows:
Step S1: with binocular camera shooting head stationkeeping, close white balance, obtain the binocular vision image;
Step S2: to the binocular vision image that obtains, under the clean background image of setting frame number, carry out background modeling, obtain background model;
Step S3: the binocular depth information that utilizes computer binocular vision to obtain, calculating pixel belongs to the probability of prospect and background;
Step S4: utilize binocular depth information and background modeling data and dynamic figure to cut algorithm, binocular vision display foreground and background are cut apart, and extracted prospect profile;
Step S5: prospect profile carries out refinement, determines the human body key point, finishes motion-captured.
2. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the step of obtaining the binocular vision image described in the step S2 comprises as follows:
Step S211: guarantee the stationkeeping of camera, do not have tangible light and shade to change in the scene;
Step S212: close the Automatic white balance of camera, in the hardware parameter of camera, the function of automatic exposure parameter and Automatic white balance is arranged generally, so that when scene light changes, realize regulating automatically the function of picture quality; In background modeling, need to set white balance parameter and fix;
Step S213: gather fixedly frame number clean background image, be stored in the internal memory.
3. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the step of carrying out background modeling described in the step S2 under the clean background image of setting frame number comprises as follows:
Step S221: utilize the Gaussian Background model to gather the coloured image of each frame in the binocular vision image, use R respectively, G, B represent red, and green and blue three-channel value, span are 0~255;
Step S221: in the background modeling process, obtained N image, each image comprises 320 * 240 pixels, calculate the brightness I of each pixel and colourity (r, g), wherein, r=R/ (R+G+B), g=G/ (R+G+B), R, G, B represents the redness in the Color Channel respectively, the value of green and blue component;
Step S221: the fusion background model of setting up pixel scale; Calculate the brightness and average and the variance of colourity in N image of each pixel, and deposit internal memory in;
Step S221: set up feature background model at brightness space, chrominance space set up based on the colourity model, deposit the colourity obtained and the background model in the brightness space in internal memory.
4. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the depth data cost to each pixel in the binocular vision image described in the step S2 is calculated, obtain the degree of depth cost of each pixel, thereby the binocular depth information is introduced, and concrete steps comprise as follows:
Step 231: gather and preservation binocular vision image, be designated as left image and right image respectively;
Step 232: set a depth value for each pixel of left image, described depth value is represented with the parallax of left image and right image;
Step 233:, calculate the difference cost of left image and right image at each depth value;
Step 234: add up the cost value in the left image, and the cost value in the left image is divided into four groups according to the size of described cost value;
Step 235: the cost value of each group is upgraded the prospect of this pixel and the cost of background, and the cost that wherein belongs to prospect reduces according to the exponential relationship to parallax, and the cost of background increases according to the exponential relationship of parallax.
5. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: the described binocular depth information and background modeling data and dynamic figure of utilizing of step S4 cut algorithm, binocular vision display foreground and background are cut apart, and the extraction prospect profile, concrete steps comprise as follows:
Step S41: background modeling reads in the binocular vision image that newly reads after finishing, and described binocular vision image comprises left image and right image;
Step S42: utilize the result of binocular vision data cost acquisition, obtain the data cost of binocular information;
Step S43: utilize background model, compare, obtain to utilize figure to cut algorithm basic principle, set up the network flow of max-flow or minimal cut based on the color cost value with the pixel of left figure;
Step S44: two data cost value utilizing step S42 and step S43 to obtain obtain figure and cut data cost value in the algorithm;
Step S45: utilize the relationship of contrast between the left pixel, the level and smooth item that figure is cut in the algorithm carries out assignment;
Step S46: utilize dynamic figure to cut algorithm, will cut apart based on the video flowing of pixel aspect, segmentation result is divided into two parts, and a part is a prospect, and a part is a background in addition;
Step S47: the prospect background that will cut apart is stored in the picture of identical size according to 0 or 1, and 0 or 1 prospect background picture is obtained edge contour;
Step S48: the mode of utilizing High frequency filter makes the edge more level and smooth the edge denoising;
Step S49: utilize the cut zone of the data error of former frames to proofread and correct.
6. according to claim 1 based on the motion capture method of background modeling to the binocular vision image, it is characterized in that: utilize picture denoising, refinement mode, obtain the key point of trunk, thereby realize that motion-captured effect step comprises as follows:
Step S51: will carry out convergent-divergent through the human body contour outline of aftertreatment;
Step S52: the human body contour outline of convergent-divergent is carried out refinement;
Step S53: the human body contour outline of refinement is enlarged, expand original size to;
Step S54: once more profile is carried out refinement;
Step S55: find neighborhood territory pixel greater than 2 node, and get its center-of-gravity value, be set at the gravity center of human body;
Step S56: search for up and down along center of gravity, find node, be set at head and waist;
Step S57: along about center of gravity, search for, find left arm and right arm, and proportionally determine ancon and shoulder with eccentricity;
Step S58: 9 key points will determining compare with former frames, obtain comparatively stable and trunk position accurately.
CN 201010602544 2010-12-23 2010-12-23 Motion capture method for binocular vision image based on background modeling Expired - Fee Related CN102034247B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN 201010602544 CN102034247B (en) 2010-12-23 2010-12-23 Motion capture method for binocular vision image based on background modeling

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN 201010602544 CN102034247B (en) 2010-12-23 2010-12-23 Motion capture method for binocular vision image based on background modeling

Publications (2)

Publication Number Publication Date
CN102034247A true CN102034247A (en) 2011-04-27
CN102034247B CN102034247B (en) 2013-01-02

Family

ID=43887100

Family Applications (1)

Application Number Title Priority Date Filing Date
CN 201010602544 Expired - Fee Related CN102034247B (en) 2010-12-23 2010-12-23 Motion capture method for binocular vision image based on background modeling

Country Status (1)

Country Link
CN (1) CN102034247B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102184008A (en) * 2011-05-03 2011-09-14 北京天盛世纪科技发展有限公司 Interactive projection system and method
CN102927652A (en) * 2012-10-09 2013-02-13 清华大学 Intelligent air conditioner control method based on positions of indoor persons and objects
CN103826071A (en) * 2014-03-11 2014-05-28 深圳市中安视科技有限公司 Three-dimensional camera shooting method for three-dimensional identification and continuous tracking
CN104243951A (en) * 2013-06-07 2014-12-24 索尼电脑娱乐公司 Image processing device, image processing system and image processing method
CN105374043A (en) * 2015-12-02 2016-03-02 福州华鹰重工机械有限公司 Method and device of background filtering of visual odometry
CN105516579A (en) * 2014-09-25 2016-04-20 联想(北京)有限公司 Image processing method and device and electronic equipment
CN106056056A (en) * 2016-05-23 2016-10-26 浙江大学 Long-distance non-contact luggage volume detection system and method thereof
CN109064511A (en) * 2018-08-22 2018-12-21 广东工业大学 A kind of gravity center of human body's height measurement method, device and relevant device
CN109214996A (en) * 2018-08-29 2019-01-15 深圳市元征科技股份有限公司 A kind of image processing method and device
CN110490877A (en) * 2019-07-04 2019-11-22 西安理工大学 Binocular stereo image based on Graph Cuts is to Target Segmentation method
CN111567036A (en) * 2017-12-07 2020-08-21 微软技术许可有限责任公司 Video capture system and method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040125207A1 (en) * 2002-08-01 2004-07-01 Anurag Mittal Robust stereo-driven video-based surveillance
US20070031037A1 (en) * 2005-08-02 2007-02-08 Microsoft Corporation Stereo image segmentation
CN101344965A (en) * 2008-09-04 2009-01-14 上海交通大学 Tracking system based on binocular camera shooting
CN101389004A (en) * 2007-09-13 2009-03-18 中国科学院自动化研究所 Moving target classification method based on on-line study

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040125207A1 (en) * 2002-08-01 2004-07-01 Anurag Mittal Robust stereo-driven video-based surveillance
US20070031037A1 (en) * 2005-08-02 2007-02-08 Microsoft Corporation Stereo image segmentation
CN101389004A (en) * 2007-09-13 2009-03-18 中国科学院自动化研究所 Moving target classification method based on on-line study
CN101344965A (en) * 2008-09-04 2009-01-14 上海交通大学 Tracking system based on binocular camera shooting

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
《IEEE Transactions on Pattern Analysis and Machine Intelligence》 20060930 Vladimir Kolmogorov et al. Probabilistic Fusion of Stereo with Color and Contrast for Bilayer Segmentation 全文 1-6 第28卷, 第9期 2 *
《LNCS》 20091231 Xiaoyu Wu et al. Video Background Segmentation Using Adaptive Background Models 全文 1-6 第5716卷, 2 *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102184008A (en) * 2011-05-03 2011-09-14 北京天盛世纪科技发展有限公司 Interactive projection system and method
CN102927652B (en) * 2012-10-09 2015-06-24 清华大学 Intelligent air conditioner control method based on positions of indoor persons and objects
CN102927652A (en) * 2012-10-09 2013-02-13 清华大学 Intelligent air conditioner control method based on positions of indoor persons and objects
CN104243951B (en) * 2013-06-07 2017-01-11 索尼电脑娱乐公司 Image processing device, image processing system and image processing method
CN104243951A (en) * 2013-06-07 2014-12-24 索尼电脑娱乐公司 Image processing device, image processing system and image processing method
US10293252B2 (en) 2013-06-07 2019-05-21 Sony Interactive Entertainment Inc. Image processing device, system and method based on position detection
CN103826071A (en) * 2014-03-11 2014-05-28 深圳市中安视科技有限公司 Three-dimensional camera shooting method for three-dimensional identification and continuous tracking
CN105516579A (en) * 2014-09-25 2016-04-20 联想(北京)有限公司 Image processing method and device and electronic equipment
CN105374043A (en) * 2015-12-02 2016-03-02 福州华鹰重工机械有限公司 Method and device of background filtering of visual odometry
CN106056056A (en) * 2016-05-23 2016-10-26 浙江大学 Long-distance non-contact luggage volume detection system and method thereof
CN111567036A (en) * 2017-12-07 2020-08-21 微软技术许可有限责任公司 Video capture system and method
CN109064511A (en) * 2018-08-22 2018-12-21 广东工业大学 A kind of gravity center of human body's height measurement method, device and relevant device
CN109064511B (en) * 2018-08-22 2022-02-15 广东工业大学 Method and device for measuring height of center of gravity of human body and related equipment
CN109214996A (en) * 2018-08-29 2019-01-15 深圳市元征科技股份有限公司 A kind of image processing method and device
CN109214996B (en) * 2018-08-29 2021-11-12 深圳市元征科技股份有限公司 Image processing method and device
CN110490877A (en) * 2019-07-04 2019-11-22 西安理工大学 Binocular stereo image based on Graph Cuts is to Target Segmentation method

Also Published As

Publication number Publication date
CN102034247B (en) 2013-01-02

Similar Documents

Publication Publication Date Title
CN102034247B (en) Motion capture method for binocular vision image based on background modeling
CN111797716B (en) Single target tracking method based on Siamese network
CN107204010B (en) A kind of monocular image depth estimation method and system
CN107423698B (en) A kind of gesture estimation method based on convolutional neural networks in parallel
CN102567727B (en) Method and device for replacing background target
CN105160310A (en) 3D (three-dimensional) convolutional neural network based human body behavior recognition method
CN110008832A (en) Based on deep learning character image automatic division method, information data processing terminal
CN109146948A (en) The quantization of crop growing state phenotypic parameter and the correlation with yield analysis method of view-based access control model
CN103455984A (en) Method and device for acquiring Kinect depth image
CN110398720A (en) A kind of anti-unmanned plane detection tracking interference system and photoelectric follow-up working method
CN102184551A (en) Automatic target tracking method and system by combining multi-characteristic matching and particle filtering
CN105869178A (en) Method for unsupervised segmentation of complex targets from dynamic scene based on multi-scale combination feature convex optimization
CN104966286A (en) 3D video saliency detection method
CN113240691A (en) Medical image segmentation method based on U-shaped network
CN112464847B (en) Human body action segmentation method and device in video
CN111179189B (en) Image processing method and device based on generation of countermeasure network GAN, electronic equipment and storage medium
CN111462027B (en) Multi-focus image fusion method based on multi-scale gradient and matting
CN112733950A (en) Power equipment fault diagnosis method based on combination of image fusion and target detection
CN109712247B (en) Live-action training system based on mixed reality technology
CN110263768A (en) A kind of face identification method based on depth residual error network
CN103413323B (en) Based on the object tracking methods of component-level apparent model
CN103440662A (en) Kinect depth image acquisition method and device
CN104063871B (en) The image sequence Scene Segmentation of wearable device
CN111382613A (en) Image processing method, apparatus, device and medium
CN107194948B (en) Video significance detection method based on integrated prediction and time-space domain propagation

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20130102

Termination date: 20151223

EXPY Termination of patent right or utility model