CN104765440B - Hand detection method and equipment - Google Patents
Hand detection method and equipment Download PDFInfo
- Publication number
- CN104765440B CN104765440B CN201410001215.5A CN201410001215A CN104765440B CN 104765440 B CN104765440 B CN 104765440B CN 201410001215 A CN201410001215 A CN 201410001215A CN 104765440 B CN104765440 B CN 104765440B
- Authority
- CN
- China
- Prior art keywords
- hand
- candidate region
- region
- wrist
- current scene
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Abstract
There is provided hand detection method and equipment, the hand detection method includes:Detect the candidate region of hand in current scene;Depth image based on current scene, determines the underarm area in current scene;Corresponding wrist information is predicted respectively using the candidate region of each hand, and predicts the wrist information using the underarm area;With based on wrist information each described, selection confidence level highest candidate region is used as the region where hand.The hand detection method and equipment due to the motion of hand causing in the object for having the similar colour of skin in image blurring, background, interactive process illumination to change, can obtain preferable testing result in the case of hand and the Various Complex such as face is overlapping.
Description
Technical field
This patent disclosure relates generally to man-machine interaction, relate more specifically to hand detection method and equipment.
Background technology
At present, man-machine interaction is from interactive development is touched to by detecting that the gesture and posture of operator is man-machine to carry out
Interaction.Specifically, exactly by capturing the scene image of the operator before screen and screen, and to the scene image of acquisition
Handled, to obtain the action of operator, the action of operator is then converted into the operational order of machine, so as to realize people
Machine is interacted.This man-machine interaction usually requires to recognize the gesture of operator, and the basis of gesture identification is the detection of hand.In view of hand
The characteristic of itself, for example, skin color and the distinctive shape of hand, people are typically based on the color or profile of hand to recognize and examine
Survey hand.
In U.S. Patent application US20100803369A, a kind of human hand detection method of view-based access control model, the party are described
Method requires that two hands occur simultaneously.Specifically, the left and right anaglyph that this method is obtained according to two video cameras calculates scene
Depth information, goes out multiple human hand candidate regions, and the candidate region combination of two to obtaining, by calculating it according to Face Detection
The information such as size is poor, depth difference, alternate position spike draw the positions of both hands.In Japanese patent application JP2005000236876,
Describe a kind of method based on Color images detecting human hand.This method detects multiple human hand candidate regions using features of skin colors
Domain, then calculates the shape of candidate region, and determines whether the region is human hand according to the complexity of shape.In addition, in the U.S.
In patent US7593552B2, a kind of gesture identification method is described, including the detection of human hand.Specifically, in this method
In, face is detected first so as to obtain position and the Skin Color Information of face, then centered on the position of face, in its left and right two
Side takes a region of search respectively, and detects area of skin color in this region, and the focus point of the area of skin color detected is people
Hand position point.But, above-mentioned hand detection method can not all tackle well have the similar colour of skin in motion blur, background object,
Situations such as illumination variation.In addition, needing initiation gesture the above method more.
The content of the invention
According to an aspect of the invention, there is provided a kind of hand detection method, including:Detect the candidate of hand in current scene
Region;Depth image based on current scene, determines the underarm area in current scene;Utilize the candidate region of each hand
Corresponding wrist information is predicted respectively, and predicts the wrist information using the underarm area;With based on wrist each described
Information, selection confidence level highest candidate region is used as the region where hand.
According to another aspect of the present invention there is provided a kind of hand detection device, including:Hand candidate region detection part,
It is configured to detect the candidate region of hand in current scene;Underarm area detection part, is configured to the depth based on current scene
Image is spent, the underarm area in current scene is determined;Wrist information prediction unit, is configured to the candidate using each hand
Corresponding wrist information is predicted in region respectively, and predicts the wrist information using the underarm area;With hand region determining section
Part, is configured to be based on each described wrist information, selection confidence level highest candidate region is used as the region where hand.
Hand detection technique according to embodiments of the present invention can be good at processing because the motion of hand causes the image blurring, back of the body
There is in the object of the similar colour of skin, interactive process illumination change in scape, hand and the Various Complex situation such as face is overlapping.Separately
Outside, hand detection technique according to embodiments of the present invention can carry out the detection of hand on single-frame images, and need not start hand
Gesture and movable information.
Brief description of the drawings
Fig. 1 schematically shows the scene using hand detection technique according to embodiments of the present invention.
Fig. 2 shows the flow chart of hand detection method according to embodiments of the present invention.
Fig. 3 shows the place that the candidate region of hand in current scene is detected in hand detection method according to embodiments of the present invention
The flow chart of reason.
Fig. 4 is shown in the processing shown in Fig. 3 determines shade contrast based on foreground depth image and prospect coloured image
The flow chart of figure, saturation degree contrast figure and Depth contrasts' figure.
Fig. 5, which is shown, calculates shade contrast's figure, saturation degree contrast figure and Depth contrasts' figure in the processing shown in Fig. 3
Weights figure flow chart.
Fig. 6 show in hand detection method according to embodiments of the present invention determine current scene in underarm area it is specific
The flow chart of processing.
Fig. 7(a)Show exemplary foreground depth image, Fig. 7(b)Show schematical depth profile, Fig. 7
(c)Show the depth profile after schematical thresholding processing, Fig. 7(d)Schematically illustrate corresponding to the straight of forearm
Line.
Fig. 8 shows the corresponding wrist of candidate region prediction using hand in hand detection method according to embodiments of the present invention
The flow chart of the processing of information.
Fig. 9 shows the processing for predicting wrist information in hand detection method according to embodiments of the present invention using underarm area
Flow chart.
Figure 10 (a), which schematically shows the wrist predicted in foreground depth image using the candidate region of hand, to be believed
Breath.
Figure 10 (b) schematically shows the wrist information predicted in foreground depth image using underarm area.
Figure 11 shows the functional configuration block diagram of hand detection device according to embodiments of the present invention.
Figure 12 shows the general hardware block diagram of hand detecting system according to embodiments of the present invention.
Embodiment
In order that those skilled in the art more fully understand the present invention, with reference to the accompanying drawings and detailed description to this hair
It is bright to be described in further detail.
Fig. 1 schematically shows the scene using hand detection technique according to embodiments of the present invention.As shown in figure 1, with
Family is just stood in the image pickup scope of video camera 101, and the program in the processing equipment 102 of such as computer is controlled using gesture.
Video camera 101 can be the video camera of any one coloured image that can provide scene and depth image, such as
PrimeSensor, Kinect etc..When user carries out gesture control, video camera 101 is shot to user, processing equipment 102
The every two field picture photographed for video camera 101, detects the positional information of the hand of user.It should be noted that Fig. 1 only schemes
Shown the present invention a kind of possible application scenarios, according to actual conditions, the equipment in application scenarios can correspondingly increase or
Reduce, and with different configurations, or other different equipment can be used.
Below, hand detection technique according to embodiments of the present invention will be described in detail.
Brief description is carried out to the basic thought of the hand detection technique of the present invention first.It is well known that hand is non-rigid
Object, with move it is fast, yielding the features such as.Therefore, causing have the similar colour of skin in image blurring, background due to the motion of hand
Object, hand and the Various Complex such as face is overlapping in the case of, it is difficult to accurately detect the position sold.For this case, this hair
The bright candidate region for proposing multiple hands using forearm information to detecting is verified, so as to obtain the position of accurate hand
Technology.Because forearm volume is larger and with less deformation, therefore its detection is easy relative to the detection of hand, so that
The position of the hand obtained using forearm Information Authentication has higher confidence level.
Fig. 2 shows the flow chart of hand detection method according to embodiments of the present invention.
As shown in Fig. 2 in step S201, detecting the candidate region of hand in current scene.
As it was noted above, detecting candidate region existing many researchs in the art of one or more hands, it is not
The key point of the present invention, those skilled in the art can detect the candidate of hand in current scene in any suitable manner
Region.Herein, it is complete for explanation, the way of example used in the present embodiment will simply be described with reference to Fig. 3.
As shown in figure 3, in step S2011, obtaining the depth image and coloured image for the current scene that video camera is shot.Such as
Well known to a person skilled in the art the value that, depth image is each pixel in image represent in scene certain point and video camera it
Between distance image, and coloured image is then a width RGB image.
In step S2012, based on the depth image and coloured image, the foreground depth image of current scene is obtained with before
Scape coloured image.
Due to not only including user in depth image and coloured image, also include other background image contents, therefore
In order to reduce subsequent detection processing amount of calculation, can be partitioned into by foreground segmentation from depth image and coloured image including
Other in user and/or scene close to the foreground area of the object of video camera, that is, obtain current scene foreground depth image and
Prospect coloured image.This processing can be realized by various existing methods in this area, such as connected domain analysis, threshold value
Segmentation etc., is no longer described in detail herein.
In subsequent step S2013-S2016, by from the tone of hand, saturation degree and depth information, Utilization prospects
The salient region that depth image and prospect Color images detecting go out in foreground area, is used as the candidate region of hand.Due to hand
Color and most of objects have a distinction, thus color contrast can as distinguish hand and most of objects feature it
One.Moreover, tone and saturation degree by handling color respectively, the object and hand included for foreground area have similar color
Situation, can also accurately distinguish and sell.On tone and saturation degree component, can by by coloured image from RGB
Color space conversion is obtained to hsv color space, and this conversion belongs to prior art, does not carry out detailed retouch to it herein
State.On the other hand, due in man-machine interactive operation, not having other objects in the distance of hand to video camera, thus hand is in depth
On be significant, it can be used as the another feature for distinguishing hand and other objects.Therefore, the detection of salient region includes color
The calculating and fusion of tune, saturation degree and the aspect of depth three.This is described below in conjunction with the specific processing in step S2013-S2016
One salient region detection process.
Specifically, in step S2013, based on the foreground depth image and prospect coloured image, determining shade contrast
Scheme CT, saturation degree contrast figure CSScheme C with Depth contrastsD。
Shade contrast's degree figure CTIt is to be obtained by the tone path computation of prospect coloured image, saturation degree contrast figure CSIt is
Obtained by the saturation degree path computation of prospect coloured image, and Depth contrasts' figure CDIt is then to be calculated by foreground depth image
Arrive, wherein in each contrast figure the pixel value of each pixel represent the pixel relative in image other pixels it is aobvious
Work value.Shade contrast schemes CT, saturation degree contrast figure CSScheme C with Depth contrastsDIt can be calculated using same method, example
As using below in conjunction with the method described in Fig. 4.That is, by performing the method shown in three times Fig. 4, calculating respectively
Go out tone comparison diagram CT, saturation degree contrast figure CSScheme C with Depth contrastsD.To have below to the method shown in Fig. 4
The explanation of body.For convenience of description, input picture is represented with I in the following description, and corresponding contrast figure then uses Cd(d=
D, T, S) represent.It is understood that when calculating shade contrast's figure, I represents the tone channel image of prospect coloured image, in meter
When calculating saturation degree contrast figure, I represents the saturation degree channel image of prospect coloured image, when calculating Depth contrasts' figure, I tables
Show foreground depth image.
As shown in figure 4, in step S401, its neighborhood territory pixel j (j=1...n are selected to image I each pixel ii),
Wherein niIt is pixel i neighborhood territory pixel number.
In this step, neighborhood territory pixel can be selected in any suitable manner.A kind of possible mode is to use
The multi-density method of sampling.The so-called multi-density method of sampling be exactly range pixel i it is nearer position sampling neighborhood territory pixel more
It is many, in the fewer of position sampling more remote range pixel i.Specifically, m deciles direction is chosen by origin of pixel i.At this
On m direction, sampled respectively by step-length of r, until image I border.M value can be come really according to specific needs
Fixed, for example, m value takes usually 8 in an experiment, and if it is desired to obtaining more accurate result, m value can be bigger(Such as
16), or if not high to the accuracy requirement of result, m value can also be smaller(Such as 4).Step-length r value can also root
Determined according to specific needs, such as r value can be 2 pixel distances.By this sampling, prospect cromogram is corresponded to respectively
The tone channel image of picture, the saturation degree channel image of prospect coloured image and foreground depth image, obtain adopting in each image
Tone value, intensity value and the depth value of sampling point.Optionally, in this step can only for each non-zero value pixel i
Its neighborhood territory pixel is selected, to reduce amount of calculation.
In step S402, to each pixel i as origin in image I, calculate itself and corresponding neighborhood territory pixel j it
Between pixel value difference.
In this step, it is possible to use expression formula [1] calculates each pixel i as origin and corresponding neighborhood territory pixel
Margin of image element d between jij。
dij=|Ii-Ij|2,i=1...N [1]
Wherein, IiIt is pixel i pixel value, IjIt is neighborhood territory pixel j pixel value, N is image I size.More specifically,
Calculating Depth contrasts' figure CDWhen, IiAnd IjRepresent pixel i and j depth value;Calculating shade contrast's degree figure CTWhen, IiAnd Ij
Pixel i and j shade of color value are represented respectively;And calculating saturation degree contrast figure CSWhen, IiAnd IjPixel i and j are represented respectively
Color saturation value.
In step S403, for each neighborhood territory pixel j, its weights is determined.
In this step, for example, it is possible to use expression formula [2] calculates neighborhood territory pixel j Gauss weight wij。
Wherein, σpThe scale factor of Gauss weights, such as can be using value as 0.25, p in testingiAnd pjIt is pixel i respectively
With j position.||pi-pj| | represent position piAnd pjEuclidean distance.Apart from more remote neighborhood it can be seen from expression formula [2]
Pixel, its weights are smaller, and apart from nearer neighborhood territory pixel, its weights is bigger.
Then, in step S404, to each pixel i as origin in image I, its contrast value is determined, to obtain
Contrast figure Cd。
In this step, for example, it is possible to use expression formula [3] calculates in image I each pixel i as origin pair
Compare angle value.
By each pixel in each for three kinds of foreground image I, above-mentioned steps S401-S404 is performed, can be with
Obtain three kinds of contrast figures, i.e. shade contrast's figure CT, saturation degree contrast figure CSScheme C with Depth contrastsD.As it was previously stated, each right
Pixel value than each pixel in degree figure represents saliency value of the pixel relative to other pixels of image, because hand exists
In terms of tone, saturation degree and depth all than other objects in scene significantly, so the bigger pixel of contrast figure intermediate value belongs to
The probability of hand is higher.
Fig. 3 is returned, in step S2014, shade contrast's figure C is calculated respectivelyT, saturation degree contrast figure CSAnd depth correlation
Degree figure CDWeights figure WT、WSAnd WD。
In the present embodiment, by calculating contrast figure CD、CTAnd CSBallot each other obtains weights figure WT、WSAnd
WD.Herein, ballot is a kind of description to difference between contrast figure.The pixel value of weights figure represents its corresponding contrast figure
Confidence level, the pixel value in weights figure is bigger, then corresponding contrast figure is more credible.Below, will be with reference to Fig. 5 to according to this reality
The processing for applying the calculating weights figure of example is described.
As shown in figure 5, in step S501, for contrast figure CD,CT,CS, calculate corresponding gradient vector figure Gd。
Gradient vector figure GdIt is two tuple (Dd,Md) (d=D, T, S), wherein, DdIt is gradient direction, MdIt is gradient magnitude.It is logical
Cross and calculate gradient for each pixel in contrast figure, gradient vector figure corresponding with contrast figure can be generated.And calculate
The gradient of pixel is common technical means in the art, therefore herein without being described in detail.
In step S502, for every width contrast figure, other each ballots of contrast figure to it are calculated.
For the ease of description, hereinafter, C is usedd(d=D, T, S) represents three width contrast figure CD,CT,CSIn it is any right
Than degree figure, C is usedc(c=D, T, S, and c ≠ d) represents to be different from CdAnother contrast figure.CcTo CdBallot describe in vacation
If CcUnder the conditions of correct, CdIt is also correct probability.
In this step, C is calculated firstdIn CcFor wrong probability under the conditions of correct.Generally, if CcIt is correct
In the case of CdIt is wrong, then the direction of their gradient vector is inevitable different, there is angle between two vectors.According to
The Vector triangle of amount, two vectorial differences are the length on the side corresponding to two vector angles.Therefore, CdIn CcFor correct bar
Wrong probability can be calculated by expression formula [4] under part.
c,d=D,T,S;c≠d.
Wherein,Represent CdBe it is wrong,Represent CcIt is correct, θ is gradient vector GcAnd GdAngle, F be used for should
To the situation that two vector angles are obtuse angle.
Then, C is calculated by expression formula [5]cTo CdBallot,
The C it can be seen from expression formula [5]cBe it is correct in the case of CdIt is that wrong probability is higher, CcTo CdBallot get over
It is small.Each pixel that processing in step S502 is directed in every width contrast figure is carried out one by one.
In step S503, based on the ballot, the weights figure of every width contrast figure is calculated.
Specifically, the weights figure of each contrast figure can be calculated by expression formula [6] in this step.
Wherein, Wd(d=D, T, S) is contrast figure CdWeights figure.Each contrast figure it can be seen from expression formula [6]
Weights figure be remaining summation of contrast figure to the ballot of the contrast figure.Processing in step S503 is directed to every width weights
Each pixel in figure is carried out one by one.
Thus, shade contrast's figure C is obtainedT, saturation degree contrast figure CSScheme C with Depth contrastsDWeights figure WT、WSAnd
WD.Optionally, in order to handle conveniently, it is possible to use expression formula [7] is to weights figure WT、WSAnd WDIt is normalized
Wherein, W'd(d=D, T, S) is normalized weights figure.
Referring back to Fig. 3, in step S2015, based on the contrast figure and corresponding weights figure, conspicuousness is obtained
Figure.
Specifically, as shown in expression formula [8], in this step, can be contrasted by using corresponding weights figure to each
Degree figure is weighted summation, obtains Saliency maps SM.
Saliency maps SM is to consider the conspicuousness description obtained after tone, saturation degree and the aspect of depth three, and this is notable
Pixel value in property figure represents that respective pixel belongs to the probability in human hand region.
Then, in step S2016, the candidate region of hand in current scene is determined based on the Saliency maps.
In this step, the candidate region of hand can be determined by carrying out binary conversion treatment to the Saliency maps.Tool
Body, such as shown in expression formula [9], binaryzation can be carried out to each pixel in Saliency maps using predetermined threshold alpha
Processing.
Wherein, the one or more candidate regions for each pixel formation hand that H values are 1.So far, current scene can be obtained
The candidate region of middle hand, can also equally obtain the candidate region of hand in the foreground depth image of current scene.
In addition, it is optional, can be for example by expression formula [10], for the candidate region of each hand thereby determined that, meter
The probability that it is the region where hand is calculated, for follow-up processing.
GH(t)=average(SM(t)) [10]
Wherein t is the candidate region of hand, and SM (t) is the significance value for each pixel being located in Saliency maps in the t of region.
According to expression formula [10], each candidate region is that the probability in the region where hand is the conspicuousness of each pixel included in the region
The average value of value.
Simply describe using the tone of hand, saturation degree and depth information to detect the time of hand in current scene above
The way of example of favored area.As it was previously stated, in fact, those skilled in the art can be examined using other any appropriate modes
Astronomical observation favored area, the detection for example with the candidate region of the hand based on the colour of skin, the detection of the candidate region of the hand based on shape
Deng.
Fig. 2 is returned to, in step S202, the depth image based on current scene determines the underarm area in current scene.
Fig. 6 shows the underarm area determined according to the depth image based on current scene of the present embodiment in current scene
Specific processing flow chart.
As shown in fig. 6, in step S601, based on the foreground depth image of the depth image acquisition by current scene, generation
Depth profile.
Specifically, in this step, for each pixel i in foreground depth image, being set respectively centered on it
One interior zone and a perimeter, wherein perimeter include interior zone, and according to such as following expression [11]-
[14], the depth profile of the pixel is calculated.
G(i)=exp(-(RI+RO)) [11]
T=min(DI)+ε [14]
Wherein, DIRepresent the depth value matrix of interior zone, min (DI) be pixel in interior zone minimum depth value
(In the present embodiment, the distance of agreement pixel to video camera is nearer, then the depth value of the pixel is bigger), ε is constant value, and it can
Set with as needed.RIRepresent in region internally, depth value is more than the number C of T pixelIAccount for the total pixel of interior zone
Number NIRatio.RORepresent to remove in the outer region in the region outside interior zone, depth value is more than of T pixel
Number COAccount for total number of pixels N in the region in the perimeter outside removing interior zoneORatio.
By the way that the depth profile of each pixel i in foreground depth image is calculated as above, depth profile is formed.Example
Such as, Fig. 7(a)Show exemplary foreground depth image, Fig. 7(b)Show the schematical depth profile to be formed.
It is understood that it is described above by for each pixel i set an interior zone and a perimeter come
The mode for calculating depth profile is only an example.Those skilled in the art can count by the way of any other is appropriate
The depth profile of pixel is calculated, for example, can obtain the distribution situation of pixel value by calculating the histogram of depth image, so that
Obtain depth profile.
In step S602, thresholding processing is carried out to the depth profile.
It is it is well known in the art that being not be described in detail herein that thresholding processing is how carried out to image.Fig. 7(c)Show
The depth profile gone out after schematical thresholding processing.
The detection of straight lines on step S603, the depth profile after thresholding processing.
The method that a variety of known detection of straight lines from image are there are in this area, such as hough transform, herein
Without being described in detail.
In step S604, based on the position of human body in the foreground depth image, determine to correspond to forearm in the straight line
Straight line.
Position of the human body in foreground depth image can be come for example, by any appropriate method such as head and shoulder Model Matching
It is determined that, it is not be described in detail herein., can be according to human figure structure, it is determined that in step after the position of human body is determined
Correspond to the straight line of forearm in a plurality of straight line detected in S603.Fig. 7(d)Schematically illustrate corresponding to the straight of forearm
Line.
Optionally, can be for example by expression formula [15], for the straight line corresponding to forearm thereby determined that, calculating it is
The probability of forearm, for follow-up processing.
GL=average(G(j)) [15]
Wherein j is the pixel being located in the linearity region, and G (j) is as described previously by being somebody's turn to do that expression formula [11] is calculated
The depth profile of each pixel j in linearity region.According to expression formula [15], identified straight line is that the probability of forearm is this
The average value of the depth profile of each pixel included in linearity region.
In step S605, according to the straight line corresponding to forearm, underarm area is determined in the foreground depth image.
, can be for example, by region after detecting the straight line corresponding to forearm in depth profile after thresholding processing
Any appropriate methods such as growth determine underarm area in foreground depth image.
More than, describe the tool that the depth image based on current scene determines the underarm area in current scene with reference to Fig. 6
Body processing.It should be appreciated that this is only a kind of example, those skilled in the art can be using any other appropriate method come really
Underarm area is determined, such as the forearm detection based on kinematic constraint, the arm detection based on joint variable.
Fig. 2 is returned to, in step S203, corresponding wrist information is predicted respectively using the candidate region of each hand,
And predict the wrist information using the underarm area.
In this step, for the candidate region of each hand detected, be based on the candidate region predict it is corresponding
Wrist information, and for identified underarm area, corresponding wrist information is predicted also based on the underarm area.It is optional
, the prediction steps can be carried out in foreground depth image.Below, it will be described in detail with reference to Fig. 8 and 9.
Fig. 8 shows that the method for corresponding wrist information is predicted in the candidate region of utilization hand according to embodiments of the present invention
Flow chart.
As shown in figure 8, in step S801, for the candidate region of any one hand, determine the hand candidate region it is outer
Rectangle is connect, and based on the principal direction of boundary rectangle calculating hand.
In this step, the principal direction of the boundary rectangle is calculated, the principal direction of hand is used as.Principal direction can pass through this area
In known a variety of methods determine, in the present embodiment, use PCA(PCA)To calculate the master of the boundary rectangle
Direction.Figure 10 (a) schematically shows the wrist information predicted in foreground depth image using the candidate region of hand.Such as
Shown in Figure 10 (a), the solid-line rectangle frame in figure is the boundary rectangle of the candidate region of hand, and arrow A represents the principal direction of hand.
In step S802, on the direction opposite with the principal direction of the hand, the first auxiliary region of a predefined size is set
Domain, first auxiliary area connects with the boundary rectangle of the candidate region of the hand.
First auxiliary area is the region that prediction includes wrist.The size of first auxiliary area can be according to wrist and hand
Limbs proportionate relationship preset.For example, can by wide twice of the wide boundary rectangle for being set to hand of the first auxiliary area,
High setting is the high half of the boundary rectangle of hand.As shown in Figure 10 (a), the dotted rectangle in figure is first auxiliary region
Domain.
In step S803, the center of gravity of first auxiliary area is determined, the position of the wrist of prediction is used as.
In this step, using the center of gravity of the first auxiliary area as wrist position Jhl(x,y).As shown in Figure 10 (a), figure
In " * " mark represent the position of wrist thereby determined that.
In step S804, the principal direction of first auxiliary area is calculated, the direction of the wrist of prediction is used as.
In the present embodiment, it is similar with step S801, also using PCA(PCA)To calculate first auxiliary
The principal direction of rectangle, is used as the direction θ of the wrist of predictionh.As shown in Figure 10 (a), the arrow B in figure represents the wrist of prediction
Direction.
Thus, wrist information corresponding with the candidate region of a hand has been predicted.Pass through the candidate regions for each hand
Domain repeats the above steps S801-S804 processing, can be directed to the candidate region of each hand, predict corresponding wrist information.
Fig. 9 shows the flow chart of the method for prediction wrist information in utilization underarm area according to embodiments of the present invention.
As shown in figure 9, in step S901, determining the boundary rectangle of underarm area, and forearm is calculated based on the boundary rectangle
Principal direction.
In this step, the principal direction of the boundary rectangle is calculated, the principal direction of forearm is used as.As it was previously stated, principal direction can
To be determined by a variety of methods as known in the art, in the present embodiment, PCA is used(PCA)To calculate this
The principal direction of boundary rectangle.Figure 10 (b) schematically shows the hand predicted in foreground depth image using underarm area
Wrist information.As shown in Figure 10 (b), the solid-line rectangle frame in figure is the boundary rectangle of underarm area, and arrow A represents the main side of forearm
To.
In step S902, on the principal direction identical direction with the forearm, the second of one predefined size of setting is auxiliary
Region is helped, second auxiliary area connects with the boundary rectangle of underarm area this described.
Second auxiliary area is the region that prediction includes wrist.The size of second auxiliary area can according to wrist with it is small
The limbs proportionate relationship of arm is preset.For example, the width of the second auxiliary area can be disposed as to the external square of underarm area
Wide twice of shape, high setting for the boundary rectangle of underarm area width.As shown in Figure 10 (b), the dotted rectangle in figure is institute
State the second auxiliary area.
In step S903, the center of gravity of second auxiliary area is determined, the position of the wrist of prediction is used as.
In this step, using the center of gravity of the second auxiliary area as wrist position Jlh(x,y).As shown in Figure 10 (b), figure
In " * " mark represent the position of wrist thereby determined that.
In step S904, the principal direction of second auxiliary area is calculated, the direction of the wrist of prediction is used as.
In the present embodiment, it is similar with step S901, also using PCA(PCA)To calculate second auxiliary
The principal direction of rectangle, is used as the direction θ of the wrist of predictionl.As shown in Figure 10 (b), the arrow B in figure represents the wrist of prediction
Direction.
Fig. 2 is returned to, in step S204, based on wrist information each described, selection confidence level highest candidate region is made
For the region where hand.
In this step, for identified each hand candidate region, believed according to the wrist predicted using the candidate region
Breath and the wrist information predicted using underarm area, calculate the confidence level of the candidate region, and select with highest confidence level
Candidate region is used as the region where hand.
The confidence level of candidate region can using it is various it is appropriate by the way of calculate.For example, a kind of possible mode is pair
In the candidate region of each hand, the probability and the straight line corresponding to forearm for by the candidate region being region where hand are forearms
Probability is weighted processing, and using weighted results as the candidate region confidence level.Specifically, can be according to expression formula [16]
To calculate confidence level.
GH|L(t)=GH(t)+R(Ht,L)GL [16]
Wherein, t is the candidate region of hand, GH|L(t) be candidate region t confidence level, GH(t) it is as described previously by table
The candidate region t calculated up to formula [10] is the probability in the region where hand, GLIt is to be calculated as described previously by expression formula [15]
Corresponding to forearm straight line be forearm probability, R (Ht, L) and it is GLWeights.
R(Ht, L) and it is according to the wrist predicted using the candidate region t wrist informations predicted and using underarm area
What information was determined.Optionally, the wrist that the wrist information predicted using underarm area is predicted with the candidate region t using hand
The difference of information is smaller, then R (Ht, L) and it is bigger.Specifically, weights R (H can be determined according to expression formula [17]t,L)。
R(Ht,L)=exp(-E) [17]
E=wJ||Jhl(x,y)-Jlh(x,y)||+wD|θl-θh|
Wherein, θlAnd θhIt is the wrist direction predicted as previously described using the candidate region of underarm area and hand, w respectivelyD
For the weights of the deviation in wrist direction predicted, Jlh(x, y) and Jhl(x, y) be respectively as previously described using underarm area and
The wrist location that the candidate region of hand is predicted, wJFor the weights of the deviation of wrist location predicted.wDAnd wJCan be according to reality
Border situation is set, such as in testing, wJCan be using value as 1/3, wDCan be using value as 2/3.
The foregoing describe a kind of method of the confidence level for the candidate region for determining each hand.It is understood that this method is only
Be an example, and be not limitation of the present invention, the confidence level of candidate region can using other it is any it is appropriate by the way of
To calculate.For example, the wrist information that simply can be predicted according to the candidate region using hand is with utilizing underarm area prediction
The difference of wrist information, to determine the confidence level of each candidate region, i.e., with the difference for the wrist information predicted using underarm area
Not smaller candidate region, its confidence level is bigger.
Optionally, the selected candidate region with highest confidence level can be verified.Specifically, by the highest
Confidence level is compared with threshold value set in advance, if highest confidence level is more than the threshold value, it is determined that put with the highest
The candidate region of reliability is the region where hand;Conversely, then illustrate that the confidence level of the candidate region of each hand detected is too low,
Think to be not detected by hand.
The foregoing describe hand detection method according to embodiments of the present invention.In the method, using forearm message to detection
To the candidate regions of multiple hands verified that the candidate region that the multiple hands detected are calculated using forearm information is put
Reliability, and select that there is the candidate region of highest confidence level in the candidate region of multiple hands as the region where hand.Because small
The detection of arm is easy relative to the detection of hand, therefore the forearm detected has higher confidence level, and then utilizes forearm letter
Ceasing the region of the hand of checking has higher confidence level.Therefore, cause to have in image blurring, background even in due to the motion of hand
Illumination is changed, in the case of hand and the Various Complex such as face is overlapping in the object of the similar colour of skin, interactive process, according to this hair
The hand detection method of bright embodiment can also obtain preferable testing result.In addition, the hand detection method do not need initiation gesture and
Movable information.
Hand detection device according to embodiments of the present invention is described below with reference to Figure 11.
Figure 11 shows the functional configuration block diagram of hand detection device 1100 according to embodiments of the present invention.
As shown in figure 11, hand detection device 1100 can include:Hand candidate region detection part 1110, is configured to detection
The candidate region of hand in current scene;Underarm area detection part 1120, is configured to the depth image based on current scene, really
Determine the underarm area in current scene;Wrist information prediction unit 1130, is configured to the candidate region using each hand
Corresponding wrist information is predicted respectively, and predicts the wrist information using the underarm area;Part is determined with hand region
1140, it is configured to be based on each described wrist information, selection confidence level highest candidate region is used as the region where hand.
Above-mentioned hand candidate region detection part 1110, underarm area detection part 1120, wrist information prediction unit 1130
And hand region determines that the concrete function of part 1140 and operation may be referred to above-mentioned Fig. 1 to Figure 10 associated description, herein not
Repeat description.
The general hardware block diagram of hand detecting system 1200 according to embodiments of the present invention is described below with reference to Figure 12.Such as Figure 12
Shown, hand detecting system 1200 can include:Input equipment 1210, for from the relevant image of outside input or information, such as taking the photograph
Depth image, coloured image that camera is shot etc., the input equipment for example can be such as video camera;Processing equipment 1220, is used
In the above-mentioned hand detection method according to the embodiment of the present invention of implementation, or above-mentioned hand detection device is embodied as, the processing is set
Standby for example can be the central processing unit of computer or other chips with disposal ability etc.;Output equipment 1230, is used
Result obtained by above-mentioned hand detection process is implemented to outside output, the position point coordinates of the hand of such as determination, the direction of hand
Etc., the output equipment for example can be display, printer etc.;And storage device 1240, for volatile or non-easy
The mode of mistake stores various images and data involved by above-mentioned hand detection process, such as depth image, coloured image, foreground depth
Image, prospect coloured image, comparison diagram, weights figure, Saliency maps, using each hand candidate region predict wrist information,
Wrist information, various threshold values set in advance and weights for being predicted using underarm area etc., the storage device for example can be
Random access memory(RAM), read-only storage(ROM), hard disk or semiconductor memory etc. it is various volatile and nonvolatile
Property memory.
The general principle of the present invention is described above in association with specific embodiment, however, it is desirable to, it is noted that to this area
For those of ordinary skill, it is to be understood that the whole or any steps or part of methods and apparatus of the present invention, Ke Yi
Any computing device(Including processor, storage medium etc.)Or in the network of computing device, with hardware, firmware, software or
Combinations thereof is realized that this is that those of ordinary skill in the art use them in the case where having read the explanation of the present invention
Basic programming skill can be achieved with.
Therefore, the purpose of the present invention can also by run on any computing device a program or batch processing come
Realize.The computing device can be known fexible unit.Therefore, the purpose of the present invention can also be included only by offer
Realize that the program product of the program code of methods described or device is realized.That is, such program product is also constituted
The present invention, and the storage medium for such program product that is stored with also constitutes the present invention.Obviously, the storage medium can be
Any known storage medium or any storage medium developed in the future.
It may also be noted that in apparatus and method of the present invention, it is clear that each part or each step are to decompose
And/or reconfigure.These decompose and/or reconfigured the equivalents that should be regarded as the present invention.Also, perform above-mentioned series
The step of processing can order naturally following the instructions perform in chronological order, but and need not necessarily sequentially in time
Perform.Some steps can be performed parallel or independently of one another.
Above-mentioned embodiment, does not constitute limiting the scope of the invention.Those skilled in the art should be bright
It is white, depending on design requirement and other factors, can occur various modifications, combination, sub-portfolio and replacement.It is any
Modifications, equivalent substitutions and improvements made within the spirit and principles in the present invention etc., should be included in the scope of the present invention
Within.
Claims (10)
1. a kind of hand detection method, including:
Detect the candidate region of hand in current scene;
Depth image based on current scene, determines the underarm area in current scene;
Corresponding wrist information is predicted respectively using the candidate region of each hand, and it is described using underarm area prediction
Wrist information;With
Based on wrist information each described, selection confidence level highest candidate region is used as the region where hand.
2. hand detection method as claimed in claim 1, wherein the depth image based on current scene determines current scene
In underarm area include:
Foreground segmentation is carried out to the depth image, to generate foreground depth image;
Based on the foreground depth image, depth profile is generated;
Thresholding processing is carried out to the depth profile;
Detection of straight lines in depth profile after thresholding processing;
Based on the position of human body in the foreground depth image, determine to correspond to the straight line of forearm in the straight line;
According to the straight line corresponding to forearm, underarm area is determined in the foreground depth image.
3. hand detection method as claimed in claim 1 or 2, wherein predicting correspondence respectively using the candidate region of each hand
Wrist information further comprise candidate region for each hand:
The boundary rectangle of the candidate region of the hand is determined, and based on the principal direction of boundary rectangle calculating hand;
On the direction opposite with the principal direction of the hand, the first auxiliary area of a predefined size, first auxiliary region are set
Domain connects with the boundary rectangle of the candidate region of the hand;
The center of gravity of first auxiliary area is determined, the position of the wrist of prediction is used as;
The principal direction of first auxiliary area is calculated, the direction of the wrist of prediction is used as.
4. hand detection method as claimed in claim 3, wherein predicting that the wrist information is further using the underarm area
Including:
The boundary rectangle of the underarm area is determined, and based on the principal direction of boundary rectangle calculating forearm;
On the principal direction identical direction with the forearm, the second auxiliary area of a predefined size is set, second auxiliary
Region connects with the boundary rectangle of underarm area this described;
The center of gravity of second auxiliary area is determined, the position of the wrist of prediction is used as;
The principal direction of second auxiliary area is calculated, the direction of the wrist of prediction is used as.
5. hand detection method as claimed in claim 1, further comprises the candidate region for each hand, calculating it is
The probability in the region where hand.
6. hand detection method as claimed in claim 5, further comprises, for the identified straight line corresponding to forearm, calculating
The straight line is the probability of forearm.
7. hand detection method as claimed in claim 6, wherein being waited based on wrist information selection confidence level highest each described
Favored area includes as the region where hand:
For the candidate region of each hand, the probability for being region where hand is with the straight line corresponding to forearm
The probability of forearm is weighted, and using weighted results as the candidate region confidence level;
Candidate region of the selection with maximum weighted result, is used as confidence level highest candidate region.
8. hand detection method as claimed in claim 7, wherein the wrist information predicted using underarm area is with utilizing hand
The difference for the wrist information that candidate region is predicted is smaller, then the straight line corresponding to forearm is that the weights of the probability of forearm are got over
Greatly.
9. hand detection method as claimed in claim 1, wherein the candidate region of hand includes in detection current scene:Utilize hand
The candidate region of hand in tone, saturation degree and depth information, detection current scene.
10. a kind of hand detection device, including:
Hand candidate region detection part, is configured to detect the candidate region of hand in current scene;
Underarm area detection part, is configured to the depth image based on current scene, determines the underarm area in current scene;
Wrist information prediction unit, is configured to predict corresponding wrist information respectively using the candidate region of each hand,
And predict the wrist information using the underarm area;With
Hand region determines part, is configured to be based on each described wrist information, selects the candidate region conduct of confidence level highest
Region where hand.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410001215.5A CN104765440B (en) | 2014-01-02 | 2014-01-02 | Hand detection method and equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410001215.5A CN104765440B (en) | 2014-01-02 | 2014-01-02 | Hand detection method and equipment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104765440A CN104765440A (en) | 2015-07-08 |
CN104765440B true CN104765440B (en) | 2017-08-11 |
Family
ID=53647331
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410001215.5A Expired - Fee Related CN104765440B (en) | 2014-01-02 | 2014-01-02 | Hand detection method and equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104765440B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105893926A (en) * | 2015-12-15 | 2016-08-24 | 乐视致新电子科技(天津)有限公司 | Hand identification method, system and device |
CN107292904B (en) * | 2016-03-31 | 2018-06-15 | 北京市商汤科技开发有限公司 | A kind of palm tracking and system based on depth image |
CN107958458B (en) * | 2016-10-17 | 2021-01-22 | 京东方科技集团股份有限公司 | Image segmentation method, image segmentation system and equipment comprising image segmentation system |
CN109101860B (en) * | 2017-06-21 | 2022-05-13 | 富泰华工业(深圳)有限公司 | Electronic equipment and gesture recognition method thereof |
CN108255351B (en) * | 2017-12-22 | 2019-08-20 | 潍坊歌尔电子有限公司 | Determination method and device, projector, the optical projection system of user's finger location information |
CN111259757B (en) * | 2020-01-13 | 2023-06-20 | 支付宝实验室(新加坡)有限公司 | Living body identification method, device and equipment based on image |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102778953A (en) * | 2012-06-28 | 2012-11-14 | 华东师范大学 | Motion sensing control method of shadow play remote digital performing based on Kinect |
DE102011075877A1 (en) * | 2011-05-16 | 2012-11-22 | Siemens Aktiengesellschaft | Evaluation method for a sequence of temporally successive depth images |
CN103424105A (en) * | 2012-05-16 | 2013-12-04 | 株式会社理光 | Object detection method and device |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8144148B2 (en) * | 2007-02-08 | 2012-03-27 | Edge 3 Technologies Llc | Method and system for vision-based interaction in a virtual environment |
-
2014
- 2014-01-02 CN CN201410001215.5A patent/CN104765440B/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102011075877A1 (en) * | 2011-05-16 | 2012-11-22 | Siemens Aktiengesellschaft | Evaluation method for a sequence of temporally successive depth images |
CN103424105A (en) * | 2012-05-16 | 2013-12-04 | 株式会社理光 | Object detection method and device |
CN102778953A (en) * | 2012-06-28 | 2012-11-14 | 华东师范大学 | Motion sensing control method of shadow play remote digital performing based on Kinect |
Also Published As
Publication number | Publication date |
---|---|
CN104765440A (en) | 2015-07-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108764024B (en) | Device and method for generating face recognition model and computer readable storage medium | |
CN104765440B (en) | Hand detection method and equipment | |
US10719743B2 (en) | License plate reader using optical character recognition on plural detected regions | |
US10936911B2 (en) | Logo detection | |
KR102554724B1 (en) | Method for identifying an object in an image and mobile device for practicing the method | |
JP5952001B2 (en) | Camera motion estimation method and apparatus using depth information, augmented reality system | |
US9710698B2 (en) | Method, apparatus and computer program product for human-face features extraction | |
JP7051267B2 (en) | Image detection methods, equipment, electronic equipment, storage media, and programs | |
EP3113114A1 (en) | Image processing method and device | |
WO2019071664A1 (en) | Human face recognition method and apparatus combined with depth information, and storage medium | |
GB2549554A (en) | Method and system for detecting an object in an image | |
KR20190028349A (en) | Electronic device and method for human segmentation in image | |
CN109389038A (en) | A kind of detection method of information, device and equipment | |
CN109919971B (en) | Image processing method, image processing device, electronic equipment and computer readable storage medium | |
US20120119984A1 (en) | Hand pose recognition | |
US10764563B2 (en) | 3D enhanced image correction | |
CN105912126B (en) | A kind of gesture motion is mapped to the adaptive adjusting gain method at interface | |
US20170374268A1 (en) | Focusing point determining method and apparatus | |
CN111401318B (en) | Action recognition method and device | |
US10122912B2 (en) | Device and method for detecting regions in an image | |
JP6997369B2 (en) | Programs, ranging methods, and ranging devices | |
JP2021073606A (en) | Crowd state recognition device, method, and program | |
JP2020154773A (en) | Image processor, image processing method, and image processing system | |
JP2018120283A (en) | Information processing device, information processing method and program | |
CN107209556A (en) | For to capturing object relative to the system and method that the depth image of the interaction of interactive flat is handled |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20170811 Termination date: 20200102 |
|
CF01 | Termination of patent right due to non-payment of annual fee |