CN105759967B - A kind of hand overall situation attitude detecting method based on depth data - Google Patents

A kind of hand overall situation attitude detecting method based on depth data Download PDF

Info

Publication number
CN105759967B
CN105759967B CN201610093720.6A CN201610093720A CN105759967B CN 105759967 B CN105759967 B CN 105759967B CN 201610093720 A CN201610093720 A CN 201610093720A CN 105759967 B CN105759967 B CN 105759967B
Authority
CN
China
Prior art keywords
point
palm
hand
finger
finger tip
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610093720.6A
Other languages
Chinese (zh)
Other versions
CN105759967A (en
Inventor
程洪
姬艳丽
李昊鑫
唐竟人
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201610093720.6A priority Critical patent/CN105759967B/en
Publication of CN105759967A publication Critical patent/CN105759967A/en
Application granted granted Critical
Publication of CN105759967B publication Critical patent/CN105759967B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of hand overall situation attitude detecting method based on depth data, it includes following sub-step: S1: finger three-dimensional detection: using hand depth data as input, detect palm center to finger general direction, simultaneously by calculating point cloud center as palm center, current D translation is obtained;S2: the three-dimensional normal vector of palm the detection of palm three-dimensional: is obtained by fit Plane;S3: palm overall situation posture indicates: palm overall situation posture shows as the D translation of palm and the three-dimensional rotation of palm, and three-dimensional rotation is combined to obtain by the rotation of finger orientation and the rotation of palm normal vector.The present invention has counted a set of simple effective method on the basis of hand depth data, and the detection of hand overall situation posture is realized using three steps, and clear in structure to be illustrated, algorithm overall simple, execution efficiency is high, has preferable practicability.

Description

A kind of hand overall situation attitude detecting method based on depth data
Technical field
The present invention relates to field of human-computer interaction more particularly to a kind of hand overall situation attitude detection sides based on depth data Method.
Background technique
Human-computer interaction wants communication channel as the group between people and robot, technology also constantly by operate it is feasible to Convenient, comfortable direction is developed, in terms of more and more researchs are placed on the interaction technique based on hand, compared to other human body portions Position, hand freedom and flexibility are responsible for a large amount of interworking in the routine work of people, and the operand completed by hand is not Win number, but based on traditional interactive mode mostly operated with hand, and three-dimension gesture interaction is compared to traditional mode, not only It is more naturally comfortable, it can also promote the most direct interactive experience of user.And in recent years, with the increasingly stream of depth transducer technology Row, the technical threshold of three-dimension gesture technology compared to being greatly reduced in the past, while various demands are continuously increased, and are promoted The further development of three-dimension gesture technical research.
The technical foundation of three-dimension gesture is exactly hand gestures estimation, at present the hand gestures estimation method master under three-dimension gesture Be divided into discriminate and production, the method combination threedimensional model of production carries out the mode of energy function optimization, often according to Rely and is accelerated with GPU, it is computationally intensive.Production method is trained model by using database, then is estimated, for Untrained posture is often ineffective, and precision is general.In actual three-dimension gesture problem, the posture of hand can be described For the posture of palm and the posture of finger, the posture of palm is global posture, and the posture of finger is influenced by the posture of palm, Therefore the posture of palm is carried out according to a preliminary estimate, the difficulty of hand entirety Attitude estimation capable of being substantially reduced.Part Methods at present In estimation hand gestures, i.e., global direction, the method for use has cloud PCA method, the Return Law, model fitting method etc..Its midpoint cloud PCA method will appear the case where apparent error judges for the hand of three-dimensional deformation, and the recurrence mode based on model is by random character Be affected, and model fitting method then needs to generate the threedimensional model under a large amount of hand overall situations direction, be not still it is optimal and Simplest method.
In conjunction at present to three-dimensional hand overall situation direction determining method advantage and disadvantage on the basis of, the present invention proposes on this basis Hand overall situation direction detection method based on depth data.
The present invention is directed to hand overall situation Attitude estimation problem, and problem is decomposed, and the finger for detecting hand respectively is three-dimensional Direction and palm three-dimensional calculate the global posture of hand, i.e. three-dimensional rotation and D translation in the result estimated herein, real Now to hand overall situation direction estimation.Method proposed by the present invention is clear in structure to be illustrated, and algorithm overall simple, execution efficiency is high, tool There is preferable practicability.
Application No. is the patents of invention of CN201410254196.7 to disclose a kind of tracking of real time hand, posture classification and interface Control.This invention proposes the devices that Electronic Design operation is carried out based on gesture, and main functional modules include finger tracking, appearance State classification, Interface Control and related circuit design etc..In gesture posture categorization module, the invention simultaneously extracts layout areas sift spy The form for levying point, carries out the classification of gesture posture, based on sift feature using SVM to determine hand posture using classification results.This Invention propose method be hand global posture detection, therefore with this invention towards the problem of different, the method to be solved Thinking is also made a world of difference.
Application No. is 201180075255.9 patents of invention to disclose a kind of man-machine friendship of the short distance based on dynamic posture Mutually.The device that the invention proposes includes the parts such as posture starting detecting and alarm, gesture recognition engine and user command, wherein posture Identify that hand gestures and hand track in engine implementation depth image, hand track are examined using masstone and depth histogram It surveys, and gesture recognition engine then extracts bias, rectangularity, compactness, orientation, center etc. profile spy to hand bianry image Sign, then using classifier carry out gesture posture classification, it can be seen that the method have with patent 1 it is a little similar, with this patent Problem direction to be solved does not have to, and method is also different.
Application No. is 201210438642.0 patents of invention to disclose a kind of space gesture posture based on depth camera Command processing method.The device that the invention proposes obtains hand point cloud data using Kinect sensor, and is registrated according to plane Data later extract outline data, realize the identification of gesture posture.The method utilizes three-dimensional after obtaining hand depth data Plane carries out point cloud registering, that is, seeks the position and direction of palm plane, then projects to a cloud and obtains two in plane It is worth image, then extracts finger tip information and calculate hand gestures.Wherein this mode is to only determining simply by cloud plane Palm plane, the effect and bad for the method under deformation gesture is different with the problem to be solved in the present invention target, and The method used in volar direction detection is also different, and the method that the present invention uses is simple, while can adapt to more Complicated gesture situation.
Application No. is the patents of invention of 20131047231.X to disclose a kind of Fingertip Detection based on depth information. The invention carries out finger tip detection using depth data, mainly includes palm segmentation and palm of the hand positioning, refers to root positioning and finger tip inspection Survey and etc..Wherein when referring to root positioning, the invention uses outline concave as hand root, but deformation gesture bottom profiled is recessed Can not be as the judgment basis for referring to root, the method that graph model is used for detection the invention of finger tip.And this patent is also adopted Finger tip detection has been carried out with profile, but only detection part finger tip, method is different, and is not distinguish to finger tip, and to refer to Point removes detection finger substantially three-dimensional, therefore problem to be solved is different, and method is also different.
Summary of the invention
The hand overall situation posture based on depth data that it is an object of the invention to overcome the deficiencies of the prior art and provide a kind of Detection method, most important problem are how to utilize the hand data of depth transducer, rapidly estimate the current gesture of user Under hand overall situation posture, i.e. the 3 d pose of palm lays the foundation for later hand gestures estimation.
The purpose of the present invention is achieved through the following technical solutions: a kind of hand overall situation posture based on depth data Detection method, it includes following sub-step:
S1: the detection of finger three-dimensional: using hand depth data as input, detect palm center to finger substantially Direction Vy, while by calculating point cloud center as palm center, obtain current D translation Tg
S2: the three-dimensional normal vector V of palm the detection of palm three-dimensional: is obtained by fit Planez
S3: palm overall situation posture indicates: palm overall situation posture shows as the D translation T of palmgWith the three-dimensional rotation of palm Turn Tr, three-dimensional rotation is combined to obtain by the rotation of finger orientation and the rotation of palm normal vector.
The step S1 includes following sub-step:
S11: image preprocessing: pre-processing depth image, i.e., thresholding processing is first carried out, by image background regions Zero setting only retains hand shape area;Closed operation is carried out on this basis;
S12: it calculates hand images center: for the image of hand depth data, carrying out hand center calculation, as point cloud Central point Pcen, wherein calculation is the mean value of each dimension of hand region pixel;
The detection of S13:2D finger tip point, including following sub-step:
S131: the bianry image obtained for pretreatment carries out profile C detection, finds the corresponding profile C of maximum perimeter, Convex polygon detection is carried out to profile C;
S132: carrying out convex closure to convex polygon and defect point detect, and the convex closure point includes finger tip point and non-finger tip point, The defect point then contains concave point and noise spot between finger;
S133: it is screened to reconnaissance: for convex closure point, calculating separately the distance between two two o'clocks, excluded distance and be less than threshold value Point, that is, exclude continuous and extra convex closure point;For defect point, calculate first each defect point to hand picture centre away from From, then calculate hand rotation Rectangular Bounding Volume, obtain the corresponding square boundary of hand region, wherein rectangle it is long with it is wide in it is larger Value is L, and the defect point by distance greater than 1/2L excludes, i.e. noise drawbacks point on exclusion profile;
S134: after obtaining defect point and convex closure point, location index C of each convex closure point in profile is calculatedt, similarly lack Location index C of the trapping spot in profileq
S135: according to convex closure order traversal convex closure point, if the corresponding outline position of some convex closure point indexes CtPositioned at two Between the outline position of a Adjacent defect point, it may be assumed that Cq1< Ct< Cq2, then it is considered as finger tip point PTip, it is otherwise non-finger tip point PNTip
As the finger tip point P metTipWhen less than two, the outline position of each defect point is indexed, detection and its most phase The outline position index of adjacent convex closure point, meets Ct1> Cq, Ct2< Cq1When, this convex closure point is calculated at a distance from 2D centroid, if It is greater than 1/3L, then is considered as finger tip point T2d, it otherwise just abandons, while carrying out the judgement of finger tip point number:
When being greater than 5, stop finger tip point detection.And if finger tip point number is 0 after this step calculates, at this time The vertical straight-on camera of hand, finger orientation are set as particular value, towards lens direction;
S14: the 2D fingertip location P of initial option is determinedTipLater, image is transformed in conjunction with profile C give directions Ttip2d, because For PTipFor finger tip dot profile index value;It is determined 3D finger tip direction, including following sub-step at this time:
S141: in conjunction with Ttip2dCorresponding depth Z is changed into the 3D point T under camera coordinates system under depth datatip3d, To a cloud central point PcenAlso P under camera coordinates system is transformed tocen3D
S142: each 3D finger tip is calculated separately to Pcen3DThree-dimensional vector: Pvi=T3di-Pcen3D, then VyIt is each The mean value of finger tip three-dimensional:
It is normalized simultaneously.
It further include a treatment on special problems step after the step S135, for three-dimensional hand, there are thumbs to erect It rises, and the case where remaining digital flexion, the finger tip point detected at this time is not actual finger general direction, needs to add thus Enter approximate wrist point detection;During step S135 detection, if the defect point of detection is 1, or the finger tip of detection When point number is 1, approximate wrist point detection, including following sub-step are carried out at this time:
S1011: for all non-finger tip convex closure point PNTip, calculate separately the depth value Z (P of its corresponding pixel pointsNTip), Select maximum Z (PNTip) corresponding PNTip_minAs preliminary wrist point;
S1012: approximate wrist point at this time is the convex closure point on profile, needs to detect its neighborhood point, finds neighborhood depth It is worth bigger profile point;P thuswrist=minZ (PNTip_min+ Neb), Neb=1: ± N, wherein PwristIt is expressed as quasi- wrist point PNTip_minThe point that depth value is bigger within the scope of neighborhood N is as approximate wrist point.
When the point number of 2D finger tip detection is 1, the information that approximate wrist point direction is added at this time is corrected, including with Lower sub-step:
PwristBe converted to three-dimensional point P under camera spacewrist3D, then orientation Vw=Pcen3D-Pwrist3D, and utilize VwIt is right VyIt is corrected:
V after normalizing at this timeyIt is exactly the substantially three-dimensional of finger, while Tg=Pcen3D
The step S2 includes following sub-step:
S21: non-finger tip point determines: during 2D finger tip detection, the non-finger tip point P that recordsNTipWith defect point one And it is denoted as palm point Ppalm, image coordinate point T is obtained by profile Cpalm, and camera is converted to according to depth value and camera parameter and is sat 3D point T under mark systempalm3D
S22: plane fitting is carried out to the non-finger tip point of 3D: sets spatial plane equation as Ax+By+Cz+D=0, remembers the non-finger tip of 3D Point is Tpalm3Di=(xi, yi, zi), then seek objective function are as follows:
Using least square method, the parameter { A, B, C, D } of plane equation is solved, plane normal vector is { A, B, C }, at this time hand Slapping direction is approximately Vz={ A, B, C }.
Palm overall situation direction is expressed as three-dimensional rotation Tr, since the freedom degree of palm rotation is higher, revolved by finger orientation Transfering the letter breath and volar direction rotation information, calculate three-dimensional rotation Tr;The step S3 includes following sub-step:
S31: primary standard finger orientation upward, is denoted as Vy0, and finger straight-on camera, direction are denoted as Vz0;It is counted by step S1 Obtained finger substantially three-dimensional Vy, find out Ty;The TyIt is indicated by spin matrix or quaternary number, calculation To calculate Vy、Vy0Cross product and angle;
S32: for the rotation information of volar direction, V is first carried outzUpdate Vz=Ty*Vz, and with side identical with step S31 Formula obtains rotation information Tz
S33: the three-dimensional rotation T of entire handr=Ty*Tz, to detect the global posture of hand.
The beneficial effects of the present invention are:
Method proposed by the present invention is based on the hand depth data of depth transducer (such as Kinect2) shooting, in hand depth For degree on the basis of, the present invention devises a set of simple effective method, realizes hand overall situation posture using three steps Detection: the detection of finger three-dimensional, the detection of palm three-dimensional and the expression of palm overall situation posture.For hand overall situation appearance State shows as the D translation and three-dimensional rotation of palm.Wherein D translation is expressed as the movement of palm, and three-dimensional rotation indicates Palm plane is varied under current pose by first state of value.Due to the flexibility of hand, freedom degree is higher, and according to the spy of palm Point, using the two normal vectors, can realize opponent by detecting finger substantially three-dimensional and palm three-dimensional normal vector respectively The estimation of portion's overall situation posture.Method proposed by the present invention is clear in structure to be illustrated, and algorithm overall simple, execution efficiency is high, have compared with Good practicability.
Detailed description of the invention
Fig. 1 is the method for the present invention flow chart.
Specific embodiment
Technical solution of the present invention is described in further detail with reference to the accompanying drawing:
The present invention includes 3 steps:
(1) finger three-dimensional detects: finger three-dimensional indicates the direction from palm center to middle fingertip, and three In the case that finger shape changes under dimension gesture, such as digital flexion, and palm rotates, relatively it is difficult to estimate finger at this time General direction.Through analyzing in the depth map of hand, often observe that part finger tip still than more prominent, therefore can pass through Finger tip and the line at hand center obtain finger three-dimensional.
(2) palm three-dimensional detects: palm three-dimensional is expressed as the normal vector of palm plane, since palm can be close Like being plane, normal vector is by palm direction outside.Three-dimension gesture assistant's palm point still in approximate plane, therefore The present invention obtains the three-dimensional normal vector of palm by fit Plane.
(3) expression of palm overall situation posture: palm overall situation posture shows as the three-dimensional rotation of the D translation and palm of palm Turn, D translation is obtained by hand center, and three-dimensional rotation then passes through the rotation of finger orientation and the rotation phase group of palm normal vector Conjunction obtains, to detect hand overall situation posture by above method, not only algorithm is simple, but also is to the three-dimensional of some of complex Also there is certain applicability, there is good practical value.
As shown in Figure 1, a kind of hand overall situation attitude detecting method based on depth data, it includes following sub-step:
S1: the detection of finger three-dimensional: finger orientation detection is mainly used as input using hand depth data, the purpose is to Detected in the case where angle is changeable and gesture is changeable palm center to finger general direction Vy, the present invention is with practical true Data are analyzed, and are found the corresponding depth map of most of gesture, are still remained all or part of tip shape, use thus The detection of 3D finger tip detection method progress finger orientation.
Using hand depth data as input, detect palm center to finger general direction Vy, while passing through calculating Point cloud center obtains current D translation T as palm centerg
The step S1 includes following sub-step:
S11: image preprocessing: pre-processing depth image, i.e., thresholding processing is first carried out, by image background regions Zero setting only retains hand shape area;It carries out closed operation on this basis, not only reduces noise and small empty profile in this way It influences, can more eliminate the influence in invalid depth region.
S12: it calculates hand images center: for the image of hand depth data, carrying out hand center calculation, as point cloud Central point Pcen, wherein calculation is the mean value of each dimension of hand region pixel;
The detection of S13:2D finger tip point, including following sub-step:
S131: the bianry image obtained for pretreatment carries out profile C detection, finds the corresponding profile C of maximum perimeter, Convex polygon detection is carried out to profile C;The convex polygon being achieved in that approximate can replace hand profile, can be avoided using wheel Wide C bring not robustness.
S132: carrying out convex closure to convex polygon and defect point detect, and the convex closure point includes finger tip point and non-finger tip point, The defect point then contains concave point and noise spot between finger;Convex closure point has generally comprised finger tip point and non-finger tip at this time Point, and defect point then contains concave point and noise spot between finger, it is therefore desirable to screened to reconnaissance.
S133: it is screened to reconnaissance:
For convex closure point, the distance between two two o'clocks are calculated separately, exclude the point that distance is less than threshold value, threshold value can be set For 10 pixels, that is, exclude continuous and extra convex closure point;
For defect point, calculate first each defect point to hand picture centre distance, then calculate hand rotation rectangle Bounding box obtains the corresponding square boundary of hand region, and wherein the long the larger value with width of rectangle is L, will be apart from greater than 1/2L Defect point exclude, i.e., exclusion profile on noise drawbacks point;In true hand structure, recess between adjacent finger away from It is less than the half of hand sizes with a distance from hand center.
S134: after obtaining defect point and convex closure point, location index C of each convex closure point in profile is calculatedt, similarly lack Location index C of the trapping spot in profileq
S135: according to convex closure order traversal convex closure point, if the corresponding outline position of some convex closure point indexes CtPositioned at two Between the outline position of a Adjacent defect point, it may be assumed that Cq1< Ct< Cq2, then it is considered as finger tip point PTip, it is otherwise non-finger tip point PNTip
As the finger tip point P metTipWhen less than two, the outline position of each defect point is indexed, detection and its most phase The outline position index of adjacent convex closure point, meets Ct1> Cq, Ct2< Cq1When, this convex closure point is calculated at a distance from 2D centroid, if It is greater than 1/3L, then is considered as finger tip point T2d, it otherwise just abandons, while carrying out the judgement of finger tip point number:
When being greater than 5, stop finger tip point detection.And if finger tip point number is 0 after this step calculates, at this time The vertical straight-on camera of hand, finger orientation are set as particular value, towards lens direction;
It further include a treatment on special problems step after the step S135, for three-dimensional hand, there are thumbs to erect It rises, and the case where remaining digital flexion, the finger tip point detected at this time is not actual finger general direction, needs to add thus Enter approximate wrist point detection;During step S135 detection, if the defect point of detection is 1, or the finger tip of detection When point number is 1, approximate wrist point detection, including following sub-step are carried out at this time:
S1011: for all non-finger tip convex closure point PNTip, calculate separately the depth value Z (P of its corresponding pixel pointsNTip), Select maximum Z (PNTip) corresponding PNTip_minAs preliminary wrist point;Because wrist is more than finger under actual gesture operation Far from camera, depth value is bigger.
S1012: approximate wrist point at this time is the convex closure point on profile, needs to detect its neighborhood point, finds neighborhood depth It is worth bigger profile point;P thuswrist=minZ (PNTip_min+ Neb), Neb=1: ± 20, wherein PwristIt is expressed as quasi- wrist point PNTip_minThe point that depth value is bigger in 20 range of neighborhood is as approximate wrist point.
S14: the 2D fingertip location P of initial option is determinedTipLater, image is transformed in conjunction with profile C give directions Ttip2d, because For PTipFor finger tip dot profile index value;It is determined 3D finger tip direction, including following sub-step at this time:
S141: in conjunction with Ttip2dCorresponding depth Z is changed into the 3D point T under camera coordinates system under depth datatip3d, To a cloud central point PcenAlso P under camera coordinates system is transformed tocen3D
S142: each 3D finger tip is calculated separately to Pcen3DThree-dimensional vector: Pvi=T3di-Pcen3D, then VyIt is each The mean value of finger tip three-dimensional:
It is normalized simultaneously.
When the point number of 2D finger tip detection is 1, the information that approximate wrist point direction is added at this time is corrected, including with Lower sub-step:
PwristBe converted to three-dimensional point P under camera spacewrist3D, then orientation Vw=Pcen3D-Pwrist3D, and utilize VwIt is right
It is corrected:
V after normalizing at this timeyIt is exactly the substantially three-dimensional of finger, while Tg=Pcen3D
S2: the three-dimensional normal vector V of palm the detection of palm three-dimensional: is obtained by fit Planez;Volar direction is by palm Normal vector VzIt is indicated, is directed toward outside from palm.
The step S2 includes following sub-step:
S21: non-finger tip point determines: during 2D finger tip detection, the non-finger tip point P that recordsNTipWith defect point one And it is denoted as palm point Ppalm, image coordinate point T is obtained by profile Cpalm, and camera is converted to according to depth value and camera parameter and is sat 3D point T under mark systempalm3D;Because the palm in-plane under three-dimension gesture is three-dimensional.
S22: carry out plane fitting to the non-finger tip point of 3D: all non-finger tip points represent the point on palm, and palm can To be approximately plane, therefore the fitting of least square method space plane is carried out based on the non-finger tip point of 3D.
If spatial plane equation is Ax+By+Cz+D=0, the note non-finger tip point of 3D is Tpalm3Di=(xi, yi, zi), then it seeks Objective function are as follows:
Using least square method, the parameter { A, B, C, D } of plane equation is solved, plane normal vector is { A, B, C }, at this time hand Slapping direction is approximately Vz={ A, B, C }.
Simultaneously because the sequence of contour detecting be it is counterclockwise, the finger normal vector for calculating acquisition is also directed towards camera side To, therefore volar direction proposed by the invention is estimated to be suitble to most of situation of the volar direction towards camera, does not consider palm The case where backwards to camera.
S3: palm overall situation posture indicates: palm overall situation posture shows as the D translation T of palmgWith the three-dimensional rotation of palm Turn Tr, three-dimensional rotation is combined to obtain by the rotation of finger orientation and the rotation of palm normal vector.
Palm overall situation direction is expressed as three-dimensional rotation Tr, since the freedom degree of palm rotation is higher, revolved by finger orientation Transfering the letter breath and volar direction rotation information, calculate three-dimensional rotation Tr;The step S3 includes following sub-step:
Palm overall situation direction is expressed as three-dimensional rotation Tr, can be by finger side since the freedom degree of palm rotation is higher To rotation information and volar direction rotation information, three-dimensional rotation T is calculatedr.Assuming that primary standard finger orientation is upward, it is denoted as Vy0, and finger straight-on camera, direction are denoted as Vz0, the finger substantially three-dimensional V that is calculated at this time by step 1y, find out Ty, And TyIt can be indicated, can also be indicated by quaternary number, calculation can calculate V by spin matrixy、Vy0Cross product and angle It calculates.For the rotation information of volar direction, V is first carried outzUpdate Vz=Ty*Vz, and rotation information T is obtained in the same wayz, The then three-dimensional rotation T of entire handr=Ty*Tz, to detect the global posture of hand.

Claims (5)

1. a kind of hand overall situation attitude detecting method based on depth data, it is characterised in that: it includes following sub-step:
S1: finger three-dimensional detection: using hand depth data as input, detect palm center to finger general direction Vy, while by calculating point cloud center as palm center, obtain current D translation Tg
S2: the three-dimensional normal vector V of palm the detection of palm three-dimensional: is obtained by fit Planez
S3: palm overall situation posture indicates: palm overall situation posture shows as the D translation T of palmgWith the three-dimensional rotation T of palmr, Three-dimensional rotation is combined to obtain by the rotation of finger orientation and the rotation of palm normal vector;
The step S1 includes following sub-step:
S11: image preprocessing: pre-processing depth image, i.e., first carries out thresholding processing, image background regions are set Zero, only retain hand shape area;Closed operation is carried out on this basis;
S12: it calculates hand images center: for the image of hand depth data, carrying out hand center calculation, as point Yun Zhongxin Point Pcen, wherein calculation is the mean value of each dimension of hand region pixel;
The detection of S13:2D finger tip point, including following sub-step:
S131: the bianry image obtained for pretreatment carries out profile C detection, the corresponding profile C of maximum perimeter is found, to wheel Wide C carries out convex polygon detection;
S132: carrying out convex closure point to convex polygon and defect point detect, and the convex closure point includes finger tip point and non-finger tip point, institute The defect point stated then contains concave point and noise spot between finger;
S133: it is screened to reconnaissance: for convex closure point, calculating separately the distance between two two o'clocks, exclude the point that distance is less than threshold value, Exclude continuous and extra convex closure point;For defect point, calculate first each defect point to hand picture centre distance, then It calculates hand and rotates Rectangular Bounding Volume, obtain the corresponding square boundary of hand region, wherein the long the larger value with width of rectangle is L, the defect point by distance greater than 1/2L exclude, i.e. noise drawbacks point on exclusion profile;
S134: after obtaining defect point and convex closure point, location index C of each convex closure point in profile is calculatedt, similarly defect point Location index C in profileq
S135: according to convex closure order traversal convex closure point, if the corresponding outline position of some convex closure point indexes CtIt is adjacent positioned at two Between the outline position of defect point, it may be assumed that Cq1< Ct< Cq2, then it is considered as finger tip point PTip, it is otherwise non-finger tip point PNTip
As the finger tip point P metTipWhen less than two, the outline position of each defect point is indexed, detection is most adjacent with it The outline position of convex closure point indexes, and meets Ct1> Cq, Ct2< CqWhen, two convex closure points are at a distance from 2D centroid at this time for calculating, such as It is greater than 1/3L to fruit, then is considered as finger tip point PTip, otherwise just abandon;The judgement of finger tip point number is carried out simultaneously: when judging the finger When sharp number is greater than 5, stops finger tip point detection and then confirm the vertical face of hand at this time when judging finger tip number for 0 Camera, finger orientation are set as particular value, towards lens direction;
S14: the 2D. fingertip location P of initial option is determinedTipLater, image is transformed in conjunction with profile C give directions Ttip2d, because PTipFor finger tip dot profile index value;It is determined 3D finger tip direction, including following sub-step at this time:
S141: in conjunction with Ttip2dCorresponding depth Z is changed into the 3D point T under camera coordinates system under depth datatip3d, to a cloud Central point PcenAlso P under camera coordinates system is transformed tocen3D
S142: each 3D finger tip is calculated separately to Pcen3DThree-dimensional vector: Pvi=Ttip3d-Pcen3D, then VyFor each finger tip The mean value of three-dimensional:
It is normalized simultaneously.
2. a kind of hand overall situation attitude detecting method based on depth data according to claim 1, it is characterised in that: institute It further include a treatment on special problems step after the step S135 stated, for three-dimensional hand, there are thumbs to hold up, and remaining The case where digital flexion, the finger tip point detected at this time are not actual finger general direction, need that approximate wrist is added thus Point detection;During step S135 detection, if the defect point of detection is 1, or the finger tip point number of detection is 1 When, approximate wrist point detection, including following sub-step are carried out at this time:
S1011: for all non-finger tip convex closure point PNTip, calculate separately the depth value ZP of its corresponding pixel pointsNTip, selection is most Big ZPNTipCorresponding PNTip_maxAs preliminary wrist point;
S1012: approximate wrist point at this time is the convex closure point on profile, needs to detect its neighborhood point, finds neighborhood depth value more Big profile point;P thuswrist=minZ (PNTip_max+ Neb), Neb=1: ± N, wherein PwristIt is expressed as preliminary wrist point PNTip_maxThe point that depth value is bigger within the scope of neighborhood N is as approximate wrist point.
3. a kind of hand overall situation attitude detecting method based on depth data according to claim 2, it is characterised in that: when When the point number of 2D finger tip detection is 1, the information that approximate wrist point direction is added at this time is corrected, including following sub-step:
Pwrist is converted to three-dimensional point P under camera spacewrist3D, then orientation Vw=Pcen3D-Pwrist3D, and utilize VwTo VyInto Row correction:
V after normalizing at this timeyIt is exactly the substantially three-dimensional of finger, while Tg=Pcen3D
4. a kind of hand overall situation attitude detecting method based on depth data according to claim 2 or 3, feature exist In: the step S2 includes following sub-step:
S21: non-finger tip point determines: during 2D finger tip detection, the non-finger tip point P that recordsNTipIt is denoted as together with defect point Palm point Ppalm, image coordinate point T is obtained by profile Cpalm, and be converted under camera coordinates system according to depth value and camera parameter 3D point Tpalm3D
S22: plane fitting is carried out to the non-finger tip point of 3D: sets spatial plane equation as Ax+By+Cz+D=0, the note non-finger tip point of 3D is Tpalm3Di=(xi, yi, zi), then seek objective function are as follows:
Using least square method, the parameter { A, B, C, D } of plane equation is solved, plane normal vector is { A, B, C }, at this time palm side To being approximately Vz={ A, B, C }.
5. a kind of hand overall situation attitude detecting method based on depth data according to claim 4, it is characterised in that: hand It slaps global direction and is expressed as three-dimensional rotation Tr, since the freedom degree of palm rotation is higher, pass through finger orientation rotation information and hand Direction rotation information is slapped, three-dimensional rotation T is calculatedr;The step S3 includes following sub-step:
S31: primary standard finger orientation upward, is denoted as Vy0, and finger straight-on camera, direction are denoted as Vz0;It is calculated by step S1 The finger arrived substantially three-dimensional Vy, find out Ty;The TyIt is indicated by spin matrix or quaternary number, calculation is meter Calculate Vy、Vy0Cross product and angle;
S32: for the rotation information of volar direction, V is first carried outzUpdate Vz=Ty*Vz, and obtained in a manner of identical with step S31 To rotation information Tz
S33: the three-dimensional rotation T of entire handr=Ty*Tz, to detect the global posture of hand.
CN201610093720.6A 2016-02-19 2016-02-19 A kind of hand overall situation attitude detecting method based on depth data Active CN105759967B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610093720.6A CN105759967B (en) 2016-02-19 2016-02-19 A kind of hand overall situation attitude detecting method based on depth data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610093720.6A CN105759967B (en) 2016-02-19 2016-02-19 A kind of hand overall situation attitude detecting method based on depth data

Publications (2)

Publication Number Publication Date
CN105759967A CN105759967A (en) 2016-07-13
CN105759967B true CN105759967B (en) 2019-07-09

Family

ID=56330196

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610093720.6A Active CN105759967B (en) 2016-02-19 2016-02-19 A kind of hand overall situation attitude detecting method based on depth data

Country Status (1)

Country Link
CN (1) CN105759967B (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106325509A (en) * 2016-08-19 2017-01-11 北京暴风魔镜科技有限公司 Three-dimensional gesture recognition method and system
CN106815578A (en) * 2017-01-23 2017-06-09 重庆邮电大学 A kind of gesture identification method based on Depth Motion figure Scale invariant features transform
TWI662482B (en) * 2017-06-28 2019-06-11 緯創資通股份有限公司 Method and Module of Gesture Recognition
CN110581987A (en) * 2018-06-07 2019-12-17 宏碁股份有限公司 Three-dimensional display with gesture sensing function
CN110880186A (en) * 2018-09-06 2020-03-13 山东理工大学 Real-time human hand three-dimensional measurement method based on one-time projection structured light parallel stripe pattern
CN109271931A (en) * 2018-09-14 2019-01-25 辽宁奇辉电子系统工程有限公司 It is a kind of that gesture real-time identifying system is pointed sword at based on edge analysis
CN109407848B (en) * 2018-11-01 2021-07-13 广东源创智能科技有限公司 Step detection method and device
CN109521879B (en) * 2018-11-19 2022-02-18 杭州易现先进科技有限公司 Interactive projection control method and device, storage medium and electronic equipment
CN110472507A (en) * 2019-07-12 2019-11-19 中国地质大学(武汉) Manpower depth image position and orientation estimation method and system based on depth residual error network
WO2021160024A1 (en) * 2020-02-14 2021-08-19 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and system of identifying a user selection at a display of a user device
CN111694432B (en) * 2020-06-11 2023-04-07 济南大学 Virtual hand position correction method and system based on virtual hand interaction
CN116958146B (en) * 2023-09-20 2024-01-12 深圳市信润富联数字科技有限公司 Acquisition method and device of 3D point cloud and electronic device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102368290A (en) * 2011-09-02 2012-03-07 华南理工大学 Hand gesture identification method based on finger advanced characteristic
CN102982557A (en) * 2012-11-06 2013-03-20 桂林电子科技大学 Method for processing space hand signal gesture command based on depth camera
CN103488972A (en) * 2013-09-09 2014-01-01 西安交通大学 Method for detection fingertips based on depth information

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102368290A (en) * 2011-09-02 2012-03-07 华南理工大学 Hand gesture identification method based on finger advanced characteristic
CN102982557A (en) * 2012-11-06 2013-03-20 桂林电子科技大学 Method for processing space hand signal gesture command based on depth camera
CN103488972A (en) * 2013-09-09 2014-01-01 西安交通大学 Method for detection fingertips based on depth information

Also Published As

Publication number Publication date
CN105759967A (en) 2016-07-13

Similar Documents

Publication Publication Date Title
CN105759967B (en) A kind of hand overall situation attitude detecting method based on depth data
US9189855B2 (en) Three dimensional close interactions
US9330307B2 (en) Learning based estimation of hand and finger pose
CN106055091B (en) A kind of hand gestures estimation method based on depth information and correcting mode
Ma et al. Kinect sensor-based long-distance hand gesture recognition and fingertip detection with depth information
CN105389539B (en) A kind of three-dimension gesture Attitude estimation method and system based on depth data
Hasan et al. RETRACTED ARTICLE: Static hand gesture recognition using neural networks
Ye et al. Accurate 3d pose estimation from a single depth image
WO2017219391A1 (en) Face recognition system based on three-dimensional data
Kanhangad et al. Contactless and pose invariant biometric identification using hand surface
CN108052942B (en) Visual image recognition method for aircraft flight attitude
Yilmaz et al. A differential geometric approach to representing the human actions
WO2012077286A1 (en) Object detection device and object detection method
CN109359514A (en) A kind of gesture tracking identification federation policies method towards deskVR
CN106446773A (en) Automatic robust three-dimensional face detection method
CN111160291A (en) Human eye detection method based on depth information and CNN
CN107479693A (en) Real-time hand recognition methods based on RGB information, storage medium, electronic equipment
CN103577792A (en) Device and method for estimating body posture
Kanaujia et al. Part segmentation of visual hull for 3d human pose estimation
Cohen et al. 3D body reconstruction for immersive interaction
Azad et al. Accurate shape-based 6-dof pose estimation of single-colored objects
CN108694348B (en) Tracking registration method and device based on natural features
Guo et al. A hybrid framework based on warped hierarchical tree for pose estimation of texture-less objects
Sunyoto et al. Wrist detection based on a minimum bounding box and geometric features
Wang Hand gesture recognition based on fingertip detection

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant