CN105160323B - A kind of gesture identification method - Google Patents

A kind of gesture identification method Download PDF

Info

Publication number
CN105160323B
CN105160323B CN201510563293.9A CN201510563293A CN105160323B CN 105160323 B CN105160323 B CN 105160323B CN 201510563293 A CN201510563293 A CN 201510563293A CN 105160323 B CN105160323 B CN 105160323B
Authority
CN
China
Prior art keywords
hand
sequence
image
characteristic point
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201510563293.9A
Other languages
Chinese (zh)
Other versions
CN105160323A (en
Inventor
朱磊
韩琦
杨晓光
李建英
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Yishe Technology Co Ltd
Original Assignee
Harbin Yishe Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Yishe Technology Co Ltd filed Critical Harbin Yishe Technology Co Ltd
Priority to CN201510563293.9A priority Critical patent/CN105160323B/en
Publication of CN105160323A publication Critical patent/CN105160323A/en
Application granted granted Critical
Publication of CN105160323B publication Critical patent/CN105160323B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm
    • G06V40/113Recognition of static hand signs

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of gesture identification method, which includes:Obtain the testing image sequence of user's hand containing depth information;According to image depth information and image color information, the hand profile of user is detected in every frame image of testing image sequence;For every hand of user, using preset hand structure template, the characteristic point sequence to be measured of this hand is determined in every frame image of testing image sequence;For every hand of user, the matching sequence of the characteristic point sequence to be measured of this hand is determined in multiple default characteristic point sequences, to determine denomination of dive and the position of this hand according to matching sequence;The gesture that selection matches with the denomination of dive of user's both hands and position in default gesture table, the gesture identification result as testing image sequence.Above-mentioned technology of the invention can accurately identify the gesture of user, and accuracy of identification is higher, and recognition speed is very fast.

Description

A kind of gesture identification method
Technical field
The present invention relates to signal processing technology more particularly to a kind of gesture identification methods.
Background technique
With the development of science and technology, the electronic equipment of laptop, mobile phone, tablet computer etc. gradually has more Carry out more functions, and gesture identification is exactly one of this various functions.
Currently, the algorithm that the identification processing procedure of existing Gesture Recognition utilizes is complex, the time is not only expended, And thus identify that the precision of gesture is lower, accuracy is poor.In addition, the reality of the existing Gesture Recognition based on pattern-recognition When property is poor, is not suitable for the occasion of real-time interaction demand.
Summary of the invention
It has been given below about brief overview of the invention, in order to provide about the basic of certain aspects of the invention Understand.It should be appreciated that this summary is not an exhaustive overview of the invention.It is not intended to determine pass of the invention Key or pith, nor is it intended to limit the scope of the present invention.Its purpose only provides certain concepts in simplified form, Taking this as a prelude to a more detailed description discussed later.
In consideration of it, being deposited the present invention provides a kind of gesture identification method at least solving existing Gesture Recognition Gesture identification precision, poor accuracy the problem of.
According to an aspect of the invention, there is provided a kind of gesture identification method, the gesture identification method include:Step One, the testing image sequence of user's hand containing depth information is obtained;Step 2: according to image depth information and image face Color information detects the hand profile of the user in every frame image of the testing image sequence;Step 3: for institute Every hand for stating user, using preset hand structure template, determining in every frame image of the testing image sequence should The characteristic point sequence to be measured of hand;Step 4: being directed to every hand of the user, determined in multiple default characteristic point sequences The matching sequence of the characteristic point sequence to be measured of this hand, to determine denomination of dive and the position of this hand according to the matching sequence It sets;Step 5: the gesture that selection matches with the denomination of dive of user's both hands and position in default gesture table, as The gesture identification result of the testing image sequence.
Further, step 3 may include:Step 3 one, every hand for the user, utilize preset hand Portion's stay in place form determines the predetermined number feature of this hand in the hand profile of every frame image of the testing image sequence Point;Step 3 two, every hand for the user are right in each frame image of the testing image sequence using this hand The predetermined number characteristic point answered obtains the characteristic point sequence to be measured of this hand.
Further, step 1 may include:By capturing the image of user's hand in predetermined imaging region, Obtain Detection Method in Optical Image SequencesAnd infrared image sequenceFor the Detection Method in Optical Image Sequences Pixel value at i-th frame image coordinate (x, y), andAt infrared image sequence the i-th frame image coordinate (x, y) Pixel value, according to the following formula obtain extract user's both hands information image sequence:
Wherein, α, β, λ are parameter preset threshold value,For the figure of user's both hands containing depth information of acquisition As sequence, as the testing image sequence, i=1,2 ..., M, the number of image frames that M includes by the testing image sequence.
Further, step 2 may include:For the testing image sequenceIn every frame imageThe color combining information deletion frame imageIn noise spot and non-area of skin color, calculated using edge detection Sub- E () is to obtained image after the deletion noise spot and the non-area of skin colorEdge detection is carried out, is obtained Edge imageThe edge imageIt as only include user's hand The image of contouring.
Further, step 3 one may include:Following processing is executed for every frame image of the testing image sequence: The finger tip point in the contour line is found according to the curvature in the contour line of the image and refers to root artis;Using the finger tip point as Setting base matches each finger root artis singly referred to, obtains benchmark of each length singly referred to as scaling;Based on described Finger tip point and the position for referring to root artis and each length singly referred to zoom in and out the corresponding hand structure template And deformation, each articulations digitorum manus characteristic point and wrist midpoint characteristic point of every hand are obtained by matching;Wherein, the hand structure mould Plate includes left-handed configuration template and right hand configurations template, and the left-handed configuration template and right hand configurations template respectively include:Each hand The fingertip characteristic point of finger, respectively refers between root joint characteristic point, wrist midpoint characteristic point and each characteristic point each articulations digitorum manus characteristic point Topological relation.
Further, step 4 may include:Step 4 one, for the characteristic point sequence to be measured of every hand, according to predetermined The characteristic point sequence to be measured is divided into multiple subsequences by time window, and obtains the corresponding mean place of each subsequence;Step Four or two, it is directed to the corresponding each subsequence of every hand, by each of the subsequence and the multiple default characteristic point sequence It is matched respectively, selects to be higher than preset matching with the matching degree of the subsequence in the multiple default characteristic point sequence Threshold value and maximum default characteristic point sequence, the matching sequence as the subsequence;Step 4 three, by each subsequence pair The mean place answered denomination of dive corresponding with the matching sequence of the subsequence is associated;Step 4 four is directed to every hand, by this The matching sequence of the corresponding each subsequence of hand is and each by multiple matching sequence as the corresponding multiple matching sequences of this hand Multiple denomination of dive of the self-corresponding denomination of dive as this hand.
Further, step 5 may include:Step 5 one presets following map listing as the default hand Gesture table:The left end of each mapping in the map listing is the position of set title pair and each denomination of dive pair;This is reflected The right end for each mapping penetrated in list is a gesture HandSignal;Step 5 two, will be every in the default gesture table The left end of a mapping is matched with the denomination of dive of user's both hands and position, wherein the matching of denomination of dive executes Stringent matching, and position is then relative position information to be calculated by the respective mean place of user's both hands, and then calculate Similarity between the relative position information and the position for mapping left end is realized.
Above-mentioned gesture identification method according to an embodiment of the present invention using first identification single-handed exercise and then passes through both hands The mode of action recognition gesture realizes that can accurately identify the gesture of user, accuracy of identification is higher, and recognition speed compared with Fastly.
Above-mentioned gesture identification method of the invention uses Hierarchical Design algorithm, and algorithm complexity is low, is easy to implement.
It is right when needing to change (such as modify, increase or decrease) in addition, using above-mentioned gesture identification method of the invention It, can be only by adjusting template (that is, by modifying the default corresponding movement of characteristic point sequence when movement and/or the definition of gesture Title changes the definition of movement, increases, subtracts movement by increasing or decreasing default characteristic point sequence and its respective action title) And default gesture table (that is, changing the definition of gesture by modifying the corresponding multiple movements of gesture in default gesture table, passes through Gesture and its respective action in default gesture table are increased or decreased to increase, subtract gesture), without changing algorithm or instructing again Practice classifier, substantially increases the adaptability of algorithm.
In addition, the strong real-time of above-mentioned gesture identification method of the invention, can be suitble to the occasion of real-time interaction demand.
By the detailed description below in conjunction with attached drawing to highly preferred embodiment of the present invention, these and other of the invention is excellent Point will be apparent from.
Detailed description of the invention
The present invention can be by reference to being better understood, wherein in institute below in association with description given by attached drawing Have and has used the same or similar appended drawing reference in attached drawing to indicate same or similar component.The attached drawing is together with following It is described in detail together comprising in the present specification and forming a part of this specification, and is used to that this is further illustrated The preferred embodiment and explanation the principle of the present invention and advantage of invention.In the accompanying drawings:
Fig. 1 is the flow chart for showing an example process of gesture identification method of the invention;
Fig. 2 is the flow chart for showing the example process of step 3 shown in FIG. 1;
Fig. 3 is the flow chart for showing the example process of step 4 shown in FIG. 1;
Fig. 4 is the flow chart for showing the example process of step 5 shown in FIG. 1.
It will be appreciated by those skilled in the art that element in attached drawing is just for the sake of showing for the sake of simple and clear, And be not necessarily drawn to scale.For example, the size of certain elements may be exaggerated relative to other elements in attached drawing, with Just the understanding to the embodiment of the present invention is helped to improve.
Specific embodiment
Exemplary embodiment of the invention is described hereinafter in connection with attached drawing.For clarity and conciseness, All features of actual implementation mode are not described in the description.It should be understood, however, that developing any this actual implementation Much decisions specific to embodiment must be made during example, to realize the objectives of developer, for example, symbol Restrictive condition those of related to system and business is closed, and these restrictive conditions may have with the difference of embodiment Changed.In addition, it will also be appreciated that although development is likely to be extremely complex and time-consuming, to having benefited from the disclosure For those skilled in the art of content, this development is only routine task.
Here, and also it should be noted is that, in order to avoid having obscured the present invention because of unnecessary details, in the accompanying drawings Illustrate only with closely related apparatus structure and/or processing step according to the solution of the present invention, and be omitted and the present invention The little other details of relationship.
The embodiment provides a kind of gesture identification method, which includes:Contain Step 1: obtaining There is the testing image sequence of user's hand of depth information;Step 2: according to image depth information and image color information, The hand profile of the user is detected in every frame image of the testing image sequence;Step 3: being directed to the user Every hand, using preset hand structure template, determined in every frame image of the testing image sequence this hand to Survey characteristic point sequence;Step 4: being directed to every hand of the user, this hand is determined in multiple default characteristic point sequences The matching sequence of characteristic point sequence to be measured, to determine denomination of dive and the position of this hand according to the matching sequence;Step 5: The gesture that selection matches with the denomination of dive of user's both hands and position in default gesture table, as described to mapping As the gesture identification result of sequence.
Fig. 1 shows the flow chart of an example process of gesture identification method of the invention.Above-mentioned gesture identification method Exemplary process start from step 1.
As shown in Figure 1, obtaining the testing image sequence of user's hand containing depth information in step 1.
According to a kind of implementation, the processing of step 1 can be implemented by the following steps:By capturing predetermined imaging The image of user's hand in region, (such as can use visible light image sensor and infrared figure in depth camera As sensor) obtain Detection Method in Optical Image SequencesAnd infrared image sequenceFor visible light figure As the pixel value at sequence the i-th frame image coordinate (x, y), andAt infrared image sequence the i-th frame image coordinate (x, y) Pixel value, according to the following formula it is available extract user's both hands information image sequence:
Wherein, α, β, λ are parameter preset threshold value, these parameter preset threshold values can set based on experience value, can also be with Determined by the method for test (such as the collected sample image training of the depth camera by actually using specific model Obtain), which is not described herein again.For the image sequence of user's both hands containing depth information of acquisition, as upper State testing image sequence.In addition, i=1,2 ..., M, M is number of image frames included in testing image sequence.
It should be noted that the difference (single or double) of the hand quantity according to used in user's gesture, is making a reservation for into As institute's captured image may be the image comprising user's both hands in region, it is also possible to only include the figure of single hand of user Picture.In addition, the testing image sequence obtained can be and obtain in a period of time, which can be previously according to experience Value setting, may be, for example, 10 seconds.
Then, in step 2, according to image depth information and image color information, in every frame figure of testing image sequence The hand profile of user is detected as in.Wherein, the hand profile detected may be both hands profile, it is also possible to single hand wheel It is wide.
According to a kind of implementation, the processing of step 2 can be implemented by the following steps:For testing image sequenceIn every frame imageThe color combining information deletion frame imageIn noise spot and the non-colour of skin Region, using edge detection operator E () to obtained image after erased noise point and non-area of skin colorCarry out side Edge detection, to obtain edge image
Edge imageIt as only include the image of user's hand profile.
Wherein, in the treatment process of " noise spot and non-area of skin color in the color combining information deletion frame image ", Existing denoising method be can use to delete the noise spot in image, and can be by calculating imageMean value obtain skin Color region, then the region except area of skin color is non-area of skin color, and the deletion to non-area of skin color can be realized.For example, To imageMean value after, float up and down a range in the mean value, obtain a color gamut comprising the mean value, work as figure The color value of certain point is fallen within this color gamut as in, then is colour of skin point by the point determination, otherwise it is assumed that not being colour of skin point; All colour of skin points constitute area of skin color, remaining is non-area of skin color.
As a result, by the processing of step 2, the hand profile of user can be quickly detected, improve and entirely handle Speed and efficiency.
Then, in step 3, for every hand of user, using preset hand structure template, in testing image The characteristic point sequence to be measured of this hand is determined in every frame image of sequence.
Wherein, hand structure template includes left-handed configuration template and right hand configurations template, left-handed configuration template and right hand knot Structure template respectively includes the topological relation between predetermined number characteristic point and each characteristic point.
In one example, left-handed configuration template and right hand configurations template can respectively include following 20 (as predetermined number Purpose example, but predetermined number is not limited to 20, can also be the numerical value such as 19,21) a characteristic point:The fingertip characteristic point (5 of each finger It is a), each articulations digitorum manus characteristic point (9), respectively refer to root joint characteristic point (5), wrist midpoint characteristic point (1).
According to a kind of implementation, the processing in step 3 can pass through step three as shown in Figure 2 one and step 3 two To realize.
As shown in Fig. 2, for every hand of user, can use above-mentioned preset hand structure in step 3 one Template, respectively by every frame image of testing image sequence hand profile and hand structure template (tiled configuration template and the right side Hand stay in place form) it matched, be aligned, obtain predetermined number (such as 20) characteristic point in the frame image hand profile.
Then, in step 3 two, for every hand of user, it can use this hand in each of testing image sequence Corresponding predetermined number characteristic point (i.e. feature point set), obtains the characteristic point sequence to be measured of this hand in frame image.
In this way, passing through hand structure template and each hand profile (the i.e. every frame figure of testing image sequence obtained before Hand profile as in) it carries out the processing such as matching, the predetermined number that can quickly and accurately obtain in each hand profile is special Sign point.Thereby, it is possible to subsequent processing using the predetermined number characteristic point in these profiles to further realize hand Gesture identification, compared with the prior art, improves the speed and accuracy of entire gesture recognition process.
In the prior art, when needing to change (such as modify, increase or decrease) to movement according to different application scene Definition when, need to modify algorithm and re -training classifier;In the present invention, template only can be acted (i.e. Default characteristic point sequence) realize the change to action definition, substantially increase the adaptability of algorithm.
In one example, the processing of step 3 one can be realized in the following way.
According to the physiological structure feature of mankind's both hands, portable 20 (example as predetermined number) a spies can be taken to every Sign point.For every frame image of testing image sequenceExecute following processing:Firstly, according to the imageIn Profile curvature of a curve finds the finger tip point in the contour line and refers to root artis;Then using finger tip point as setting base, matching Each finger root artis singly referred to, obtains benchmark of each length singly referred to as scaling;Finally based on the finger tip found Point and refer to the position of root artis and obtained each length for singly referring to of both parameter to corresponding hand structure template into Row scaling and deformation obtain remaining 10 characteristic point of every hand, i.e., each articulations digitorum manus characteristic point and hand of every hand by matching Wrist midpoint characteristic point.
For example, looking for contour lineIn finger tip point and refer to root artis during, can be by its mean curvature most Big salient point is as finger tip point, using the concave point of maximum curvature as webs minimum point, and by each finger tip point to the finger tip point phase Adjacent the distance between webs minimum point is defined as the corresponding unit length of the finger tip point.It is minimum to every two adjacent webs Point, by this 2 points midpoint, toward volar direction extension one third unit length, (unit length at this time is thus between 2 points again The corresponding unit length of finger tip point) point, the corresponding finger root artis of the finger tip point is defined as, it is hereby achieved that every hand The finger root artis of centre 3.It in addition to this, can be by during subsequent scaling and deformation for every hand Obtain two finger root artis of head and the tail of this hand;Alternatively, can also be adjacent by two (such as arbitrarily the selecting two) of this hand The distance between webs minimum point be used as finger reference width, then by two webs minimum points of the head and the tail of this hand respectively along cutting Line direction, extend outwardly half of finger reference width, the two finger root artis of head and the tail of obtained point respectively as this hand.
It should be noted that if the salient point found for single hand more than 5, can be by itself and hand structure mould Plate remove extra salient point during matching alignment.
It is matched to obtain 20 characteristic point Pl={ pl of the corresponding left hand of each frame image with such method as a result,1, pl2..., pl20And the right hand 20 characteristic point Pr={ pr1, pr2..., pr20}.It should be noted that if user's gesture is only It is then 20 characteristic point (referred to as features of the single hand in every frame image by matching above obtained comprising single hand Point set), i.e. Pl={ pl1, pl2..., pl20Or Pr={ pr1, pr2..., pr20}.Wherein, pl1,pl2,…,pl20Respectively The position of 20 characteristic points of left hand, and pr1,pr2,…,pr20The respectively position of 20 characteristic points of the right hand.
If user's gesture includes both hands, pass through the available left hand of the processing of above step 31 and step 3 two Characteristic point sequence { Pl to be measuredi, i=1,2 ..., M and the right hand characteristic point sequence { Pr to be measuredi, i=1,2 ..., M }.Its In, PliFor user's left hand, corresponding 20 (examples as predetermined number) are a in the i-th frame image of testing image sequence Characteristic point, and PriFor user's right hand corresponding 20 (as showing for predetermined number in the i-th frame image of testing image sequence Example) a characteristic point.
If user's gesture only includes single hand, every frame image in the testing image sequence captured is only comprising being somebody's turn to do The image of single hand, so that the characteristic point to be measured of the single hand can be obtained later by the processing of step 3 one and step 3 two Sequence, i.e. { Pli, i=1,2 ..., M } or { Pri, i=1,2 ..., M }.
In this way, executing step four as shown in Figure 1 after executing the step three.
In step 4, for every hand of user, the to be measured of this hand is determined in multiple default characteristic point sequences The matching sequence of characteristic point sequence, to determine denomination of dive and the position of this hand according to the matching sequence.
As a result, by step 4, the characteristic point sequence to be measured and each default characteristic point sequence progress of every hand are utilized Match, obtain matching sequence, and then quickly determines the movement of this hand according to the corresponding denomination of dive of matching sequence.
According to a kind of implementation, step 4 can be realized to step 4 four by step four as shown in Figure 3 one Processing.
Firstly, a hand motion list of file names is preset, including basic hand motion, such as:It waves, push away, drawing, opening and closing, turning There is unique name identification Deng, each movement and with normalized hand-characteristic point sequence (preset characteristic point sequence) The template of expression.It should be noted that every hand all has an above-mentioned hand motion name for the both hands of user List.That is, include in the hand motion list of file names (abbreviation left hand acts list of file names) of left hand is every for left hand A movement also has a left hand template (i.e. default characteristic point sequence of left hand other than being respectively provided with respective title Column);For the right hand, each movement for including in the hand motion list of file names of the right hand (the abbreviation right hand acts list of file names) in addition to It is respectively provided with except respective title, also there is a right hand template (i.e. a default characteristic point sequence of the right hand).
For example, multiple default characteristic point sequences of single hand are denoted as sequence A respectively1, sequence A2..., sequence AH, wherein H The sequence number that above-mentioned multiple default characteristic point sequences for the single hand are included, then in the hand motion list of file names of the single hand In:The name identification of movement 1 is " waving " and corresponding template (i.e. default characteristic point sequence) is sequence A1;The title mark of movement 2 Knowing is " pushing away " and corresponding template is sequence A2 ;…;The name identification for acting H is " turning " and corresponding template is sequence AH
In step 4 one, for the characteristic point sequence to be measured of every hand, according to predetermined time window by the characteristic point to be measured Sequences segmentation is multiple subsequences, and obtains the corresponding mean place of each subsequence.Wherein, each subsequence is corresponding average Position can choose average bit of the specific characteristic point (such as wrist midpoint, or can also be other characteristic points) in the subsequence It sets.Wherein, predetermined time window is about the time of a singlehanded elemental motion (i.e. singlehanded holds, grabs) from start to end, can It sets, or can be determined by the method for test based on experience value, such as can be 2.5 seconds.
In one example, it is assumed that characteristic point sequence to be measured was acquired in 10 seconds, can be with using 2.5 seconds time windows The characteristic point sequence to be measured of the characteristic point sequence to be measured of left hand and the right hand is divided into 4 subsequences respectively.With left hand to Survey characteristic point sequence { Pli, i=1,2 ..., M for the (characteristic point sequence { Pr to be measured of the right handi, i=1,2 ..., M } and its Similar, I will not elaborate), it is assumed that 10 frame images of acquisition per second, then it is 100 frame images, i.e. M that characteristic point sequence to be measured is corresponding =100, that is to say, that { Pli, i=1,2 ..., M } it include 100 groups of feature point set Pl1、Pl2、…、Pl100.In this way, by upper 2.5 seconds time windows are stated, it can be by { Pli, i=1,2 ..., M } it is divided into { Pli, i=1,2 ..., 25, { Pli, i=25, 26,...,50}、{Pli, i=51,52 ..., 75 } and { Pli, i=76,77 ..., 100 } and 4 subsequences, and every sub- sequence Each 25 frame image of correspondence is arranged, that is, each subsequence respectively includes 25 groups of feature point sets.Specific characteristic point chooses wrist midpoint, with Subsequence { Pli, i=1,2 ..., 25 } for (its excess-three sub- sequence handled to it similar, I will not elaborate), in wrist Point is in { Pli, i=1,2 ..., 25 } position in corresponding 25 groups of feature point sets is respectively position p1、p2、…、p25, then hand Wrist midpoint is in subsequence { Pli, i=1,2 ..., 25 in mean place be (p1+p2+…+p25)/25, as subsequence {Pli, i=1,2 ..., 25 } corresponding mean place.
Then, in step 4 two, for the corresponding each subsequence of every hand, by the subsequence and multiple default features Each of point sequence is matched respectively, is selected in multiple default characteristic point sequences high with the matching degree of the subsequence In preset matching threshold, (matching threshold can be set based on experience value, or can also be by the method for test come really It is fixed) and that maximum default characteristic point sequence, the matching sequence as the subsequence.Wherein it is possible to calculate subsequence with Similarity between default characteristic point sequence, as matching degree between the two.
It should be noted that for each subsequence not necessarily this can be found in multiple default characteristic point sequences The corresponding matching sequence of subsequence.When some subsequence for single hand, which does not find it, matches sequence, then by the sub- sequence The matching sequence of column is denoted as " sky ", but the mean place of the subsequence can not be " sky ".According to a kind of implementation, if sub- sequence The matching sequence of column is " sky ", then the mean place of the subsequence is set as " sky ";According to another implementation, if subsequence Matching sequence be " sky ", the mean place of the subsequence is the actual average position of specified characteristic point in the subsequence;According to The mean place of the subsequence is set as "+∞ " if the matching sequence of subsequence is " sky " by a kind of other implementations.
In addition, according to a kind of implementation, if there is no specific characteristic points (namely the specific characteristic is not present in subsequence The actual average position of point), the mean place of the subsequence can be set as "+∞ ".
Then, in step 4 three, the corresponding mean place of each subsequence is corresponding with the matching sequence of the subsequence Denomination of dive it is associated.
It, can be using the matching sequence of the corresponding each subsequence of this hand as this for every hand in step 4 four The corresponding multiple matching sequences of hand, and the corresponding denomination of dive of multiple matching sequence (after sorting in chronological order) is made For multiple denomination of dive of this hand.
For example, it is assumed that being { Pl for multiple subsequences of the characteristic point sequence to be measured of left handi, i=1,2 ..., 25 }, {Pli, i=25,26 ..., 50, { Pli, i=51,52 ..., 75 } and { Pli, i=76,77 ..., 100 }, respectively in left hand Multiple default characteristic point sequences in find { Pli, i=1,2 ..., 25, { Pli, i=25,26 ..., 50, { Pli, i= 51,52 ..., 75 matching sequence be followed successively by Pl1'、Pl2'、Pl3', and { Pl is not foundi, i=76,77 ..., 100 With sequence.Assuming that Pl1'、 Pl2'、Pl3' left hand movement list of file names in corresponding denomination of dive respectively be " waving ", " pushing away ", " drawing ", { Pli, i=1,2 ..., 25, { Pli, i=25,26 ..., 50, { Pli, i=51,52 ..., 75 } and { Pli,i =76,77 ..., 100 } respective mean place is respectively pm1、pm2、pm3And pm4, then the movement name of thus obtained left hand Claim and position includes:" waving " (position pm1);" pushing away " (position pm2);" drawing " (position pm3);" sky " (position " pm4").It needs to infuse It anticipates to being, in different embodiments, pm4It may be actual position value, it is also possible to " sky " or "+∞ " etc..
The corresponding multiple movements of every hand of user can be obtained to the processing of step 4 four by step 4 one as a result, Title (denomination of dive as this hand, that is to say, that the denomination of dive of this hand), and each denomination of dive has been respectively associated One mean place (it include one or more mean places as the position of this hand, in " position of this hand ", quantity and dynamic The quantity for making title is identical).For only identifying individual part as the identification technology of gesture, above-mentioned processing is identified The respective multiple movements of both hands and position, provide more flexible combination, on the one hand make the accuracy of identification of gesture more Height, the gesture on the other hand making it possible to identify are more various, abundant.
Finally, selection matches with the denomination of dive of user's both hands and position in default gesture table in step 5 Gesture, the gesture identification result as testing image sequence.
According to a kind of implementation, step 5 can be realized by step May Day as shown in Figure 4 and step 5 two Processing.
As shown in figure 4, step 5 one, predefined one make manually from two and the element of position two is arranged to the mapping of gesture Table is as default gesture table:The left end of each mapping is the position of set title pair and each denomination of dive pair;Each reflect The right end penetrated is a gesture HandSignal.Then, step 5 two is executed.
Wherein, " set title to " includes multiple denomination of dive pair, and each denomination of dive is to including that left hand acts name Claim ActNameleftWith right hand denomination of dive ActNameright, the position of each denomination of dive pair includes the opposite position of two hands It sets.
For example, in default gesture table, mapping one for (" drawing ", " sky "), (" drawing ", " drawing "), (" sky ", " conjunction "), (" sky ", " sky ") } (as element one), { (x1, y1), (x2, y2), (x3, y3), (x4, y4) (relative position, as element two) To the mapping of gesture " switch ";Mapping two for (" drawing ", " drawing "), (" opening ", " opening "), (" sky ", " sky "), (" sky ", " sky ") }, { (x5, y5), (x6, y6), (x7, y7), (x8, y8) arrive gesture " explosion " mapping;Etc..Wherein, each movement pair The denomination of dive on the left side is acted corresponding to left hand in (such as (" drawing ", " sky ")), and the denomination of dive on the right is acted corresponding to the right hand.
By taking mapping one as an example, (x1, y1) what is indicated is between left hand first element " drawing " and right hand first element " sky " Relative position (act to act left hand in (" drawings ", " sky ") and act the relative positions of corresponding two hands with the right hand); (x2, y2) what is indicated is the relative position between second, left hand movement " drawing " and second, right hand movement " drawing ";(x3, y3) table What is shown is the relative position between left hand third movement " sky " and right hand third movement " conjunction ";And (x4, y4) what is indicated is left Relative position between the 4th, hand movement " sky " and the 4th, right hand movement " sky ".Other mapping in elocutionary meaning with it is such Seemingly, it repeats no more.
In step 5 two, by the denomination of dive and position of the left end of each mapping in default gesture table and user's both hands It sets and is matched.
Wherein, the matching of denomination of dive executes stringent matching, that is, situation of verbatim account between two denomination of dive Lower the two denomination of dive of judgement are matched;And position is then that phase is calculated by the respective mean place of user's both hands To location information, and then calculating the similarity between the relative position information and the position for mapping left end (such as can be with come what is realized A similarity threshold is set, determines that position is matched when the similarity of calculating is greater than or equal to the similarity threshold).
For example, it is assumed that by step 4 obtain the respective denomination of dive of user's both hands be (" drawing ", " drawing "), (" opening ", " opening "), (" sky ", " sky "), (" sky ", " sky "), position be { (x11, y12)、(x21, y22)、(x31, y32)、(x41, y42) (right Answer left hand);(x'11, y '12)、(x’21, y '22)、(x’31, y '32)、 (x’41, y '42) (the corresponding right hand).
Then, the denomination of dive of user's both hands is matched with the left end of each mapping in default gesture table.
When being matched with mapping one, it can be deduced that, the left end of the denomination of dive and mapping one of user's both hands is moved Make title mismatch, therefore ignore mapping one, continues matching mapping two.
When being matched with mapping two, it can be deduced that, the left end of the denomination of dive and mapping two of user's both hands is moved Make title exact matching, then again matches the position of user's both hands with the relative position of the left end of mapping two.
During the relative position of the position of user's both hands and the left end of mapping two is carried out matched, calculate first The relative position of user's both hands is as follows:{(x'11-x11, y '12-y12)、(x’21-x21, y '22-y22)、 (x’31-x31, y '32- y32)、(x’41-x41, y '42-y42) (corresponding left hand).Then, by the above-mentioned relative position for the user's both hands being calculated with Map the relative position { (x of two left ends5, y5), (x6, y6), (x7, y7), (x8, y8) matched, that is, calculating { (x '11- x11, y '12-y12)、(x’21-x21, y '22-y22)、 (x’31-x31, y '32-y32)、(x’41-x41, y '42-y42) (corresponding left hand) with {(x5, y5), (x6, y6), (x7, y7), (x8, y8) between similarity, it is assumed that the similarity being calculated be 95%.In the example In son, if similarity threshold is 80%, two left ends of relative position and mapping for the user's both hands being calculated then are determined Relative position is matched.As a result, in this example embodiment, the result of gesture identification is " explosion ".
Above step May Day and step 5 two, by between the respective multiple movements of both hands and position and prearranged gesture table Matching, to determine that the gesture of user, the precision of identification are higher;When according to different application scene need change (such as modification, increase Add deduct few etc.) to the definition of gesture when, do not need modification algorithm or re -training classifier, can be only by adjusting predetermined hand The modes such as gesture title or the corresponding denomination of dive of gesture in gesture table realize the change to definition of gesture, substantially increase The adaptability of algorithm.
Above-mentioned gesture identification method according to an embodiment of the present invention using first identification single-handed exercise and then passes through both hands The mode of action recognition gesture realizes that can accurately identify the gesture of user, accuracy of identification is higher, and recognition speed compared with Fastly.
Above-mentioned gesture identification method of the invention uses Hierarchical Design algorithm, and algorithm complexity is low, is easy to implement.
It is right when needing to change (such as modify, increase or decrease) in addition, using above-mentioned gesture identification method of the invention When movement and/or the definition of gesture, can only (it be modified in advance that is, passing through by adjusting template and/or default gesture table to realize If the corresponding denomination of dive of characteristic point sequence changes the definition of movement, by increasing or decreasing default characteristic point sequence and its right Denomination of dive is answered to increase, subtract movement;Change the definition of gesture by modifying the corresponding multiple movements of gesture in default gesture table, Increased by increasing or decreasing gesture in default gesture table and its respective action, subtract gesture), without changing algorithm or again New training classifier, substantially increases the adaptability of algorithm.
In addition, the strong real-time of above-mentioned gesture identification method of the invention, can be suitble to the occasion of real-time interaction demand.
Although the embodiment according to limited quantity describes the present invention, above description, the art are benefited from It is interior it is clear for the skilled person that in the scope of the present invention thus described, it can be envisaged that other embodiments.Additionally, it should be noted that Language used in this specification primarily to readable and introduction purpose and select, rather than in order to explain or limit Determine subject of the present invention and selects.Therefore, without departing from the scope and spirit of the appended claims, for this Many modifications and changes are obvious for the those of ordinary skill of technical field.For the scope of the present invention, to this Invent done disclosure be it is illustrative and not restrictive, it is intended that the scope of the present invention be defined by the claims appended hereto.

Claims (6)

1. a kind of gesture identification method, which is characterized in that the gesture identification method includes:
Step 1: obtaining the testing image sequence of user's hand containing depth information;
Step 2: being detected in every frame image of the testing image sequence according to image depth information and image color information The hand profile of the user out;
Step 3: every hand of the user is directed to, using preset hand structure template, in the testing image sequence The characteristic point sequence to be measured of this hand is determined in every frame image;
Step 4: being directed to every hand of the user, the feature to be measured of this hand is determined in multiple default characteristic point sequences The matching sequence of point sequence, to determine denomination of dive and the position of this hand according to the matching sequence;
Step 5: the gesture that selection matches with the denomination of dive of user's both hands and position in default gesture table, makees For the gesture identification result of the testing image sequence;
Wherein, step 5 includes:
Step 5 one presets following map listing as the default gesture table:Each mapping in the map listing Left end be set title pair and each denomination of dive pair position;The right end of each mapping in the map listing is one A gesture;
Step 5 two, denomination of dive and position by the left end of each mapping in the default gesture table and user's both hands It sets and is matched, wherein the matching of denomination of dive executes stringent matching, and position is then respective average by user's both hands Relative position information is calculated in position, and then calculates the similarity between the relative position information and the position for mapping left end and come It realizes.
2. gesture identification method according to claim 1, which is characterized in that step 3 includes:
Step 3 one, every hand for the user, using preset hand structure template, in the testing image sequence Every frame image hand profile in determine this hand predetermined number characteristic point;
Step 3 two, every hand for the user, using this hand in each frame image of the testing image sequence Corresponding predetermined number characteristic point obtains the characteristic point sequence to be measured of this hand.
3. gesture identification method according to claim 1 or 2, which is characterized in that step 1 includes:
By capturing the image of user's hand in predetermined imaging region, Detection Method in Optical Image Sequences is obtainedWith Infrared image sequenceFor the pixel at Detection Method in Optical Image Sequences the i-th frame image coordinate (x, y) Value, andFor the pixel value at infrared image sequence the i-th frame image coordinate (x, y), taken out according to the following formula Take the image sequence of user's both hands information:
Wherein, α, β, λ are parameter preset threshold value,For the image sequence of user's both hands containing depth information of acquisition Column, as the testing image sequence, i=1,2 ..., M, the number of image frames that M includes by the testing image sequence.
4. gesture identification method according to claim 1 or 2, which is characterized in that step 2 includes:
For the testing image sequenceIn every frame imageThe color combining information deletion frame imageIn noise spot and non-area of skin color, using edge detection operator E () to deleting the noise spot and the non-skin Obtained image behind color regionEdge detection is carried out, edge image is obtained
The edge imageIt as only include the image of user's hand profile.
5. gesture identification method according to claim 2, which is characterized in that step 3 one includes:
Following processing is executed for every frame image of the testing image sequence:According to the profile curvature of a curve in the image, look for To the finger tip point in the contour line and refer to root artis;Using the finger tip point as setting base, each finger root singly referred to is matched Artis obtains benchmark of each length singly referred to as scaling;Based on the finger tip point and the position for referring to root artis It sets and each length singly referred to is zoomed in and out to the corresponding hand structure template and deformation, every hand is obtained by matching Each articulations digitorum manus characteristic point and wrist midpoint characteristic point;
Wherein, the hand structure template includes left-handed configuration template and right hand configurations template, the left-handed configuration template and the right side Hand stay in place form respectively includes:The fingertip characteristic point of each finger, each articulations digitorum manus characteristic point respectively refer to root joint characteristic point, in wrist Topological relation between point feature point and each characteristic point.
6. gesture identification method according to claim 1 or 2, which is characterized in that step 4 includes:
Step 4 one, for the characteristic point sequence to be measured of every hand, the characteristic point sequence to be measured is divided according to predetermined time window For multiple subsequences, and obtain the corresponding mean place of each subsequence;
Step 4 two is directed to the corresponding each subsequence of every hand, will be in the subsequence and the multiple default characteristic point sequence Each matched respectively, select to be higher than with the matching degree of the subsequence in the multiple default characteristic point sequence pre- If matching threshold and maximum default characteristic point sequence, the matching sequence as the subsequence;
It is step 4 three, the corresponding mean place of each subsequence is related to the corresponding denomination of dive of matching sequence of the subsequence Connection;
Step 4 four is directed to every hand, and the matching sequence of the corresponding each subsequence of this hand is corresponding multiple as this hand Sequence is matched, and using multiple matching corresponding denomination of dive of sequence as multiple denomination of dive of this hand.
CN201510563293.9A 2015-09-07 2015-09-07 A kind of gesture identification method Expired - Fee Related CN105160323B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510563293.9A CN105160323B (en) 2015-09-07 2015-09-07 A kind of gesture identification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510563293.9A CN105160323B (en) 2015-09-07 2015-09-07 A kind of gesture identification method

Publications (2)

Publication Number Publication Date
CN105160323A CN105160323A (en) 2015-12-16
CN105160323B true CN105160323B (en) 2018-11-27

Family

ID=54801175

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510563293.9A Expired - Fee Related CN105160323B (en) 2015-09-07 2015-09-07 A kind of gesture identification method

Country Status (1)

Country Link
CN (1) CN105160323B (en)

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106971131A (en) * 2016-01-14 2017-07-21 芋头科技(杭州)有限公司 A kind of gesture identification method based on center
CN106971130A (en) * 2016-01-14 2017-07-21 芋头科技(杭州)有限公司 A kind of gesture identification method using face as reference
CN106971135A (en) * 2016-01-14 2017-07-21 芋头科技(杭州)有限公司 A kind of slip gesture recognition methods
CN106327486B (en) * 2016-08-16 2018-12-28 广州视源电子科技股份有限公司 Method and device for tracking finger web position
CN106648068A (en) * 2016-11-11 2017-05-10 哈尔滨工业大学深圳研究生院 Method for recognizing three-dimensional dynamic gesture by two hands
CN106650628B (en) * 2016-11-21 2021-03-23 南京邮电大学 Fingertip detection method based on three-dimensional K curvature
CN106791763B (en) * 2016-11-24 2019-02-22 深圳奥比中光科技有限公司 A kind of application specific processor for 3D display and 3D interaction
CN107038424B (en) * 2017-04-20 2019-12-24 华中师范大学 Gesture recognition method
CN111045511B (en) * 2018-10-15 2022-06-07 华为技术有限公司 Gesture-based control method and terminal equipment
CN109685037B (en) * 2019-01-08 2021-03-05 北京汉王智远科技有限公司 Real-time action recognition method and device and electronic equipment
CN110298314A (en) * 2019-06-28 2019-10-01 海尔优家智能科技(北京)有限公司 The recognition methods of gesture area and device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426480A (en) * 2011-11-03 2012-04-25 康佳集团股份有限公司 Man-machine interactive system and real-time gesture tracking processing method for same
CN104281265A (en) * 2014-10-14 2015-01-14 京东方科技集团股份有限公司 Application program control method, application program control device and electronic equipment
CN104571482A (en) * 2013-10-22 2015-04-29 中国传媒大学 Digital device control method based on somatosensory recognition
CN104598915A (en) * 2014-01-24 2015-05-06 深圳奥比中光科技有限公司 Gesture recognition method and gesture recognition device
CN104750397A (en) * 2015-04-09 2015-07-01 重庆邮电大学 Somatosensory-based natural interaction method for virtual mine

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2003217587A1 (en) * 2002-02-15 2003-09-09 Canesta, Inc. Gesture recognition system using depth perceptive sensors

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102426480A (en) * 2011-11-03 2012-04-25 康佳集团股份有限公司 Man-machine interactive system and real-time gesture tracking processing method for same
CN104571482A (en) * 2013-10-22 2015-04-29 中国传媒大学 Digital device control method based on somatosensory recognition
CN104598915A (en) * 2014-01-24 2015-05-06 深圳奥比中光科技有限公司 Gesture recognition method and gesture recognition device
CN104281265A (en) * 2014-10-14 2015-01-14 京东方科技集团股份有限公司 Application program control method, application program control device and electronic equipment
CN104750397A (en) * 2015-04-09 2015-07-01 重庆邮电大学 Somatosensory-based natural interaction method for virtual mine

Also Published As

Publication number Publication date
CN105160323A (en) 2015-12-16

Similar Documents

Publication Publication Date Title
CN105160323B (en) A kind of gesture identification method
CN105045398B (en) A kind of virtual reality interactive device based on gesture identification
CN105069444B (en) A kind of gesture identifying device
US8254627B2 (en) Method for automatically following hand movements in an image sequence
CN105045399B (en) A kind of electronic equipment with 3D camera assemblies
CN107563494B (en) First-view-angle fingertip detection method based on convolutional neural network and heat map
CN104317391B (en) A kind of three-dimensional palm gesture recognition exchange method and system based on stereoscopic vision
CN103135758B (en) Realize the method and system of shortcut function
CN105302295B (en) A kind of virtual reality interactive device with 3D camera assemblies
CN105493078B (en) Colored sketches picture search
CN103488294B (en) A kind of Non-contact gesture based on user's interaction habits controls to map method of adjustment
CN105302294B (en) A kind of interactive virtual reality apparatus for demonstrating
KR100660725B1 (en) Portable terminal having apparatus for tracking human face
KR20160101973A (en) System and method for identifying faces in unconstrained media
CN110147767A (en) Three-dimension gesture attitude prediction method based on two dimensional image
CN106503651B (en) A kind of extracting method and system of images of gestures
CN105068662B (en) A kind of electronic equipment for man-machine interaction
US11282257B2 (en) Pose selection and animation of characters using video data and training techniques
CN109451634B (en) Gesture-based electric lamp control method and intelligent electric lamp system thereof
CN105046249B (en) A kind of man-machine interaction method
CN103996019A (en) System and method used for detecting and tracking a plurality of portions on an object
CN205080499U (en) Mutual equipment of virtual reality based on gesture recognition
CN109948450A (en) A kind of user behavior detection method, device and storage medium based on image
CN112215129A (en) Crowd counting method and system based on sequencing loss and double-branch network
US20140328516A1 (en) Gesture Recognition Method, An Apparatus and a Computer Program for the Same

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 150016 Heilongjiang Province, Harbin Economic Development Zone haping Road District Dalian road and Xingkai road junction

Applicant after: HARBIN YISHE TECHNOLOGY Co.,Ltd.

Address before: 150016 Heilongjiang City, Harbin province Daoli District, quiet street, unit 54, unit 2, layer 4, No. 3

Applicant before: HARBIN YISHE TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20181127