CN109190496A - A kind of monocular static gesture identification method based on multi-feature fusion - Google Patents

A kind of monocular static gesture identification method based on multi-feature fusion Download PDF

Info

Publication number
CN109190496A
CN109190496A CN201810900949.5A CN201810900949A CN109190496A CN 109190496 A CN109190496 A CN 109190496A CN 201810900949 A CN201810900949 A CN 201810900949A CN 109190496 A CN109190496 A CN 109190496A
Authority
CN
China
Prior art keywords
gesture
image
hand
value
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810900949.5A
Other languages
Chinese (zh)
Inventor
周智恒
许冰媛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201810900949.5A priority Critical patent/CN109190496A/en
Publication of CN109190496A publication Critical patent/CN109190496A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm
    • G06V40/113Recognition of static hand signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of monocular static gesture identification method based on multi-feature fusion, step includes: images of gestures acquisition, includes the RGB image of gesture with monocular cam acquisition;Image preprocessing, skin color segmentation is carried out using human body complexion information, using the geometrical characteristic of Morphological scale-space and combination hand, hand is separated with complex background, the positioning centre of the palm is operated by range conversion and removes arm regions existing for hand, obtains gesture bianry image;Gesture feature extracts, and calculates the perimeter and area ratio, Hu square and Fourier descriptor feature of gesture, constitutes gesture feature vector;Gesture identification, input gesture feature vector training BP neural network, realizes static gesture classification.The geometrical characteristic of present invention combination Skin Color Information and hand realizes accurate Hand Gesture Segmentation under monocular vision using Morphological scale-space and range conversion operation;By combination various gestures feature and training BP neural network, the high gesture classifier of strong robustness, accuracy rate is obtained.

Description

A kind of monocular static gesture identification method based on multi-feature fusion
Technical field
The present invention relates to field of image recognition, and in particular to a kind of monocular static gesture based on multi-feature fusion identification side Method.
Background technique
Gesture as it is a kind of from however intuitive interactive mode, gradually development is the research heat of field of human-computer interaction Point, and be widely used in terms of somatic sensation television game, robot control, computer.Compared to the hand based on data glove Gesture identification technology, the Gesture Recognition of view-based access control model has many advantages, such as low for equipment requirements, interaction nature, and becomes gesture and know Other main way.
Hand Gesture Segmentation is the key link in the gesture identification of view-based access control model, the influential effect of segmentation feature extraction, into And influence gesture classification result.In the static gesture identification method based on monocular vision, due to the influence of complex background environment, The result of Hand Gesture Segmentation is not satisfactory.With the appearance of kinect camera, depth information is used for gesture and separates with complex background Research in.It due to kinect camera higher cost, is not widely used, therefore such gesture identification method is unable to get popularization Using.The gesture feature that existing monocular static gesture identification method uses is relatively simple, leads to the robust of gesture recognition system Property is weaker, and recognition accuracy is not high.It is therefore proposed that a kind of can accurately divide under complex background and identify that the monocular of gesture is quiet State gesture identification method is current letter problem to be solved.
Summary of the invention
The purpose of the present invention is to solve drawbacks described above in the prior art, provide a kind of list based on multi-feature fusion Mesh static gesture identification method.
The purpose of the present invention can be reached by adopting the following technical scheme that:
A kind of monocular static gesture identification method based on multi-feature fusion, the recognition methods include:
Images of gestures acquisition step includes the RGB image of gesture with monocular cam acquisition;
Image preprocessing step, using human body complexion information carry out skin color segmentation, by image the colour of skin and class colour of skin area Domain extracts, and using the geometrical characteristic of Morphological scale-space and combination hand, hand is separated with complex background, passes through distance and becomes It changes the operation positioning centre of the palm and removes arm regions existing for hand, obtain gesture bianry image;
Gesture feature extraction step calculates the perimeter and area ratio, Hu square and Fourier descriptor feature of gesture, constitutes Gesture feature vector;
Gesture identification step passes through training BP nerve using the gesture feature vector of extraction as the input of BP neural network The classification of network implementations static gesture.
Further, the image preprocessing step includes:
The image of input is carried out the conversion of color space by skin color segmentation, is divided by Chroma threshold by the skin in image Color and class area of skin color extract, and obtain bianry image;
Morphological scale-space is carried out Morphological scale-space to the bianry image after skin color segmentation, is first operated using opening operation, then adopt It is operated with closed operation, eliminates the isolated noise in image;
The segmentation of hand geometry realizes that face and hand are separated with complex background, calculates remaining two connected regions The shape complexity C in domain extracts hand bianry image by judging with threshold value T;
By the arm removal in hand bianry image, gesture bianry image is obtained.
Further, the image by input carries out the conversion of color space, is divided by Chroma threshold by image In the colour of skin and class area of skin color extract, it is as follows to obtain bianry image process:
The image of input is transformed into YCr ' Cb ' color space, conversion formula by rgb color space by convert color spaces It is as follows:
Y=0.299 × r+0.587 × g+0.114 × b
Wherein r, g, b are respectively image three components of the red, green, blue in rgb color space, and y, cr', cb' are respectively Luminance component, red chrominance component, chroma blue component of the image in YCr ' Cb ' color space;
Chroma threshold segmentation, if two chromatic components of pixel meet the colour of skin in the threshold value model of cr' and cb' component simultaneously When enclosing, enable the pixel value be 1, otherwise value be 0, thus by image the colour of skin and class area of skin color extract, obtain To bianry image.
Further, hand geometry segmentation realizes that face and hand are separated with complex background, calculates surplus Under two connected regions shape complexity C, by with threshold value T judge, extract hand bianry image process it is as follows:
Area screening calculates the area of different connected regions in image, it is larger to extract area using eight connectivity distinguished number Two connected regions, realize separating for face and hand and complex background;
Shape complexity threshold decision calculates the shape complexity C of remaining two connected regions, if the shape of connected region Shape complexity C value is greater than threshold value T, then it is assumed that the region is non-hand region and removes, to obtain hand bianry image.
Further, the arm by hand bianry image removes, and it is as follows to obtain gesture bianry image process:
Centre of the palm positioning, is operated using range conversion, calculates minimum range of the hand pixel from hand boundary, and distance value takes For preimage vegetarian refreshments value, except remaining region value of hand is 0, the maximum picture of value in the image obtained after range conversion operation Vegetarian refreshments is the centre of the palm, and corresponding value is R0
Palm cutting will be less than R at a distance from the centre of the palm1Pixel value be 0, to remove palm area, wherein R1 =1.35 × R0
Threshold method judges that arm whether there is, and positions the pixel P of maximum value in image, and corresponding value is Pvalue, Calculate Pvalue/R0If value is greater than threshold value T1, then P point region is arm regions, continues next removal arm behaviour Make, arm regions are otherwise not present in hand bianry image, go to gesture feature extraction step;
Arm is removed, using eight connectivity distinguished number, removes P point region;
XOR operation finally obtains gesture bianry image using the XOR operation between image.
Further, the gesture feature extraction step includes:
The 7 invariant moments of gesture are calculated, Hu moment characteristics are constituted;
Using eight connectivity distinguished number, the perimeter and area of gesture area are calculated, calculates the ratio of perimeter and area;
Calculate the Fourier descriptor feature of gesture profile;
Combine Hu moment characteristics, perimeter and area ratio and Fourier descriptor feature, constitute 18 dimension gesture feature to Amount.
Further, the Fourier descriptor characteristic procedure of the calculating gesture profile is as follows:
By the coordinate { (x of gesture contour edgek,yk) complex representation is used, constitute sequence of complex numbers { ck, ckIt is expressed as follows:
ck=xk+iyk, k=0,1,2 ..., N-1;
To discrete series { ckMake Fourier transformation, Fourier coefficient sequence { C (u) } is obtained, formula is as follows:
10 Fourier coefficients that u=1 starts are extracted, to its modulus value and are normalized, Fourier descriptor feature is constituted.
Further, the BP neural network includes input layer, hidden layer and output layer, and input layer has d neuron, It is determined by the dimension of gesture feature vector, output layer has s neuron, is determined by gesture species number, and hidden layer has q nerve Member, the connection weight between h-th of neuron of i-th of neuron of input layer and hidden layer are vih, h-th of neuron of hidden layer with Connection weight between j-th of neuron of output layer is whj, the threshold value of h-th of neuron of hidden layer isJ-th of output layer The threshold value of neuron is θj
Further, before the gesture identification step, further includes:
BP neural network training step inputs the gesture feature vector training BP neural network of training sample, and process is as follows:
Random initializtion weight and threshold value, the value range for initializing weight is [- 1,1], the value model of initial threshold value It encloses for [- 0.5,0.5];
Input the gesture feature vector (x of training sample1,x2,...,x18);
Calculate the output data of each layer, wherein the BP neural network is using sigmoid function as each layer neuron Activation primitive, formula is as follows:
The output valve of h-th of neuron of hidden layer is αh, calculation formula is as follows:
The output valve of j-th of neuron of output layer isCalculation formula is as follows:
Mean square error E is calculated, calculation formula is as follows:
Wherein (y1,y2,...,y8) be training sample class label;
Parameter updates, and when E is greater than setting error, updates the weight and threshold value of network, using gradient descent method with amendment Current BP neural network;When E is less than setting error, deconditioning network obtains optimal model parameter.
Further, the calculation formula of the shape complexity C is as follows:
Wherein, A is the area of connected region, and p is the perimeter of connected region.
The present invention has the following advantages and effects with respect to the prior art:
(1) present invention utilizes the Extraction of Geometrical Features hand region of human body complexion information and hand, is grasped using range conversion It removes existing arm regions, realizes being precisely separating for gesture and complex background;
(2) present invention realizes strong robustness, a standard by combining a variety of effective gesture feature training BP neural networks The high gesture recognition system of true rate;
(3), identification low with equipment cost that the present invention is based on the static gesture identification methods of common monocular cam is accurately The advantages that rate is high, easy to promote and utilize.
Detailed description of the invention
Fig. 1 is the flow chart of monocular static gesture identification method based on multi-feature fusion disclosed in the present invention;
Fig. 2 is image preprocessing stream in monocular static gesture identification method based on multi-feature fusion disclosed in the present invention Cheng Tu;
Fig. 3 is that gesture feature extracts in monocular static gesture identification method based on multi-feature fusion disclosed in the present invention Flow chart;
Fig. 4 is gesture identification process in monocular static gesture identification method based on multi-feature fusion disclosed in the present invention Figure.
Specific embodiment
In order to make the object, technical scheme and advantages of the embodiment of the invention clearer, below in conjunction with the embodiment of the present invention In attached drawing, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that described embodiment is A part of the embodiment of the present invention, instead of all the embodiments.Based on the embodiments of the present invention, those of ordinary skill in the art Every other embodiment obtained without making creative work, shall fall within the protection scope of the present invention.
Embodiment
As shown in Figure 1, a kind of monocular static gesture identification method based on multi-feature fusion, process are as follows: gesture figure As acquisition step, image preprocessing step, gesture feature extraction step and gesture identification step.
Wherein, S1, images of gestures acquisition step:
It include the RGB image of gesture using monocular cam acquisition, the monocular cam should be located at immediately ahead of human body Position, face and hand are biggish two of area in all colours of skin and class area of skin color in the image of acquisition.
Wherein, S2, image preprocessing step:
As shown in Fig. 2, image preprocessing step process is as follows:
S201, skin color segmentation, detailed process is as follows:
The image of input is transformed into YCr ' Cb ' color space by rgb color space by S2011, convert color spaces, tool Body conversion formula is as follows:
Y=0.299 × r+0.587 × g+0.114 × b
Wherein r, g, b are respectively image three components of the red, green, blue in rgb color space;Y, cr', cb' are respectively Luminance component, red chrominance component, chroma blue component of the image in YCr ' Cb ' color space.
S2012, Chroma threshold segmentation, if two chromatic components of pixel meet the colour of skin in cr' and cb' component simultaneously When threshold range, enable the pixel value be 1, otherwise value be 0, thus by image the colour of skin and class area of skin color extract Come, obtains a secondary bianry image.
S202, Morphological scale-space, detailed process is as follows:
Morphological scale-space is carried out to the bianry image after skin color segmentation, is first operated using opening operation, then is grasped using closed operation Make, a large amount of isolated noise in image can be eliminated.
S203, the segmentation of hand geometry, detailed process is as follows:
S2031, area screening calculate the area of different connected regions in image, extract face using eight connectivity distinguished number Biggish two connected regions of product realize separating for face and hand and complex background.
S2032, shape complexity threshold decision calculate the shape complexity of remaining two connected regions, if connected region The shape complexity C value in domain is greater than threshold value T, then it is assumed that the region is non-hand region and removes, to obtain hand two-value Image, the present invention take T=0.3 best, shape complexity C's specific formula is as follows:
Wherein, A is the area of connected region, and p is the perimeter of connected region.
S204, arm removal, specific practice are as follows:
S2041, centre of the palm positioning, are operated using range conversion, calculate minimum range of the hand pixel from hand boundary, away from Replace preimage vegetarian refreshments value from value, except remaining region value of hand is 0, value is most in obtained image after range conversion operation Big pixel is the centre of the palm, and corresponding value is R0
S2042, palm cutting, will be less than R at a distance from the centre of the palm1Pixel value be 0, to remove palm area, The present invention takes R1=1.35 × R0Most preferably.
S2043, threshold method judge that arm whether there is, and position the pixel P of maximum value in image, and corresponding value is Pvalue, calculate Pvalue/R0If value is greater than threshold value T1, then P point region is arm regions, need to carry out removing arm behaviour Make, arm regions are otherwise not present in hand bianry image, can be directly used for extracting gesture feature, the present invention takes T1=0.35 most It is good.
S2044, arm is removed, using eight connectivity distinguished number, removes P point region.
S2045, XOR operation finally obtain gesture bianry image using the XOR operation between image.
Wherein, S3, gesture feature extraction step:
As shown in figure 3, extracting gesture feature to gesture bianry image, comprising the following steps:
S301, the 7 invariant moments for calculating gesture constitute Hu moment characteristics.
S302, using eight connectivity distinguished number, calculate the perimeter and area of gesture area, calculate the ratio of perimeter and area Value.
S303, the Fourier descriptor feature for calculating gesture profile, specific practice are as follows:
S3031, the coordinate { (x by gesture contour edgek,yk) complex representation is used, constitute sequence of complex numbers { ck, ckSpecifically It is expressed as follows:
ck=xk+iyk, k=0,1,2 ..., N-1;
S3032, to discrete series { ckMake Fourier transformation, Fourier coefficient sequence { C (u) } is obtained, specific formula is such as Under:
S3033,10 Fourier coefficients that u=1 starts are extracted, to its modulus value and normalized, constitute Fourier's description Subcharacter.
S304, the ratio and Fourier descriptor feature for combining Hu moment characteristics, perimeter and area, the gesture for constituting 18 dimensions are special Levy vector.
Wherein, S4, gesture identification step:
The present invention is realized using the gesture feature vector of extraction as the input of BP neural network by training BP neural network Static gesture classification.The BP neural network includes input layer, hidden layer and output layer, and input layer has d neuron, by hand The dimension of gesture feature vector determines that output layer has s neuron, determined by gesture species number, and hidden layer has q neuron, defeated Entering the connection weight between h-th of neuron of i-th of neuron of layer and hidden layer is vih, h-th of neuron of hidden layer and output Connection weight between j-th of neuron of layer is whj, the threshold value of h-th of neuron of hidden layer isJ-th of nerve of output layer The threshold value of member is θj.The present invention chooses d=18, q=10, s=8, and the present invention realizes the classification of 8 kinds of static gestures.
As shown in figure 4, before gesture identification step, further includes:
BP neural network training step inputs the gesture feature vector training BP neural network of training sample, detailed process It is as follows:
I. random initializtion weight and threshold value, the value range for initializing weight is [- 1,1], the value of initial threshold value Range is [- 0.5,0.5].
Ii. the gesture feature vector (x of training sample is inputted1,x2,...,x18)。
Iii. the output data of each layer is calculated:
BP neural network of the present invention is all made of activation primitive of the sigmoid function as each layer neuron, and specific formula is such as Under:
The output valve of h-th of neuron of hidden layer is αh, calculation formula is as follows:
The output valve of j-th of neuron of output layer isCalculation formula is as follows:
Iv. mean square error E is calculated, specific formula is as follows:
Wherein (y1,y2,...,y8) be training sample class label.
V. parameter updates, and when E is greater than setting error, updates the weight and threshold value of network, using gradient descent method to repair Proper preceding BP neural network;When E is less than setting error, deconditioning network obtains optimal model parameter.
Sorting phase carries out gesture classification with trained BP neural network model, and specific practice is as follows: input is to be sorted The gesture feature vector of sample, calculates the output data of each layer, obtains the output valve of network, the as classification results of gesture.
In conclusion present embodiment discloses a kind of monocular static gesture identification method based on multi-feature fusion, the party Method utilizes the Extraction of Geometrical Features hand region of human body complexion information and hand, removes existing arm using range conversion operation Region, realization gesture and complex background are precisely separating.This method is by combining a variety of effective gesture feature training BP nerves Network realizes the high gesture recognition system of strong robustness, an accuracy rate.In addition, this method is real using common monocular cam Existing static gesture identification, has many advantages, such as that equipment cost is low, recognition accuracy is high, easy to promote and utilize.
The above embodiment is a preferred embodiment of the present invention, but embodiments of the present invention are not by above-described embodiment Limitation, other any changes, modifications, substitutions, combinations, simplifications made without departing from the spirit and principles of the present invention, It should be equivalent substitute mode, be included within the scope of the present invention.

Claims (10)

1. a kind of monocular static gesture identification method based on multi-feature fusion, which is characterized in that the recognition methods includes:
Images of gestures acquisition step includes the RGB image of gesture with monocular cam acquisition;
Image preprocessing step, using human body complexion information carry out skin color segmentation, by image the colour of skin and class area of skin color mention It takes out, using the geometrical characteristic of Morphological scale-space and combination hand, hand is separated with complex background, is grasped by range conversion Make the positioning centre of the palm and remove arm regions existing for hand, obtains gesture bianry image;
Gesture feature extraction step calculates the perimeter and area ratio, Hu square and Fourier descriptor feature of gesture, constitutes gesture Feature vector;
Gesture identification step passes through training BP neural network using the gesture feature vector of extraction as the input of BP neural network Realize static gesture classification.
2. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 1, which is characterized in that The image preprocessing step includes:
The image of input is carried out the conversion of color space by skin color segmentation, by Chroma threshold segmentation by image the colour of skin and Class area of skin color extracts, and obtains bianry image;
Morphological scale-space is carried out Morphological scale-space to the bianry image after skin color segmentation, is first operated using opening operation, then used and close Arithmetic operation eliminates the isolated noise in image;
The segmentation of hand geometry realizes that face and hand are separated with complex background, calculates remaining two connected regions Shape complexity C extracts hand bianry image by judging with threshold value T;
By the arm removal in hand bianry image, gesture bianry image is obtained.
3. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 2, which is characterized in that The described image by input carries out the conversion of color space, by Chroma threshold segmentation by image the colour of skin and class colour of skin area Domain extracts, and it is as follows to obtain bianry image process:
The image of input is transformed into YCr ' Cb ' color space by rgb color space by convert color spaces, and conversion formula is as follows:
Y=0.299 × r+0.587 × g+0.114 × b
Wherein r, g, b are respectively image three components of the red, green, blue in rgb color space, and y, cr', cb' are respectively image Luminance component, red chrominance component, chroma blue component in YCr ' Cb ' color space;
Chroma threshold segmentation, if two chromatic components of pixel meet the colour of skin in the threshold range of cr' and cb' component simultaneously When, enable the pixel value be 1, otherwise value be 0, thus by image the colour of skin and class area of skin color extract, obtain Bianry image.
4. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 2, which is characterized in that The hand geometry segmentation, realizes that face and hand are separated with complex background, calculates remaining two connected regions Shape complexity C, by with threshold value T judge, extract hand bianry image process it is as follows:
Area screening calculates the area of different connected regions in image, extracts area biggish two using eight connectivity distinguished number A connected region, realization face and hand are separated with complex background;
Shape complexity threshold decision calculates the shape complexity C of remaining two connected regions, if the shape of connected region is multiple Miscellaneous degree C value is greater than threshold value T, then it is assumed that the region is non-hand region and removes, to obtain hand bianry image.
5. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 2, which is characterized in that Described removes the arm in hand bianry image, and it is as follows to obtain gesture bianry image process:
Centre of the palm positioning, is operated using range conversion, calculates minimum range of the hand pixel from hand boundary, and distance value replaces former Pixel value, except remaining region value of hand is 0, the maximum pixel of value in the image obtained after range conversion operation For the centre of the palm, corresponding value is R0
Palm cutting will be less than R at a distance from the centre of the palm1Pixel value be 0, to remove palm area, wherein R1= 1.35×R0
Threshold method judges that arm whether there is, and positions the pixel P of maximum value in image, and corresponding value is Pvalue, calculate Pvalue/R0If value is greater than threshold value T1, then P point region is arm regions, continue next removal arm operation, Otherwise arm regions are not present in hand bianry image, go to gesture feature extraction step;
Arm is removed, using eight connectivity distinguished number, removes P point region;
XOR operation finally obtains gesture bianry image using the XOR operation between image.
6. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 1, which is characterized in that The gesture feature extraction step includes:
The 7 invariant moments of gesture are calculated, Hu moment characteristics are constituted;
Using eight connectivity distinguished number, the perimeter and area of gesture area are calculated, calculates the ratio of perimeter and area;
Calculate the Fourier descriptor feature of gesture profile;
The ratio and Fourier descriptor feature of Hu moment characteristics, perimeter and area are combined, the gesture feature vector of 18 dimensions is constituted.
7. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 6, which is characterized in that The Fourier descriptor characteristic procedure of the calculating gesture profile is as follows:
By the coordinate { (x of gesture contour edgek, yk) complex representation is used, constitute sequence of complex numbers { ck, ckIt is expressed as follows:
ck=xk+iyk, k=0,1,2 ..., N-1;
To discrete series { ckMake Fourier transformation, Fourier coefficient sequence { C (u) } is obtained, formula is as follows:
10 Fourier coefficients that u=1 starts are extracted, to its modulus value and are normalized, Fourier descriptor feature is constituted.
8. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 1, which is characterized in that The BP neural network includes input layer, hidden layer and output layer, and input layer has d neuron, by gesture feature vector Dimension determines that output layer has s neuron, determined by gesture species number, and hidden layer has q neuron, i-th of nerve of input layer Connection weight between member and h-th of neuron of hidden layer is vih, h-th of neuron of hidden layer and j-th of neuron of output layer Between connection weight be whj, the threshold value of h-th of neuron of hidden layer isThe threshold value of j-th of neuron of output layer is θj
9. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 8, which is characterized in that Before the gesture identification step, further includes:
BP neural network training step inputs the gesture feature vector training BP neural network of training sample, and process is as follows:
Random initializtion weight and threshold value, the value range for initializing weight is [- 1,1], and the value range of initial threshold value is [-0.5,0.5];
Input the gesture feature vector (x of training sample1,x2,...,x18);
Calculate the output data of each layer, wherein the BP neural network swashing as each layer neuron using sigmoid function Function living, formula are as follows:
The output valve of h-th of neuron of hidden layer is αh, calculation formula is as follows:
The output valve of j-th of neuron of output layer isCalculation formula is as follows:
Mean square error E is calculated, calculation formula is as follows:
Wherein (y1,y2,...,y8) be training sample class label;
Parameter updates, and when E is greater than setting error, the weight and threshold value of network are updated using gradient descent method, current to correct BP neural network;When E is less than setting error, deconditioning network obtains optimal model parameter.
10. a kind of monocular static gesture identification method based on multi-feature fusion according to claim 4, feature exist In the calculation formula of the shape complexity C is as follows:
Wherein, A is the area of connected region, and p is the perimeter of connected region.
CN201810900949.5A 2018-08-09 2018-08-09 A kind of monocular static gesture identification method based on multi-feature fusion Pending CN109190496A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810900949.5A CN109190496A (en) 2018-08-09 2018-08-09 A kind of monocular static gesture identification method based on multi-feature fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810900949.5A CN109190496A (en) 2018-08-09 2018-08-09 A kind of monocular static gesture identification method based on multi-feature fusion

Publications (1)

Publication Number Publication Date
CN109190496A true CN109190496A (en) 2019-01-11

Family

ID=64921162

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810900949.5A Pending CN109190496A (en) 2018-08-09 2018-08-09 A kind of monocular static gesture identification method based on multi-feature fusion

Country Status (1)

Country Link
CN (1) CN109190496A (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109934159A (en) * 2019-03-11 2019-06-25 西安邮电大学 A kind of gesture identification method of multiple features fusion
CN110796033A (en) * 2019-10-12 2020-02-14 江苏科技大学 Static gesture recognition method based on bounding box model
CN111160194A (en) * 2019-12-23 2020-05-15 浙江理工大学 Static gesture image recognition method based on multi-feature fusion
CN111258430A (en) * 2020-01-21 2020-06-09 哈尔滨拓博科技有限公司 Desktop interaction system based on monocular gesture control
CN111339970A (en) * 2020-03-02 2020-06-26 上海化学工业区公共管廊有限公司 Smoking behavior detection method suitable for public environment
WO2020173024A1 (en) * 2019-02-26 2020-09-03 南京邮电大学 Multi-gesture precise segmentation method for smart home scenario
CN111901681A (en) * 2020-05-04 2020-11-06 东南大学 Intelligent television control device and method based on face recognition and gesture recognition
WO2020237519A1 (en) * 2019-05-29 2020-12-03 深圳大学 Identification method, apparatus and device, and storage medium
CN112034981A (en) * 2020-08-20 2020-12-04 深圳创维-Rgb电子有限公司 Display terminal control method, display terminal, and computer-readable storage medium
CN112068705A (en) * 2020-09-15 2020-12-11 山东建筑大学 Bionic robot fish interaction control method and system based on gesture recognition
CN112232217A (en) * 2020-10-16 2021-01-15 怀化新大地电脑有限公司 Gesture recognition system
CN112906550A (en) * 2021-02-09 2021-06-04 哈尔滨理工大学 Static gesture recognition method based on watershed transformation

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106293057A (en) * 2016-07-20 2017-01-04 西安中科比奇创新科技有限责任公司 Gesture identification method based on BP neutral net
CN106503619A (en) * 2016-09-23 2017-03-15 南京理工大学 Gesture identification method based on BP neural network
CN106503651A (en) * 2016-10-21 2017-03-15 上海未来伙伴机器人有限公司 A kind of extracting method of images of gestures and system
CN107133562A (en) * 2017-03-17 2017-09-05 华南理工大学 A kind of gesture identification method based on extreme learning machine
CN108108648A (en) * 2016-11-24 2018-06-01 广州映博智能科技有限公司 A kind of new gesture recognition system device and method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106293057A (en) * 2016-07-20 2017-01-04 西安中科比奇创新科技有限责任公司 Gesture identification method based on BP neutral net
CN106503619A (en) * 2016-09-23 2017-03-15 南京理工大学 Gesture identification method based on BP neural network
CN106503651A (en) * 2016-10-21 2017-03-15 上海未来伙伴机器人有限公司 A kind of extracting method of images of gestures and system
CN108108648A (en) * 2016-11-24 2018-06-01 广州映博智能科技有限公司 A kind of new gesture recognition system device and method
CN107133562A (en) * 2017-03-17 2017-09-05 华南理工大学 A kind of gesture identification method based on extreme learning machine

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
周龙: "基于神经网络方法的手势行为识别技术研究", 《中国优秀硕士论文全文数据库信息科技辑》 *
徐常青等: "《数学实验与软件计算》", 31 January 2014 *
曹建秋等: "基于改进YCrCb颜色空间的肤色分割", 《重庆交通大学学报自然科学版》 *
杨帆等: "《精通图像处理经典算法(MATLAB版)》", 30 April 2014 *
赵小川等: "《MATLAB数字图像处理-从仿真到C/C++代码的自动生成》", 30 September 2015 *

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2021517281A (en) * 2019-02-26 2021-07-15 南京郵電大学Nanjing University Of Posts And Telecommunications Multi-gesture fine division method for smart home scenes
WO2020173024A1 (en) * 2019-02-26 2020-09-03 南京邮电大学 Multi-gesture precise segmentation method for smart home scenario
CN109934159A (en) * 2019-03-11 2019-06-25 西安邮电大学 A kind of gesture identification method of multiple features fusion
WO2020237519A1 (en) * 2019-05-29 2020-12-03 深圳大学 Identification method, apparatus and device, and storage medium
CN110796033A (en) * 2019-10-12 2020-02-14 江苏科技大学 Static gesture recognition method based on bounding box model
CN110796033B (en) * 2019-10-12 2023-07-28 江苏科技大学 Static gesture recognition method based on bounding box model
CN111160194A (en) * 2019-12-23 2020-05-15 浙江理工大学 Static gesture image recognition method based on multi-feature fusion
CN111258430A (en) * 2020-01-21 2020-06-09 哈尔滨拓博科技有限公司 Desktop interaction system based on monocular gesture control
CN111339970A (en) * 2020-03-02 2020-06-26 上海化学工业区公共管廊有限公司 Smoking behavior detection method suitable for public environment
CN111339970B (en) * 2020-03-02 2023-04-07 上海化学工业区公共管廊有限公司 Smoking behavior detection method suitable for public environment
CN111901681B (en) * 2020-05-04 2022-09-30 东南大学 Intelligent television control device and method based on face recognition and gesture recognition
CN111901681A (en) * 2020-05-04 2020-11-06 东南大学 Intelligent television control device and method based on face recognition and gesture recognition
CN112034981A (en) * 2020-08-20 2020-12-04 深圳创维-Rgb电子有限公司 Display terminal control method, display terminal, and computer-readable storage medium
CN112068705A (en) * 2020-09-15 2020-12-11 山东建筑大学 Bionic robot fish interaction control method and system based on gesture recognition
CN112232217A (en) * 2020-10-16 2021-01-15 怀化新大地电脑有限公司 Gesture recognition system
CN112232217B (en) * 2020-10-16 2022-08-02 怀化新大地电脑有限公司 Gesture recognition system
CN112906550A (en) * 2021-02-09 2021-06-04 哈尔滨理工大学 Static gesture recognition method based on watershed transformation
CN112906550B (en) * 2021-02-09 2022-07-19 哈尔滨理工大学 Static gesture recognition method based on watershed transformation

Similar Documents

Publication Publication Date Title
CN109190496A (en) A kind of monocular static gesture identification method based on multi-feature fusion
Nikam et al. Sign language recognition using image based hand gesture recognition techniques
CN104268583B (en) Pedestrian re-recognition method and system based on color area features
CN108875787A (en) A kind of image-recognizing method and device, computer equipment and storage medium
CN108256421A (en) A kind of dynamic gesture sequence real-time identification method, system and device
CN106778785B (en) Construct the method for image Feature Selection Model and the method, apparatus of image recognition
CN109558832A (en) A kind of human body attitude detection method, device, equipment and storage medium
Konwar et al. An American sign language detection system using HSV color model and edge detection
CN111274921B (en) Method for recognizing human body behaviors by using gesture mask
CN109214297A (en) A kind of static gesture identification method of combination depth information and Skin Color Information
Bilal et al. A hybrid method using haar-like and skin-color algorithm for hand posture detection, recognition and tracking
CN106446862A (en) Face detection method and system
CN111158491A (en) Gesture recognition man-machine interaction method applied to vehicle-mounted HUD
CN112906550B (en) Static gesture recognition method based on watershed transformation
Vishwakarma et al. Simple and intelligent system to recognize the expression of speech-disabled person
CN106097354A (en) A kind of combining adaptive Gauss Face Detection and the hand images dividing method of region growing
Barkoky et al. Static hand gesture recognition of Persian sign numbers using thinning method
CN111080670A (en) Image extraction method, device, equipment and storage medium
CN110046544A (en) Digital gesture identification method based on convolutional neural networks
CN110956099A (en) Dynamic gesture instruction identification method
CN111160194B (en) Static gesture image recognition method based on multi-feature fusion
CN103778430B (en) Rapid face detection method based on combination between skin color segmentation and AdaBoost
CN106909884A (en) A kind of hand region detection method and device based on hierarchy and deformable part sub-model
Lei et al. A novel side face contour extraction algorithm for driving fatigue statue recognition
Sokhib et al. A combined method of skin-and depth-based hand gesture recognition.

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190111