CN107480600A - A kind of gesture identification method based on depth convolutional neural networks - Google Patents

A kind of gesture identification method based on depth convolutional neural networks Download PDF

Info

Publication number
CN107480600A
CN107480600A CN201710597440.3A CN201710597440A CN107480600A CN 107480600 A CN107480600 A CN 107480600A CN 201710597440 A CN201710597440 A CN 201710597440A CN 107480600 A CN107480600 A CN 107480600A
Authority
CN
China
Prior art keywords
mrow
convolutional neural
neural networks
hand
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710597440.3A
Other languages
Chinese (zh)
Inventor
王修晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Jiliang University
Original Assignee
China Jiliang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Jiliang University filed Critical China Jiliang University
Priority to CN201710597440.3A priority Critical patent/CN107480600A/en
Publication of CN107480600A publication Critical patent/CN107480600A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm
    • G06V40/113Recognition of static hand signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of gesture identification method based on depth convolutional neural networks, including:(1) division of edge detection process and sample set is carried out to the sample image of training set;(2)Build depth convolutional neural networks;(3)Determine activation primitive and loss function;(4)Train deep neural network;(5)Gesture identification is realized according to the depth convolutional neural networks after training:Its step includes:A) hand-type image is extracted from gesture data to be identified;B) hand-type image is subjected to rim detection and size normalized;C) normalized hand-type image is input in depth convolutional neural networks, the output valve according to output layer judges the ownership class of current gesture.The present invention uses multiple down-sampling technique construction depth convolutional neural networks, and carries out the training of neutral net as activation primitive using hyperbolic tangent function, can not only improve the efficiency of gesture identification, and can improve the accuracy rate of gesture identification.

Description

A kind of gesture identification method based on depth convolutional neural networks
Technical field
The present invention relates to living things feature recognition field, more particularly to a kind of gesture identification based on depth convolutional neural networks Method.
Background technology
Living things feature recognition is one of key technologies in field such as video monitoring, safety certification.Biological characteristic can be divided into Physiological characteristic and behavioural characteristic.Physiological characteristic mainly includes face, fingerprint and iris etc., and behavioural characteristic then includes gait, gesture Deng.Typically the recognition methods based on physiological characteristic has fingerprint recognition, palm shape and outline identification, recognition of face, and iris is known Not etc..Fingerprint recognition is one of current most widely used personal identification method based on biological characteristic.Fingerprint recognition has skill The advantages that art is ripe, and cost is cheap.Its shortcoming is contact, has the property invaded, and the problem of health aspect, while fingerprint be present It is and easy to wear.Face recognition technology is a very active in recent years research field, have intuitive it is good, conveniently, friend Well, the advantages of being easily accepted.Recognition of face is contactless, passive discerning, it is not necessary to which people's cooperates with one's own initiative;But shortcoming The influence of illumination, visual angle, shelter, environment, expression etc. is susceptible to, causes identification difficult.The safety of iris feature identification Degree and precision are very high, but collection apparatus process is extremely difficult.The identity recognizing technology of Behavior-based control feature, common are Gait Recognition and gesture identification.The input of Gait Recognition is the sequence of video images of one section of walking, and data volume is very big, causes to calculate Complexity is high, deals with relatively difficult.And important component of the gesture identification as non-contact type human-machine interaction, it is big at present In terms of most researchers are primarily focused on the final identification of gesture, it will usually gesture background is simplified, and in the single back of the body Gesture is split using the algorithm studied under scape, then passed through the implication that gesture is expressed using conventional recognition methods Network analysis comes out.
The content of the invention
In view of the above-mentioned deficiencies in the prior art, it is an object of the present invention to provide a kind of gesture based on depth convolutional neural networks Recognition methods.
The purpose of the present invention is achieved through the following technical solutions:A kind of gesture identification based on deep neural network Method, comprise the following steps:
(1) division of edge detection process and sample set is carried out to the sample image of training set:First to the hand of training Power set image carries out hand-type detection and edge detection process, and by the hand-type Image Adjusting of extraction to unified size;It will locate in advance Data after reason are divided into training sample set and checking sample set;
(2) depth convolutional neural networks are built:If I and O are respectively the input layer and output layer of depth convolutional neural networks, Hidden layer between I and O is H1, H2 ..., Hn.Wherein, the hand-type image that input layer obtains for step (1), output layer one The gesture feature vector that individual length is N, hidden layer use multiple down-sampling technology, it is allowed to have between the piecemeal of down-sampling overlapping;
(3) activation primitive and loss function are determined:The non-linear hyperbolic tan shown in formula (1) is selected as nerve The activation primitive of member;Select the loss function shown in formula (2).Wherein, n is the number of sample in training set, and x is hand-type figure Picture, y are the gesture feature vector corresponding to x, and θ is parameter vector;
(4) deep neural network is trained:Concentrated from training sample and choose m training sample, using steepest descent method Calculate gradient;Then, verified using checking sample set, when accuracy exceedes predetermined threshold value 99.5%, terminate training, from And obtain with the deep neural network for determining weight w and bias term b;
(5) gesture identification is realized according to the depth convolutional neural networks after training:Its step includes:A) from gesture to be identified Extracting data hand-type image;B) hand-type image is subjected to rim detection and size normalized;C) hand-type image is inputted Into depth convolutional neural networks, the output valve according to output layer judges the ownership class of current gesture.
The beneficial effects of the invention are as follows:The gesture identification method based on depth convolutional neural networks of the present invention, using more Weight down-sampling technique construction depth convolutional neural networks, and neutral net is carried out as activation primitive using hyperbolic tangent function Training, can not only improve the efficiency of gesture identification, and can improve the accuracy rate of gesture identification.
Brief description of the drawings
Fig. 1 is the implementing procedure of this method;
Fig. 2 hand-type image zooming-out schematic diagrames;
The multiple down-sampling schematic diagrames of Fig. 3;
Fig. 4 hyperbolic tangent function curves;
The error rate that Fig. 5 discriminations compare.
Embodiment
The present invention is described in detail below in conjunction with the accompanying drawings.
As shown in figure 1, the gesture identification method of the invention based on depth convolutional neural networks is divided into training stage and identification Stage.
In the training stage, depth convolutional neural networks are built, and its weights and offset parameter are determined using training set data Value.Specifically include following sub-step:
1st, hand-type image is extracted from training data:For the frame in gesture interaction, as shown in Fig. 2 (a), the present invention is first Rough hand region image is first extracted according to features of skin colors, as shown in Fig. 2 (b);Secondly, it is swollen using filtering and quick ecological Swollen, erosion algorithm is modified to hand area image, further hand-type image is obtained, as shown in Fig. 2 (c);
2nd, rim detection and size normalized are carried out to hand-type image:The hand-type image obtained to step 1 carries out side Edge detects and binary conversion treatment, preliminary hand-type contour images is obtained, as shown in Fig. 2 (d);Then, hand-type contour curve is entered Row trimming and perfect, and be adjusted to unified size obtain corresponding to hand-type image;
3rd, depth convolutional neural networks are built:The hand-type image obtained for step 2, the present invention construct depth convolution god It is used for gesture identification through network.If I and O are respectively the input layer and output layer of depth convolutional neural networks, hidden between I and O Tibetan layer be H1, H2 ..., Hn.Wherein, input layer is the hand-type image that step 2 obtains, and output layer is the gesture that a length is N Characteristic vector, hidden layer use multiple down-sampling technology, it is allowed to have between the piecemeal of down-sampling it is overlapping, as shown in Figure 3;
4th, activation primitive and loss function are determined:As shown in figure 4, selection non-linear hyperbolic tan is as neuron Activation primitive;The squared error function shown in formula (2) is selected as loss function.
5th, depth convolutional neural networks are trained:
Concentrated from training sample and choose m training sample, gradient is calculated using steepest descent method, and according to loss letter Several parameters to each hidden layer are iterated optimization;Then, verified using checking sample set, when accuracy exceedes default threshold During value 99.5%, terminate training, so as to obtain with the deep neural network for determining weight w and bias term b.
In cognitive phase, first from extracting data hand-type image to be identified, and rim detection and size normalization are carried out Processing;Then, the depth convolutional neural networks that train are entered into judge the ownership class of current gesture.Finally, it incite somebody to action this Method compares with the method using data glove and sequence similarity detection (SSDA) method, and Fig. 5 gives mistake Rate experimental data.

Claims (1)

1. a kind of gesture identification method based on depth convolutional neural networks, comprises the following steps:
(1) division of edge detection process and sample set is carried out to the sample image of training set:First to the gesture collection of training Image carries out hand-type detection and edge detection process, and by the hand-type Image Adjusting of extraction to unified size;After pre-processing Data be divided into training sample set and checking sample set;
(2) depth convolutional neural networks are built:If I and O are respectively the input layer and output layer of depth convolutional neural networks, I and O Between hidden layer for H1, H2 ..., Hn.Wherein, input layer is the hand-type image that step (1) obtains, and output layer is a length For N gesture feature vector, hidden layer uses multiple down-sampling technology, it is allowed to has between the piecemeal of down-sampling overlapping;
(3) activation primitive and loss function are determined:The non-linear hyperbolic tan shown in formula (1) is selected as neuron Activation primitive;Select the loss function shown in formula (2).Wherein, n is the number of sample in training set, and x is in hand-type image Point, y are the output valve corresponding to x, and θ is parameter vector;
<mrow> <mi>tanh</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>)</mo> </mrow> <mo>=</mo> <mfrac> <mrow> <mn>1</mn> <mo>-</mo> <msup> <mi>e</mi> <mrow> <mo>-</mo> <mn>2</mn> <mi>x</mi> </mrow> </msup> </mrow> <mrow> <mn>1</mn> <mo>+</mo> <msup> <mi>e</mi> <mrow> <mo>-</mo> <mn>2</mn> <mi>x</mi> </mrow> </msup> </mrow> </mfrac> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>1</mn> <mo>)</mo> </mrow> </mrow>
<mrow> <mi>L</mi> <mo>=</mo> <mfrac> <mn>1</mn> <mi>n</mi> </mfrac> <munder> <mo>&amp;Sigma;</mo> <mi>x</mi> </munder> <mo>|</mo> <mo>|</mo> <mi>y</mi> <mo>-</mo> <mi>f</mi> <mrow> <mo>(</mo> <mi>x</mi> <mo>,</mo> <mi>&amp;theta;</mi> <mo>)</mo> </mrow> <mo>|</mo> <msup> <mo>|</mo> <mn>2</mn> </msup> <mo>-</mo> <mo>-</mo> <mo>-</mo> <mrow> <mo>(</mo> <mn>2</mn> <mo>)</mo> </mrow> </mrow>
(4) deep neural network is trained:Concentrated from training sample and choose m training sample, calculated using steepest descent method Gradient;Then, verified using checking sample set, when accuracy exceedes predetermined threshold value, terminate training, so as to be had Determine weight w and bias term b deep neural network;
(5) gesture identification is realized according to the depth convolutional neural networks after training:Its step includes:A) from gesture data to be identified Middle extraction hand-type image;B) hand-type image is subjected to rim detection and size normalized;C) hand-type image is input to depth Spend in convolutional neural networks, the output valve according to output layer judges the ownership class of current gesture.
CN201710597440.3A 2017-07-20 2017-07-20 A kind of gesture identification method based on depth convolutional neural networks Pending CN107480600A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710597440.3A CN107480600A (en) 2017-07-20 2017-07-20 A kind of gesture identification method based on depth convolutional neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710597440.3A CN107480600A (en) 2017-07-20 2017-07-20 A kind of gesture identification method based on depth convolutional neural networks

Publications (1)

Publication Number Publication Date
CN107480600A true CN107480600A (en) 2017-12-15

Family

ID=60595146

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710597440.3A Pending CN107480600A (en) 2017-07-20 2017-07-20 A kind of gesture identification method based on depth convolutional neural networks

Country Status (1)

Country Link
CN (1) CN107480600A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108334814A (en) * 2018-01-11 2018-07-27 浙江工业大学 A kind of AR system gesture identification methods based on convolutional neural networks combination user's habituation behavioural analysis
CN109117742A (en) * 2018-07-20 2019-01-01 百度在线网络技术(北京)有限公司 Gestures detection model treatment method, apparatus, equipment and storage medium
CN109890573A (en) * 2019-01-04 2019-06-14 珊口(上海)智能科技有限公司 Control method, device, mobile robot and the storage medium of mobile robot
CN111338470A (en) * 2020-02-10 2020-06-26 烟台持久钟表有限公司 Method for controlling big clock through gestures
CN112889075A (en) * 2018-10-29 2021-06-01 Sk电信有限公司 Improving prediction performance using asymmetric hyperbolic tangent activation function
CN113703581A (en) * 2021-09-03 2021-11-26 广州朗国电子科技股份有限公司 Window adjusting method based on gesture switching, electronic whiteboard and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529470A (en) * 2016-11-09 2017-03-22 济南大学 Gesture recognition method based on multistage depth convolution neural network
US20170161607A1 (en) * 2015-12-04 2017-06-08 Pilot Ai Labs, Inc. System and method for improved gesture recognition using neural networks

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170161607A1 (en) * 2015-12-04 2017-06-08 Pilot Ai Labs, Inc. System and method for improved gesture recognition using neural networks
CN106529470A (en) * 2016-11-09 2017-03-22 济南大学 Gesture recognition method based on multistage depth convolution neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
LEE的白板报: "卷积神经网络初探", 《斯坦福深度学习HTTPS://MY.OSCHINA.NET/FINDBILL/BLOG/550565》 *
蔡娟: "基于卷积神经网络的手势识别", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108334814A (en) * 2018-01-11 2018-07-27 浙江工业大学 A kind of AR system gesture identification methods based on convolutional neural networks combination user's habituation behavioural analysis
CN108334814B (en) * 2018-01-11 2020-10-30 浙江工业大学 Gesture recognition method of AR system
CN109117742A (en) * 2018-07-20 2019-01-01 百度在线网络技术(北京)有限公司 Gestures detection model treatment method, apparatus, equipment and storage medium
CN109117742B (en) * 2018-07-20 2022-12-27 百度在线网络技术(北京)有限公司 Gesture detection model processing method, device, equipment and storage medium
CN112889075A (en) * 2018-10-29 2021-06-01 Sk电信有限公司 Improving prediction performance using asymmetric hyperbolic tangent activation function
CN112889075B (en) * 2018-10-29 2024-01-26 Sk电信有限公司 Improved predictive performance using asymmetric hyperbolic tangent activation function
CN109890573A (en) * 2019-01-04 2019-06-14 珊口(上海)智能科技有限公司 Control method, device, mobile robot and the storage medium of mobile robot
CN109890573B (en) * 2019-01-04 2022-05-03 上海阿科伯特机器人有限公司 Control method and device for mobile robot, mobile robot and storage medium
CN111338470A (en) * 2020-02-10 2020-06-26 烟台持久钟表有限公司 Method for controlling big clock through gestures
CN111338470B (en) * 2020-02-10 2022-10-21 烟台持久钟表有限公司 Method for controlling big clock through gestures
CN113703581A (en) * 2021-09-03 2021-11-26 广州朗国电子科技股份有限公司 Window adjusting method based on gesture switching, electronic whiteboard and storage medium

Similar Documents

Publication Publication Date Title
CN107480600A (en) A kind of gesture identification method based on depth convolutional neural networks
CN106529468B (en) A kind of finger vein identification method and system based on convolutional neural networks
CN100492400C (en) Matching identification method by extracting characters of vein from finger
CN100395770C (en) Hand-characteristic mix-together identifying method based on characteristic relation measure
CN101777117B (en) Method for extracting finger vein feature for matching identification
CN108460318B (en) Identity authentication/identification method and equipment based on electrocardiosignals
CN104463100A (en) Intelligent wheelchair man-machine interaction system and method based on facial expression recognition mode
CN101281598A (en) Method for recognizing human face based on amalgamation of multicomponent and multiple characteristics
CN104346607A (en) Face recognition method based on convolutional neural network
CN104239769A (en) Identity recognition method and system based on finger vein characteristics
CN106529504B (en) A kind of bimodal video feeling recognition methods of compound space-time characteristic
CN102542281A (en) Non-contact biometric feature identification method and system
CN107392114A (en) A kind of finger vein identification method and system based on neural network model
CN101789075A (en) Finger vein identifying method based on characteristic value normalization and bidirectional weighting
CN110555380A (en) Finger vein identification method based on Center Loss function
Podder et al. An efficient iris segmentation model based on eyelids and eyelashes detection in iris recognition system
CN106778498A (en) A kind of method for strengthening Fingerprint recognition
Wang et al. Damaged fingerprint classification by Deep Learning with fuzzy feature points
Yeh et al. ResNeXt with cutout for finger vein analysis
Chen et al. A finger vein recognition algorithm based on deep learning
CN107977935A (en) A kind of vena metacarpea image texture filtering reinforcement method
CN103886291B (en) Palmprint image wrinkle detection method based on dynamic template
Gaxiola et al. Modular neural networks for person recognition using segmentation and the iris biometric measurement with image pre-processing
Thiyaneswaran et al. Iris Recognition using Left and Right Iris Feature of the Human Eye for Biometric Security System
CN114973307A (en) Finger vein identification method and system for generating countermeasure and cosine ternary loss function

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20171215