CN111913575B - Method for recognizing hand-language words - Google Patents

Method for recognizing hand-language words Download PDF

Info

Publication number
CN111913575B
CN111913575B CN202010721233.6A CN202010721233A CN111913575B CN 111913575 B CN111913575 B CN 111913575B CN 202010721233 A CN202010721233 A CN 202010721233A CN 111913575 B CN111913575 B CN 111913575B
Authority
CN
China
Prior art keywords
word
sign language
words
hand
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010721233.6A
Other languages
Chinese (zh)
Other versions
CN111913575A (en
Inventor
王青山
郑志文
朱钰
王�琦
张江涛
胡汇源
丁景宏
马晓迪
王鑫炎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN202010721233.6A priority Critical patent/CN111913575B/en
Publication of CN111913575A publication Critical patent/CN111913575A/en
Application granted granted Critical
Publication of CN111913575B publication Critical patent/CN111913575B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • G06F2218/04Denoising
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Abstract

The invention discloses a method for recognizing sign language words, which comprises collecting signals generated when a subject performs sign language gesture actions by utilizing a bracelet of a sensor, firstly, removing high-frequency noise generated by equipment reasons from the acquired sign language gesture signals by using a Butterworth filter, in the feature extraction stage, the invention uses a new feature DTW to mark the signal changes and fluctuations during sign language action, meanwhile, the invention carries out word embedding operation on a predetermined number of the middle general sign language words to obtain high-dimensional sign language word vectors, improves the t-SNE algorithm, combines a sign language corpus, reducing dimensions of each high-dimensional hand word vector to obtain semantic features of the hand words, designing and identifying each hand word based on an LSTM neural network, therefore, the problem that the existing sign language gesture recognition cannot recognize the sign language under the condition of large scale is solved.

Description

Method for recognizing hand-language words
Technical Field
The invention relates to the technical field of intelligent recognition, in particular to a method for recognizing hand-written words.
Background
In recent years, with the development of 5G technology and the popularization of mobile sensor devices, attention is paid to the use of wearable sensor devices to assist deaf-mute communication. The wearable sensor is used for identifying the sign language of the deaf-mute, and essentially belongs to human body action identification. The existing research on motion recognition is mainly divided into three categories: video-based, radio frequency-based, and sensor-based.
The video-based method depends on action equipment such as a camera, and the method often causes serious privacy problems, and has the problem of strong invasiveness due to the limitation of light and environment. In the action recognition based on the radio frequency signal, the existing WiFi device is often used to collect Channel State Information (CSI) signal changes caused by actions of people to presume actions performed by the human body, and the method needs to improve recognition accuracy, often only can recognize simple actions with large amplitude, and has a problem of being greatly influenced by the environment.
Disclosure of Invention
The invention provides a method for recognizing hand-language words, and aims to solve the problem that the existing hand-language gesture recognition cannot recognize hand languages under the condition of large scale.
The application provides a method for recognizing a hand word, which comprises the following steps:
enabling the arms of a plurality of testees to wear sensor bracelets, respectively making sign language gestures of a preset number of sign language words, and acquiring and recording sensor signals corresponding to each sign language gesture through the sensor bracelets;
filtering the sensor signal;
extracting a characteristic value of each sensor signal by adopting a dynamic time normalization algorithm;
performing word embedding operation on semantic information of each hand word to generate a word vector corresponding to the hand word;
performing dimension reduction processing on the word vector to generate a sign language word meaning characteristic vector corresponding to the word vector;
inputting the characteristic values and the semantic feature vectors of the hand words into a preset long-short term memory neural network, and training through the preset long-short term memory neural network to obtain a classification recognition model;
and recognizing the sign language gestures through the classification recognition model, and further recognizing the sign language words with a preset number.
The embodiment of the invention collects the signals generated when the subject performs sign language gesture actions by utilizing the bracelet with the IMU signal sensor and the sEMG signal sensor, the bracelet is worn on the dominant hand of the subject, firstly, removing high-frequency noise generated by equipment reasons from the acquired sign language gesture signals by using a Butterworth filter, in the feature extraction stage, the invention uses a new feature DTW to mark the signal changes and fluctuations during sign language action, meanwhile, the invention carries out word embedding operation on a predetermined number of general sign language words to obtain high-dimensional sign language word vectors, improves the t-SNE algorithm, combines a sign language corpus, reducing dimensions of each high-dimensional hand word vector to obtain semantic features of the hand words, designing and identifying each hand word based on an LSTM neural network, therefore, the problem that the existing sign language gesture recognition cannot recognize the sign language under the condition of large scale is solved.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic flowchart of a method for recognizing a handwritten word according to an embodiment of the present invention;
FIG. 2 is a block diagram schematically illustrating a method for recognizing a handwritten word according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a continuous bag-of-words model of a recognition method for sign language words according to an embodiment of the present invention;
FIG. 4 is a LSTM model structure diagram of a method for recognizing a handwritten word according to an embodiment of the present invention;
FIG. 5 is a diagram illustrating an initial state of a gesture action of a subject performing a "hello" handwritten word according to an embodiment of the present invention;
FIG. 6 is a first state diagram of a gesture action of a "hello" handwritten word by a subject according to the method for recognizing the handwritten word provided by the embodiment of the present invention;
FIG. 7 is a second state diagram of a gesture action of a "hello" handwritten word by a subject according to the method for recognizing handwritten words of the present invention;
FIG. 8 is a diagram illustrating the variation of the three types of indicators, i.e., accuracy, recall and F1-score, with the number of the handwritten words according to the method for recognizing the handwritten words in the embodiment of the present invention;
fig. 9 is a schematic diagram illustrating a change in recognition accuracy of different individuals according to the method for recognizing a handwritten word according to the embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It will be understood that the terms "comprises" and/or "comprising," when used in this specification and the appended claims, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
It is also to be understood that the terminology used in the description of the invention herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used in the specification of the present invention and the appended claims, the singular forms "a," "an," and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise.
It should be further understood that the term "and/or" as used in this specification and the appended claims refers to and includes any and all possible combinations of one or more of the associated listed items.
Referring to fig. 1, fig. 1 is a schematic flow chart of a method for recognizing a handwritten word according to an embodiment of the present application. As shown in fig. 1, the method for recognizing a handwritten word includes steps S101 to S104.
Step S101: let a plurality of examinee's arms wear the sensor bracelet to make the sign language gesture of predetermined number hand word respectively, pass through the sensor bracelet gathers and takes notes every the sensor signal that the sign language gesture corresponds.
Specifically, as shown in fig. 2, the user performs sign language gesture motions, and the IMU signal and the sEMG signal generated by the sensor bracelet are transmitted to the computer host through bluetooth.
Step S102: and filtering the sensor signal.
Specifically, denoising the sign language gesture signal acquired in the step (1) to remove background noise existing in the signal and generated due to equipment reasons. The normalized Butterworth filter is defined in the frequency domain as follows:
Figure BDA0002600079070000041
wherein n represents the stage number of the Butterworth filter, j is an imaginary unit, i.e., j is-1, and ω is the cut-off angle frequency of the signal, and radian/second is the unit, in the present invention, 200 hand words are taken as an example, and the bracelet is FsData were acquired at a rate of 200 samples/second, and during the gesture, the finger motion frequency was f 15Hz, thus setting the cut-off frequency ω of the butterworth filter to be
Figure BDA0002600079070000042
Wherein, the time of the subject for performing a sign language gesture is 2 seconds, and 200 pieces of data are collected. Wherein the signals are 18-dimensional in total and represent 8-dimensional sEMG signals and 10-dimensional IMU signals, respectively.
Step S103: and extracting the characteristic value of each sensor signal by adopting a dynamic time integral algorithm.
Specifically, by using the DTW to define the change of each line of data, since the subject has a corresponding change in each dimension of signal during the gesture motion, we can use the DTW algorithm to find the points on two adjacent lines of data to obtain their distances relatively more accurately based on the first line of data of the gesture word motion signal.
Next, describing a DTW calculation method, assuming that two adjacent rows of data sequences are a and B, respectively, and their lengths are m and n, respectively, and specifically expressed as follows:
A=a1,a2,a3,···,ai,···,an,
B=b1,b2,b3,···,bj,···,bm,
wherein, aiAmplitude of i-th frame representing sequence A, bjRepresenting the amplitude of the jth frame of sequence B.
Definition aiAnd bjThe DTW values in between are as follows:
DTW(i,j)=d(ai,bj)+min{DTW(i-1,j-1),DTW(i-1,j),(i,j-1)},
wherein d (a)i,bj) Presentation data aiAnd bjThe euclidean distance of (c). The two sequences are matched from the initial end (0,0) point of the data, and the similarity of the two adjacent lines of data can be obtained after the two sequences reach the end point (n, m), namely the DTW (n, m) obtained by the invention.
Then, the invention extracts the common data characteristic value: maximum, minimum, standard deviation, skewness and kurtosis, which are derived from each line of data for each spoken word. The feature values including the DTW extracted above are all the motion features of the hand words.
Step S104: and performing word embedding operation on the semantic information of each sign language word to generate a word vector corresponding to the sign language word.
Specifically, taking 200 handlanguage words as an example, the present invention defines the vector representation of the 200 common handlanguage words (which may be, for example, chinese common handlanguage words) through a word embedding operation. Specifically, the method uses the sign language corpus to perform word embedding operation on the 200 sign language words by using a continuous bag-Of-words model (CBOW). The CBOW model may use existing context to predict the current word. As shown in fig. 3, a schematic diagram of the CBOW model is shown, assuming that the context is "Todayweather", the current word is "good", and Σ g (templates) represents the sum of the words "Today" and "weather". The model is essentially a two-class neural network, assuming that the context is h and the real target vocabulary corresponding to the context is wtThe noise vocabulary is
Figure BDA0002600079070000051
Noise sequence is Pnoise. Thus, the optimization function of the neural network is:
Figure BDA0002600079070000052
wherein, when D is 1, wtlnQ for real target vocabularyθ(D=1|wtH) is a vector word wtThe probability obtained by Logistic regression corresponding to h, using a binary loss function to train the model, can obtain the value of the hidden layer as the embedded expression of the corresponding word as follows:
V={vi|1≤i≤200},
wherein V represents the word vector group of the 200 hand-written words, ViA word vector representing each hand word, which is represented as follows:
vi={vi1,vi2,...,vi200}.
wherein, the ith hand word viRepresented as a 200-dimensional vector.
Step S105: and performing dimension reduction processing on the word vector to generate a sign language word meaning characteristic vector corresponding to the word vector.
Specifically, a t-distributed Stochastic Neighbor Embedding (t-SNE) algorithm based on semantics is designed, and dimension reduction is performed on the high-dimensional word vectors obtained in step S104 to obtain sign language word meaning feature representation. Firstly, high-dimensional word vectors are arranged from high to low according to the frequency in the daily use process of sign language, and the accuracy of dimension reduction is improved by giving weight to each symbol. The invention gives each word vector viAdding a weight
Figure BDA0002600079070000053
The definition is as follows:
Figure BDA0002600079070000061
wherein p isiIs the frequency of occurrence of sign language word vectors in the corpus. Defining a word vector viAs vjConditional probability p of neighborsj|iThe following were used:
Figure BDA0002600079070000062
wherein the content of the first and second substances,
Figure BDA0002600079070000063
represents by viA square with center obeying Gaussian distribution, | | · | | non-woven phosphor2Is a two-norm, high-dimensional word vector viAnd vjThe corresponding low-dimensional word vectors are respectively
Figure BDA0002600079070000064
And
Figure BDA0002600079070000065
and the corresponding low-dimensional space obeys t-distribution. In the same way, with
Figure BDA0002600079070000066
As
Figure BDA0002600079070000067
The conditional probability of the neighbors of (a) is:
Figure BDA0002600079070000068
analogously, define
Figure BDA0002600079070000069
And
Figure BDA00026000790700000610
the joint probability distribution of (c) is:
Figure BDA00026000790700000611
the invention uses KL (Kullback-Leibler, divergence) distance as our optimization target at this time:
Figure BDA00026000790700000612
at this point, the optimal gradient is:
Figure BDA00026000790700000613
then training the model by using a stochastic gradient descent algorithm, and carrying out t-SNE reduction on the word vectors of 200 common symbol words to two dimensions. The reduced symbolic words form a semantic space in which each point represents a symbolic word.
Step S106: and inputting the characteristic values and the sign language word meaning characteristic vectors into a preset long-short term memory neural network, and training through the preset long-short term memory neural network to obtain a classification recognition model.
Specifically, the LSTM recurrent neural network is adopted, the action signal characteristics of each hand word extracted in the step and the semantic characteristics of each hand word are input into the LSTM recurrent neural network, and the LSTM recurrent neural network is used for training to obtain a classification recognition model. The invention uses LSTM neural network to train the hand word feature, and the input vector is the action feature and semantic feature of the hand word. For example, 70% of the feature data may be used as training data and 30% of the feature data may be used as test data. Performing label setting before training, wherein the performing label setting before training comprises: before training, input training data is labeled, namely the sequence number of each sign language gesture feature and sign language semantic feature data is set according to sign language words corresponding to the training data, then the model is trained according to the training times set by the method, and finally the trained model is obtained. Then the test data of the invention is sent to a model for detection to obtain the probability of a predicted value. The LSTM model is shown in fig. 4.
Step S107: and recognizing the sign language gestures through the classification recognition model, and further recognizing the sign language words with a preset number.
Specifically, the hand ring provided with an IMU signal sensor and an sEMG signal sensor is used for collecting signals generated when a subject performs sign language gesture actions, the hand ring is worn on the main hand of the subject, the collected sign language gesture signals are firstly removed by a Butterworth filter to remove high-frequency noise generated due to equipment, in the feature extraction stage, the invention uses a new feature DTW to mark the signal changes and fluctuations during sign language action, meanwhile, the invention carries out word embedding operation on a predetermined number of general sign language words to obtain high-dimensional sign language word vectors, improves the t-SNE algorithm, combines a sign language corpus, reducing dimensions of each high-dimensional hand word vector to obtain semantic features of the hand words, designing and identifying each hand word based on an LSTM neural network, therefore, the problem that the existing sign language gesture recognition cannot recognize the sign language under the condition of large scale is solved.
In the experiment, we had 10 subjects (5 men and 5 women) with an age between 18 and 25 years. With sign language training 8 hours before the experiment, they can accurately execute 200 sign language words in the whole sign language corpus. During the experiment, the subject worn the bracelet on his dominant hand (i.e., right or left hand) and made a gesture in front of the PC. As shown in fig. 5-7, the subject is performing a gesture action of "hello" hand words, wherein fig. 5 is the subject in an initial state, fig. 6 is the subject in a first state of the hand words, and fig. 7 is the subject in a second state of the hand words.
According to the above, the experiment sends the motion characteristics of the hand words and the semantic characteristics of the hand words to the LSTM neural network for training. In the experiment, the recognition effect of the model is judged by using three indexes of accuracy, recall rate and F1-score, and the experimental result is shown in FIG. 8. The experimental result shows that under the condition of various sign language word quantities, the three indexes are maintained to be more than 95%.
FIG. 9 shows the change in recognition accuracy of the model for different individuals. The horizontal axis represents 10 different subjects, and the vertical axis represents recognition accuracy. The accuracy obtained is different due to different habits of each person in sign language. The experimental result shows that the identification accuracy rate of 10 subjects reaches more than 95%.
While the invention has been described with reference to specific embodiments, the scope of the invention is not limited thereto, and those skilled in the art can easily conceive various equivalent modifications or substitutions within the technical scope of the invention. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (8)

1. A method for recognizing a hand-written word, comprising:
enabling the arms of a plurality of testees to wear sensor bracelets, respectively making sign language gestures of a preset number of sign language words, and acquiring and recording sensor signals corresponding to each sign language gesture through the sensor bracelets;
filtering the sensor signal;
extracting a characteristic value of each sensor signal by adopting a dynamic time normalization algorithm;
performing word embedding operation on semantic information of each hand word to generate a word vector corresponding to the hand word;
performing dimension reduction processing on the word vector to generate a sign language word meaning characteristic vector corresponding to the word vector;
inputting the characteristic values and the semantic feature vectors of the hand words into a preset long-short term memory neural network, and training through the preset long-short term memory neural network to obtain a classification recognition model;
and recognizing the sign language gestures through the classification recognition model, and further recognizing the sign language words with a preset number.
2. The method for recognizing hand-language words according to claim 1, wherein the acquiring and recording the sensor signal corresponding to each hand-language gesture through the sensor bracelet comprises:
and acquiring IMU signals and sEMG signals generated corresponding to each sign language gesture through the sensor bracelet, and transmitting the IMU signals and the sEMG signals to a processor for recording and storing.
3. The method for recognizing sign language words according to claim 2, wherein said filtering the sensor signal comprises filtering the sensor signal through a normalized Butterworth filter, the expression of which is:
Figure FDA0003022344320000011
where n represents the number of stages of the butterworth filter, j is the unit of an imaginary number, and ω is the cut-off angular frequency of the signal.
4. The method according to claim 3, wherein the feature values include DTW feature values and common feature values, and the common feature values include maximum values, minimum values, standard deviations, skewness, and kurtosis of the sensor signal data.
5. The method according to claim 4, wherein said performing word embedding operations on the semantics of each of said handwritten word comprises performing word embedding operations on said predetermined number of said handwritten word through a continuous bag-of-words model,
the optimization function of the continuous bag-of-words model is as follows:
Figure FDA0003022344320000021
wherein h is a context, wtFor the real target vocabulary corresponding to the context,
Figure FDA0003022344320000022
is a noisy vocabulary, PnoiseFor a noise sequence, θ represents a weight parameter of the neural network, and when D is 1, wtlnQ for real target vocabularyθ(D=1|wtH) is a vector word wtThe probability obtained by Logistic regression corresponding to h is used for training the model by utilizing a binary loss function, and the value of the hidden layer can be obtained and used as the word embedding expression of the corresponding hand word as follows:
V={vi/1≤i≤z}
wherein z represents the number of the hand words, V represents the word vector group of the z hand words, ViA word vector representing each hand word, which is represented as follows:
vi={vi1,vi2,...,viz}
wherein, the ith hand word viRepresented as a vector in the z dimension.
6. The method according to claim 5, wherein said performing dimension reduction on said word vector to generate a sign language word sense feature vector corresponding to said word vector comprises:
arranging the word vectors from high to low according to the frequency in the daily use process of the sign language, giving a weight to each symbol to improve the accuracy of dimension reduction, and adding a weight to each word vector, wherein the weights are defined as follows:
Figure FDA0003022344320000023
wherein p isiIs the frequency of appearance of sign language word vectors in the corpus;
defining a word vector viAs vjConditional probability p of neighborsj|iThe following were used:
Figure FDA0003022344320000024
wherein the content of the first and second substances,
Figure FDA0003022344320000025
represents by viA square with center obeying Gaussian distribution, | | · | | non-woven phosphor2Is a two-norm, high-dimensional word vector viAnd vjThe corresponding low-dimensional word vectors are respectively
Figure FDA0003022344320000026
And
Figure FDA0003022344320000027
the corresponding low-dimensional space obeys t distribution;
to be provided with
Figure FDA0003022344320000028
As
Figure FDA0003022344320000029
The conditional probability of the neighbors of (a) is:
Figure FDA00030223443200000210
definition of
Figure FDA00030223443200000211
And
Figure FDA00030223443200000212
the joint probability distribution of (c) is:
Figure FDA00030223443200000213
wherein the content of the first and second substances,
Figure FDA00030223443200000214
and
Figure FDA00030223443200000215
respectively representing the k-th and l-th low-dimensional word vectors, v, in the sign language lexiconk、vlRespectively representing the corresponding high-dimensional word vectors;
Kullback-Leibler divergence distance was used as the optimization target:
Figure FDA0003022344320000031
at this time, the optimal gradient is expressed as:
Figure FDA0003022344320000032
wherein, yiRepresenting a hand word viThe gradient of (a) of (b) is,
Figure FDA0003022344320000033
and
Figure FDA0003022344320000034
respectively represent the gradient yiAnd gradientyjAnd training the model by utilizing a random gradient descent algorithm according to the corresponding weight coefficient, reducing the word vector of z commonly used symbolic words to two dimensions by carrying out t-distribution random adjacent embedding operation, and forming a semantic space by the dimensionality reduced symbolic words, wherein each point represents a symbolic word.
7. The method according to claim 6, wherein the inputting the feature values of the sensor signals and the feature vectors of the meaning of the sign language words into a preset long-short term memory neural network, and the training through the preset long-short term memory neural network to obtain the classification recognition model comprises:
and training the characteristic values of the sensor signals and the sign language word meaning characteristic vectors by using the preset long-term and short-term memory neural network, using preset percentages of characteristic data as training data, using the rest characteristic data as test data, setting labels before training, training the classification recognition model according to preset training times, and finally obtaining the trained model.
8. The method according to claim 7, wherein the setting of labels before training comprises: and marking the input training data before training, namely setting the sequence number of each sign language gesture feature and sign language semantic feature data according to the sign language word corresponding to the training data.
CN202010721233.6A 2020-07-24 2020-07-24 Method for recognizing hand-language words Active CN111913575B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010721233.6A CN111913575B (en) 2020-07-24 2020-07-24 Method for recognizing hand-language words

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010721233.6A CN111913575B (en) 2020-07-24 2020-07-24 Method for recognizing hand-language words

Publications (2)

Publication Number Publication Date
CN111913575A CN111913575A (en) 2020-11-10
CN111913575B true CN111913575B (en) 2021-06-11

Family

ID=73281416

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010721233.6A Active CN111913575B (en) 2020-07-24 2020-07-24 Method for recognizing hand-language words

Country Status (1)

Country Link
CN (1) CN111913575B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114115531B (en) * 2021-11-11 2022-09-30 合肥工业大学 End-to-end sign language recognition method based on attention mechanism
CN116386149B (en) * 2023-06-05 2023-08-22 果不其然无障碍科技(苏州)有限公司 Sign language information processing method and system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956529A (en) * 2016-04-25 2016-09-21 福州大学 Chinese sign language identification method based on LSTM type RNN
JP2018124801A (en) * 2017-02-01 2018-08-09 株式会社エクスビジョン Gesture recognition device and gesture recognition program
CN109063615A (en) * 2018-07-20 2018-12-21 中国科学技术大学 A kind of sign Language Recognition Method and system
CN109754113A (en) * 2018-11-29 2019-05-14 南京邮电大学 Load forecasting method based on dynamic time warping Yu length time memory
KR20190098806A (en) * 2018-01-31 2019-08-23 계명대학교 산학협력단 A smart hand device for gesture recognition and control method thereof
CN110414468A (en) * 2019-08-05 2019-11-05 合肥工业大学 Based on the auth method of hand signal under WiFi environment

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180260531A1 (en) * 2017-03-10 2018-09-13 Microsoft Technology Licensing, Llc Training random decision trees for sensor data processing
CN108766434B (en) * 2018-05-11 2022-01-04 东北大学 Sign language recognition and translation system and method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105956529A (en) * 2016-04-25 2016-09-21 福州大学 Chinese sign language identification method based on LSTM type RNN
JP2018124801A (en) * 2017-02-01 2018-08-09 株式会社エクスビジョン Gesture recognition device and gesture recognition program
KR20190098806A (en) * 2018-01-31 2019-08-23 계명대학교 산학협력단 A smart hand device for gesture recognition and control method thereof
CN109063615A (en) * 2018-07-20 2018-12-21 中国科学技术大学 A kind of sign Language Recognition Method and system
CN109754113A (en) * 2018-11-29 2019-05-14 南京邮电大学 Load forecasting method based on dynamic time warping Yu length time memory
CN110414468A (en) * 2019-08-05 2019-11-05 合肥工业大学 Based on the auth method of hand signal under WiFi environment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于传感器的手势识别方案研究及其在可穿戴设备中的实现;唐继玉;《中国优秀硕士学位论文全文数据库》;20161231;1-61 *

Also Published As

Publication number Publication date
CN111913575A (en) 2020-11-10

Similar Documents

Publication Publication Date Title
CN108899050B (en) Voice signal analysis subsystem based on multi-modal emotion recognition system
Alzohairi et al. Image based Arabic sign language recognition system
Karnati et al. LieNet: a deep convolution neural network framework for detecting deception
Yanay et al. Air-writing recognition using smart-bands
CN109886068B (en) Motion data-based action behavior identification method
CN111913575B (en) Method for recognizing hand-language words
CN104484644A (en) Gesture identification method and device
Shin et al. Korean sign language recognition using EMG and IMU sensors based on group-dependent NN models
Sharma et al. Trbaggboost: An ensemble-based transfer learning method applied to Indian Sign Language recognition
Xu et al. Intelligent emotion detection method based on deep learning in medical and health data
CN109902554A (en) A kind of recognition methods of the sign language based on commercial Wi-Fi
Futane et al. Video gestures identification and recognition using Fourier descriptor and general fuzzy minmax neural network for subset of Indian sign language
CN114384999B (en) User-independent myoelectric gesture recognition system based on self-adaptive learning
Zhang et al. Hand gesture recognition with SURF-BOF based on Gray threshold segmentation
CN111914724B (en) Continuous Chinese sign language identification method and system based on sliding window segmentation
Aksoy et al. Detection of Turkish sign language using deep learning and image processing methods
CN109766559B (en) Sign language recognition translation system and recognition method thereof
Singh et al. A reliable and efficient machine learning pipeline for american sign language gesture recognition using EMG sensors
Surekha et al. Hand Gesture Recognition and voice, text conversion using
Wang et al. Handwriting recognition under natural writing habits based on a low-cost inertial sensor
Li et al. Cross-people mobile-phone based airwriting character recognition
Viswanathan et al. Recent developments in Indian sign language recognition: an analysis
Robert et al. A review on computational methods based automated sign language recognition system for hearing and speech impaired community
CN114764580A (en) Real-time human body gesture recognition method based on no-wearing equipment
CN114863572A (en) Myoelectric gesture recognition method of multi-channel heterogeneous sensor

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant