CN113111831A - Gesture recognition technology based on multi-mode information fusion - Google Patents
Gesture recognition technology based on multi-mode information fusion Download PDFInfo
- Publication number
- CN113111831A CN113111831A CN202110441452.3A CN202110441452A CN113111831A CN 113111831 A CN113111831 A CN 113111831A CN 202110441452 A CN202110441452 A CN 202110441452A CN 113111831 A CN113111831 A CN 113111831A
- Authority
- CN
- China
- Prior art keywords
- signals
- gesture
- signal
- neural network
- gesture recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/107—Static hand or arm
- G06V40/113—Recognition of static hand signs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2135—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/12—Computing arrangements based on biological models using genetic models
- G06N3/126—Evolutionary algorithms, e.g. genetic algorithms or genetic programming
Abstract
A gesture recognition technology based on multi-mode information fusion belongs to the technical field of gesture recognition. It comprises the following steps: the method comprises the following steps: collecting a gesture action acquisition signal executed by a subject; step two: denoising myoelectric signals and triaxial acceleration and angle original data, extracting signal moving segments and marking labels; step three: extracting time domain characteristics and AR model parameters of the electromyographic signals, and extracting absolute average values and waveform length characteristics of the triaxial acceleration signals and the angle signals; step four: fusing and reducing the dimensions of the extracted features, and inputting the data subjected to dimension reduction into a classifier; step five: training to obtain an optimized prediction model; step six: and inputting the acquired gesture signals needing to be predicted into a back propagation neural network optimized based on a genetic algorithm for recognition, and outputting a result. The invention extracts information from multiple angles, overcomes the defect of incomplete extraction of single electromyographic signal information, increases the diversity of gesture information and is beneficial to improving the gesture recognition accuracy.
Description
Technical Field
The invention belongs to the technical field of gesture recognition, and particularly relates to a gesture recognition technology based on multi-mode information fusion.
Background
With the development of artificial intelligence, human-computer interaction technology as a core of the artificial intelligence development is receiving more and more attention. The essence of human-computer interaction is the information exchange process between a person and a computer for completing a certain task in a certain interaction mode. The hand is an important organ of the human body, not only has the function of working, but also is an important medium for information communication. The gesture is used as an effective information transmission means, has the characteristics of non-contact and diversity, is a common control means in the field of human-computer interaction, and the gesture recognition technology also becomes a hot topic in the fields of intelligent artificial limbs, rehabilitation training, sign language recognition, intelligent home furnishing and the like.
According to the current total population of China and the related data of the second national disabled people sampling survey, the proportion of the disabled population to the total population of the whole country is 6.34%, the total number of various disabled people is about 8876 thousands of people, wherein the proportion of the disabled people to the physical disabilities is the largest, and is 29.07%, and the number of people is about 2580 thousands. The intelligent artificial limb is arranged to help the disabled patients to relieve inconvenience and recover partial missing functions. The traditional gesture recognition method cannot be applied to the control of the artificial limb, and the gesture recognition based on the surface electromyogram signal can be used for meeting the requirements of the use and the control of the artificial limb.
At present, the population of China is aging seriously, so that more and more people face the trouble of apoplexy of middle-aged and old people, and the labor cost gradually rises. The cost of rehabilitation treatment for patients with limb function impairment will be higher and higher, so the way of using machines to replace manual work to help patients to perform rehabilitation training is a trend. The nervous system of a human has a recovery ability, so that the paralysis of limbs caused by a stroke can recover partial functions or even completely recover through rehabilitation training. Clinical practice shows that rehabilitation training matched with the active intention of a human body has a very positive effect on the rehabilitation of a nervous system. The surface muscle electrical signal can well reflect the active movement intention of the human body, so that the patient can carry out more effective rehabilitation training by matching with the gesture recognition technology of the myoelectric signal.
Sign language is a language of a structure language, is not limited by a vocal organ, is a main mode of daily communication of deaf-mutes, and normal people are not familiar with the expression mode of the sign language, which also causes communication barrier between the normal people and the deaf-mutes.
With the development of virtual reality and intelligent home technology, a user can realize the control of equipment through an information interaction means, and can experience more convenient and faster life style for the group with inconvenient movement, such as control of lamplight, household appliances, wheelchairs and even small aircrafts.
The hand motion recognition technology based on the surface electromyogram signal is researched, the hand motion recognition with rapidness, reliability and low cost is realized, the control and the use of the intelligent artificial limb by the disabled can be met in the aspect of medical rehabilitation, the rehabilitation training of the motor neuropathy patient is carried out based on the active will, and the more convenient and more universal information output in the man-machine interaction process can be met.
The Surface ElectroMyoGram (sEMG) is used as bioelectricity generated by human body activities, the acquisition of the bioelectricity has the advantages of non-invasiveness and non-wound, the psychology and the body of an acquired person are not burdened, but the single sensor collects information in a one-sided mode, the sEMG can effectively reflect the flexion and extension of fingers but cannot reflect the pose information of the hand, so that the number of gestures which can be covered by only identifying the gestures through the sEMG is small, and the actions of parts of the fingers which are similar to the actions are easy to identify errors.
Kernel Principal Component Analysis (KPCA) is used as a nonlinear generalization of Principal Component Analysis (PCA), and is implemented by mapping sample data from an input space to a high-dimensional feature space through nonlinear transformation, and then performing feature extraction in the high-dimensional feature space by using PCA. PCA is often used for dimensionality reduction of linear features, the advantage of which cannot be highlighted for nonlinear data, whereas sEMG signals are nonlinear. Therefore, the myoelectric signal and pose signal characteristics are fused and dimension reduced by using the KPCA method, and the original signal aspect is more reasonable. And the characteristics after dimension reduction and fusion are sent to a back propagation neural network based on genetic algorithm optimization, so that the identification process is accelerated.
Genetic Algorithm (GA), an evolutionary Algorithm designed according to the rules of evolution of organisms in nature, was first proposed by John holland in the united states in the 70 th 20 th century. The basic principle is an evolution algorithm simulating Darwinian biological evolution theory, namely 'competitive selection of a matter and survival of a suitable person'. The algorithm encodes problem parameters into chromosomes, and then performs operations such as selection, intersection, variation and the like in an iterative mode to exchange information of the chromosomes in a population, so as to finally generate the chromosomes meeting the optimization target. When a complex combined optimization problem is solved, a better optimization result can be obtained faster compared with some conventional optimization algorithms. Genetic algorithms have been widely used by people in the fields of combinatorial optimization, machine learning, signal processing, adaptive control, artificial life, and the like.
The bp (back propagation) neural network is a concept proposed by scientists including Rumelhart and McClelland in 1986, is a multi-layer feedforward neural network trained according to an error back propagation algorithm, and is the most widely applied neural network. The network structure of the device is composed of three neutral network structures including an input layer, a hidden layer and an output layer, and a single hidden layer can sufficiently meet the use requirement. The input layer receives data, the output layer outputs data, the neuron in the previous layer is connected to the neuron in the next layer, information transmitted by the neuron in the previous layer is collected, and a value is transmitted to the neuron in the next layer through 'activation'. Although the BP neural network is the most widely applied algorithm in the artificial neural network, the selection of the network structure, the initial connection weight and the threshold value of the BP neural network greatly improves the network training, but the BP neural network cannot be obtained accurately, and the genetic algorithm can improve the BP neural network.
Disclosure of Invention
Aiming at the problems in the prior art, the invention aims to provide a gesture recognition technology, which is a method for acquiring the myoelectric signals and the pose signals of the hands by a sensor to recognize hand actions.
The invention provides the following technical scheme: 1. a gesture recognition technology based on multi-modal information fusion is characterized in that: the method specifically comprises the following steps:
the method comprises the following steps: arranging a myoelectricity sensor at a proper position and placing a pose sensor at the back of a hand according to the muscle related to the gesture to be recognized, and executing the gesture action by a subject to acquire signals;
step two: denoising original data of electromyographic signals, three-axis acceleration signals of gesture actions and angle signals by a filtering method, extracting signal active segments by a dual-threshold segmentation method, and marking labels;
step three: extracting time domain characteristics and AR model parameters of the electromyographic signals, and extracting absolute average values and waveform length characteristics of the triaxial acceleration signals and the angle signals;
step four: performing fusion dimensionality reduction on the features extracted in the third step by a Kernel Principal Component Analysis (KPCA) method, and inputting dimensionality-reduced data into a classifier;
step five: optimizing an initial weight threshold of the BP neural network through a genetic algorithm, and training to obtain an optimized prediction model;
step six: and inputting the collected signals of the gesture motion needing to be predicted into a back propagation neural network GA-BPNN optimized based on a genetic algorithm through the characteristics after dimension reduction and fusion for identification, and outputting the result.
The gesture recognition technology based on multi-modal information fusion is characterized in that in the second step, a low-pass filtering mode is adopted, signals passing through a Butterworth low-pass filter with the cut-off frequency of 5Hz are subtracted from original signals, the base line drift of electromyographic signals is removed, low-frequency motion artifact interference and high-frequency noise interference are removed through a Butterworth band-pass filter with the pass band of 20Hz-460Hz, a comb-shaped wave trap is used for removing 50Hz power frequency and harmonic interference thereof, and noise interference mixed in the process of collecting acceleration signals and angle signals is removed through moving average filtering.
The gesture recognition technology based on multi-modal information fusion is characterized in that in the second step, data are divided through overlapped sliding windows, a first threshold value sigma is set by extracting an energy average value of multi-channel electromyographic signals, the size of the first threshold value sigma is 1.3 times of the average value of the multi-channel electromyographic signals in a relaxed state, a second threshold value mu is set to be 6, a calculated value in the window is compared with the first threshold value sigma, when the calculated value is larger than the first threshold value sigma and the length of a continuous window is larger than the second threshold value mu, the signal is judged to be an active segment, and the active segment of the pose signal is divided according to the active segment time of the electromyographic signals.
The gesture recognition technology based on multi-modal information fusion is characterized in that in the fourth step, a kernel function in the kernel principal component analysis method uses a Gaussian radial basis kernel function, and the first n principal components with contribution rate of more than 90% are selected as data after dimensionality reduction.
The gesture recognition technology based on multi-modal information fusion is characterized in that in the fifth step, the characteristics obtained after the test data are processed in the first step to the fourth step are input into a BP neural network, the absolute value of the error of a prediction result is used as the fitness of a genetic algorithm, the individual length is determined by the BP neural network structure, if the number of nodes of an input layer, a hidden layer and an output layer of the BP neural network is l, m and n respectively, the individual length is (l + n +1) x m + n, the optimal initial weight threshold obtained by optimizing the genetic algorithm is used as the initial weight threshold of the BP neural network, and a network model obtained after training is used for prediction.
The gesture recognition technology based on multi-modal information fusion is characterized in that in the sixth step, signals to be predicted are processed in the first step to the fourth step, characteristics are input into a trained neural network, and the class with the highest probability in results is output as a recognized gesture class.
By adopting the technology, compared with the prior art, the invention has the following beneficial effects:
1) the invention adopts a mode of extracting electromyographic signals and pose signals as original data, extracts information from multiple angles, complements the defect of incomplete extraction of single electromyographic signal information, increases the diversity of gesture information, and is beneficial to the improvement of gesture recognition accuracy;
2) the method adopts a method of electromyographic signal average energy double thresholds to segment the data active segment, limits the signal amplitude and the duration, reduces the interference of short-time noise interference on the segmentation of the active segment, improves the accuracy of information effective segment acquisition, and is beneficial to the time synchronization of multi-mode signals by corresponding acceleration signals and angle signals to the time segment of electromyographic signal segmentation;
3) the invention adopts a kernel principal component analysis method to perform fusion dimensionality reduction on the features extracted from the multi-modal information, is closer to the characteristic of feature nonlinearity of the electromyographic signal, is beneficial to retaining the nonlinearity characteristic of the signal, improves the retention degree of the information contained in the original signal while performing feature fusion dimensionality reduction, and contributes to improving the gesture recognition efficiency and reducing the recognition time by reducing the feature dimensionality of an input classifier;
4) the initial weight threshold of the BP neural network is optimized by adopting a genetic algorithm, so that the problem that the initial weight threshold with larger influence on the network performance can only be determined randomly is solved, the training time is reduced, and the identification accuracy is improved.
Drawings
FIG. 1 is a flow chart of a gesture recognition technique based on modal information fusion in accordance with the present invention;
FIG. 2 is a flowchart of the training algorithm of the GA-optimized BP neural network model according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
On the contrary, the invention is intended to cover alternatives, modifications, equivalents and alternatives which may be included within the spirit and scope of the invention as defined by the appended claims. Furthermore, in the following detailed description of the present invention, certain specific details are set forth in order to provide a better understanding of the present invention. It will be apparent to one skilled in the art that the present invention may be practiced without these specific details.
Referring to fig. 1-2, a gesture recognition technology based on multimodal information fusion specifically includes the following steps:
the method comprises the following steps: as shown in fig. 1, signals from both electromyogram signals and position signals are selected as gesture recognition samples. Arranging myoelectric sensors at arm muscles related to gestures of a subject, arranging IMU sensors at the back of hands, executing the action gestures of the subject according to a cyclic mode of 'relaxation-action-relaxation', wherein the action and relaxation time lasts for about 1s, each group of actions is performed for 30s, a rest interval is not less than 2min before the next group of actions is performed, and each group of actions is repeated for 3 times.
Step two: as shown in fig. 1, the original surface electromyogram signal is easily interfered by physiological and environmental noises such as motion artifacts and power frequency noises, whereas the acceleration signal and the angle signal are easily interfered by high frequency environmental noises, and the signals need to be preprocessed to improve the signal-to-noise ratio of the signals. Firstly, removing baseline drift of a signal obtained by subtracting low-pass filtering from an original signal in a low-pass filtering mode; then, a Butterworth band-pass filter of 20Hz-460Hz is utilized to remove low-frequency motion artifact interference and high-frequency noise interference; and finally, removing power frequency noise of 50Hz and harmonic interference thereof by using a comb-shaped wave trap. And eliminating fluctuation of the acceleration signal and the angle signal due to noise interference by means of sliding window mean filtering.
Then, the signals are alive by the mean energy double-threshold method of the electromyographic signalsThe expression of the mean energy of the electromyographic signals isWherein t represents a certain time, L represents the window length, C represents the number of electromyographic signal channels, semga(i) Is the electromyogram signal value of the i moment under the a channel, wherein the window length is 50 ms. Extracting a numerical value by a window for extracting the average energy with a sliding distance of 10ms, setting a threshold value sigma to be 1.3 times of the average energy of the electromyographic signals in a relaxed state, setting a threshold value mu to be 6, comparing a calculated value in the window with the threshold value sigma, judging the signal of the section to be an active section when the calculated value is greater than the threshold value sigma and the length of a continuous window is greater than the threshold value mu, wherein the starting time point of the active section is the starting point of the window which initially meets the threshold value condition, the ending time point of the active section is the starting point of the first window which does not meet the threshold value condition in a subsequent window, and the active section of the pose signal is divided according.
Step three: the features are extracted according to the window length of 200ms and the window sliding length of 50ms, and three time domain features are extracted for the electromyographic signals, namely an absolute average value (MAV), a root mean square value (RMS) and a Variance (VAR).
The absolute average expression is:where N is the window length, xiIs the electromyographic signal at time i.
for an AR parameter model, p parameters are used for representing sEMG signals in the method, and the invention adopts an AR parameter model with 4-order extraction, wherein the expression is as follows:wherein u isiIs white noise residual, p is AR model order, anIs the nth coefficient in the AR model.
And aiming at the extraction of the acceleration signal and the angle signal, the window of the acceleration signal and the angle signal is the same as the electromyographic signal, and the absolute average value is the same as the above.
step four: and (5) performing fusion dimensionality reduction on the features extracted in the step three by using a KPCA (kernel principal component analysis) method. Because different feature scales of signal extraction are different greatly, normalization processing needs to be performed on feature values first, and the feature values are normalized into data x with a mean value of 0 and a variance of 10The formula is as follows:where x is the raw feature value data, μ is the mean of the data, and σ is the variance of the data. And mapping the normalized data from a data space by using a Gaussian radial basis kernel function to obtain a kernel matrix K, wherein the Gaussian radial basis kernel function formula is as follows:where σ is the nuclear parameter. Then, the kernel matrix is centralized, and the formula is as follows: kc=K-lNK-KlN+lNKlNWherein l isNIs an N × N matrix, each element is 1/N, and N is the number of samples. Then calculating the matrix KcTo obtain a corresponding eigenvector of lambda1,···,λnSince the eigenvalues determine the magnitude v of the variance1,···,νnAnd sorting the effective information in descending order of the characteristic values and adjusting the corresponding characteristic vectors if the characteristic values are larger and the effective information is more. Orthogonalizing and unitizing the feature vectors by a Schmidt orthogonalization method to obtain a1,···,an. Calculating the accumulated contribution rate r of the reordered eigenvalues1,···,rnTaking the part with the contribution rate more than 90%, the formula is as follows:selecting a1,···,atAnd the data is the data after the dimensionality reduction and fusion.
Step five: as shown in fig. two, firstly, determining a BP neural network structure, and determining the number of nodes of an input layer, a hidden layer and an output layer, where the number of nodes of the input layer is dimension t after dimension reduction and fusion in the step four, the number of nodes of the output layer is a gesture type n to be recognized, and the number of nodes of the hidden layer m may refer to one of the following 3 formulas:
m<t-1
m=log2t
wherein a is a constant between 0 and 10 and includes 0 and 10. After the network structure is determined, the total weight and threshold in the neural network can be calculated as follows: and (t + n +1) x m + n, and coding the data by taking the weight value and the threshold length as the population length. And then, obtaining an initialization weight threshold value according to the individual for constructing a BP neural network, and training the neural network by the preprocessed training data, wherein the data preprocessing process comprises the four steps. Predicting by using the trained neural network, and taking the absolute value of the error between the predicted output and the expected output as the individual fitness of the genetic algorithm, wherein the formula is as follows:
where n is the number of neural network output nodes, yiIs the expected output of the ith node of the neural network, and y' is the predicted output of the ith node of the neural network.
And then a selection operation, which essentially selects excellent individuals from the old population with a certain probability from the viewpoint of natural selection and inheritance for next generation propagation, because the present invention uses the sum of absolute values of errors as a fitness function, the more excellent individuals should have smaller errors. Method of selectionThere are many, the present invention uses roulette, each individual i having a probability of selection piThe formula is as follows:
wherein SiThe fitness of the individual i is better as the error requirement is smaller, and therefore the reciprocal is taken, and N is the number of population individuals.
Then, performing crossover operation to the selected individuals and selecting the k-th chromosome dnakAnd the l chromosome dnalPerforming a crossover operation at m bits, wherein the formula is as follows:
wherein r is a random number between [0,1 ].
The organism has variation in the genetic process, so the genetic algorithm has variation operation, and the jth gene dna of the ith individual is selectedijPerforming variation by the formula:
wherein dnamaxAnd dnaminAre respectively based on dnaijX and r are two [0,1]]G is the current iteration number, gmaxIs the maximum number of iterations.
And if the iteration times after the mutation is finished are smaller than the maximum iteration times, returning to the selection step to continue the selection, crossing and mutation until the specified iteration times are reached, taking the generation with the minimum error as the optimal initial weight and the threshold value, substituting the optimal initial weight and the threshold value into the BP neural network, and training the neural network to obtain the optimal BP network model.
Step six: and acquiring myoelectric signals and pose signals of the testee during gesture execution in real time, processing the myoelectric signals and the pose signals in the first step to the fourth step, sending the processed myoelectric signals and pose signals to the optimal BP neural network trained in the fifth step, and outputting the result, wherein the category with the highest possibility is used as a final recognition result.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents and improvements made within the spirit and principle of the present invention are intended to be included within the scope of the present invention.
Claims (6)
1. A gesture recognition technology based on multi-modal information fusion is characterized in that: the method specifically comprises the following steps:
the method comprises the following steps: arranging a myoelectricity sensor at a proper position and placing a pose sensor at the back of a hand according to the muscle related to the gesture to be recognized, and executing the gesture action by a subject to acquire signals;
step two: denoising original data of electromyographic signals, three-axis acceleration signals of gesture actions and angle signals by a filtering method, extracting signal active segments by a dual-threshold segmentation method, and marking labels;
step three: extracting time domain characteristics and AR model parameters of the electromyographic signals, and extracting absolute average values and waveform length characteristics of the triaxial acceleration signals and the angle signals;
step four: performing fusion dimensionality reduction on the features extracted in the third step by a Kernel Principal Component Analysis (KPCA) method, and inputting dimensionality-reduced data into a classifier;
step five: optimizing an initial weight threshold of the BP neural network through a genetic algorithm, and training to obtain an optimized prediction model;
step six: and inputting the collected signals of the gesture motion needing to be predicted into a back propagation neural network GA-BPNN optimized based on a genetic algorithm through the characteristics after dimension reduction and fusion for identification, and outputting the result.
2. The gesture recognition technology based on multimodal information fusion as claimed in claim 1, wherein in the second step, a low pass filtering mode is adopted to subtract the signal passing through the butterworth low pass filter with the cut-off frequency of 5Hz from the original signal to remove the baseline drift of the electromyographic signal, the butterworth band pass filter with the pass band of 20Hz-460Hz is adopted to remove the low-frequency motion artifact interference and the high-frequency noise interference, the comb-shaped trap is adopted to remove the 50Hz power frequency and the harmonic interference thereof, and the sliding average filtering is adopted to remove the noise interference mixed in the acquisition process of the acceleration signal and the angle signal.
3. The multi-modal information fusion-based gesture recognition technology of claim 1, wherein in the second step, the data is divided through overlapped sliding windows, a first threshold value σ is set by extracting the energy mean value of the multi-channel electromyographic signals, the size of the first threshold value σ is 1.3 times of the mean value of the multi-channel electromyographic signals in the relaxed state, a second threshold value μ is set to 6, the calculated value in the window is compared with the first threshold value σ, when the calculated value is greater than the first threshold value σ and the length of the continuous window is greater than the second threshold value μ, the signal is judged to be an active segment, and the active segment of the pose signal is divided according to the active segment time of the electromyographic signals.
4. The gesture recognition technology based on multi-modal information fusion as claimed in claim 1, wherein in the fourth step, the kernel function in the kernel principal component analysis method uses a gaussian radial basis kernel function, and the first n principal components with contribution ratios of 90% or more are selected as the data after dimensionality reduction.
5. The multi-modal information fusion-based gesture recognition technology of claim 1, wherein in the fifth step, the features obtained by processing the test data in the first to fourth steps are input into a BP neural network, the absolute value of the error of the prediction result is used as the fitness of a genetic algorithm, the individual length is determined by the BP neural network structure, if the number of nodes of an input layer, a hidden layer and an output layer of the BP neural network is l, m and n respectively, the individual length is (l + n +1) xm + n, the optimal initial weight threshold obtained by optimizing the genetic algorithm is used as the initial weight threshold of the BP neural network, and the trained network model is used for prediction.
6. The gesture recognition technology based on multi-modal information fusion as claimed in claim 1, wherein in the sixth step, the signal to be predicted is processed through the first to fourth steps, the features are input into the trained neural network, and the class with the highest probability in the result is output as the recognized gesture class.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110441452.3A CN113111831A (en) | 2021-04-23 | 2021-04-23 | Gesture recognition technology based on multi-mode information fusion |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110441452.3A CN113111831A (en) | 2021-04-23 | 2021-04-23 | Gesture recognition technology based on multi-mode information fusion |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113111831A true CN113111831A (en) | 2021-07-13 |
Family
ID=76719565
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110441452.3A Withdrawn CN113111831A (en) | 2021-04-23 | 2021-04-23 | Gesture recognition technology based on multi-mode information fusion |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113111831A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114098768A (en) * | 2021-11-25 | 2022-03-01 | 哈尔滨工业大学 | Cross-individual surface electromyographic signal gesture recognition method based on dynamic threshold and easy TL |
CN114626412A (en) * | 2022-02-28 | 2022-06-14 | 长沙融创智胜电子科技有限公司 | Multi-class target identification method and system for unattended sensor system |
CN115719514A (en) * | 2022-11-23 | 2023-02-28 | 南京理工大学 | Gesture recognition-oriented field self-adaptive method and system |
CN115778408A (en) * | 2022-12-10 | 2023-03-14 | 福州大学 | Step-by-step electromyographic signal activity segment detection method |
CN116725556A (en) * | 2023-07-12 | 2023-09-12 | 中国科学院苏州生物医学工程技术研究所 | Motion intention recognition method and device based on surface electromyographic signals |
CN117238037A (en) * | 2023-11-13 | 2023-12-15 | 中国科学技术大学 | Dynamic action recognition method, device, equipment and storage medium |
-
2021
- 2021-04-23 CN CN202110441452.3A patent/CN113111831A/en not_active Withdrawn
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114098768A (en) * | 2021-11-25 | 2022-03-01 | 哈尔滨工业大学 | Cross-individual surface electromyographic signal gesture recognition method based on dynamic threshold and easy TL |
CN114098768B (en) * | 2021-11-25 | 2024-05-03 | 哈尔滨工业大学 | Cross-individual surface electromyographic signal gesture recognition method based on dynamic threshold and EasyTL |
CN114626412A (en) * | 2022-02-28 | 2022-06-14 | 长沙融创智胜电子科技有限公司 | Multi-class target identification method and system for unattended sensor system |
CN114626412B (en) * | 2022-02-28 | 2024-04-02 | 长沙融创智胜电子科技有限公司 | Multi-class target identification method and system for unattended sensor system |
CN115719514A (en) * | 2022-11-23 | 2023-02-28 | 南京理工大学 | Gesture recognition-oriented field self-adaptive method and system |
CN115719514B (en) * | 2022-11-23 | 2023-06-30 | 南京理工大学 | Gesture recognition-oriented field self-adaption method and system |
CN115778408A (en) * | 2022-12-10 | 2023-03-14 | 福州大学 | Step-by-step electromyographic signal activity segment detection method |
CN116725556A (en) * | 2023-07-12 | 2023-09-12 | 中国科学院苏州生物医学工程技术研究所 | Motion intention recognition method and device based on surface electromyographic signals |
CN117238037A (en) * | 2023-11-13 | 2023-12-15 | 中国科学技术大学 | Dynamic action recognition method, device, equipment and storage medium |
CN117238037B (en) * | 2023-11-13 | 2024-03-29 | 中国科学技术大学 | Dynamic action recognition method, device, equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Tuncer et al. | Surface EMG signal classification using ternary pattern and discrete wavelet transform based feature extraction for hand movement recognition | |
CN113111831A (en) | Gesture recognition technology based on multi-mode information fusion | |
CN102622605B (en) | Surface electromyogram signal feature extraction and action pattern recognition method | |
CN111209885A (en) | Gesture information processing method and device, electronic equipment and storage medium | |
CN110826625B (en) | Finger gesture classification method based on surface electromyographic signals | |
CN111544856B (en) | Brain-myoelectricity intelligent full limb rehabilitation method based on novel transfer learning model | |
CN112043473B (en) | Parallel nested and autonomous preferred classifier for brain-myoelectricity fusion perception of intelligent artificial limb | |
CN106236336A (en) | A kind of myoelectric limb gesture and dynamics control method | |
AlOmari et al. | Analysis of extracted forearm sEMG signal using LDA, QDA, K-NN classification algorithms | |
CN111584029B (en) | Electroencephalogram self-adaptive model based on discriminant confrontation network and application of electroencephalogram self-adaptive model in rehabilitation | |
Li et al. | EEG signal classification method based on feature priority analysis and CNN | |
CN113239744A (en) | Electroencephalogram-electromyogram signal fusion recognition method based on deep learning | |
Amanpreet | Machine learning-based novel approach to classify the shoulder motion of upper limb amputees | |
Wang et al. | Deep convolutional neural network for decoding EMG for human computer interaction | |
Zhao et al. | Decoding finger movement patterns from microscopic neural drive information based on deep learning | |
Lin et al. | A normalisation approach improves the performance of inter-subject sEMG-based hand gesture recognition with a ConvNet | |
AlOmari et al. | Novel hybrid soft computing pattern recognition system SVM–GAPSO for classification of eight different hand motions | |
CN112401905B (en) | Natural action electroencephalogram recognition method based on source localization and brain network | |
CN116831874A (en) | Lower limb rehabilitation device control method based on electromyographic signals | |
CN111923048A (en) | Electromyographic signal classification and exoskeleton robot control method and device | |
CN110604578A (en) | Human hand and hand motion recognition method based on SEMG | |
CN114098768A (en) | Cross-individual surface electromyographic signal gesture recognition method based on dynamic threshold and easy TL | |
CN114343679A (en) | Surface electromyogram signal upper limb action recognition method and system based on transfer learning | |
Zohirov | Classification Of Some Sensitive Motion Of Fingers To Create Modern Biointerface | |
Orban et al. | An Approach for Accurate Pattern Recognition of Four Hand Gestures Based on sEMG Signals |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WW01 | Invention patent application withdrawn after publication | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20210713 |