CN112183375A - Gesture recognition method fusing electromyographic signals and visual images - Google Patents

Gesture recognition method fusing electromyographic signals and visual images Download PDF

Info

Publication number
CN112183375A
CN112183375A CN202011053444.3A CN202011053444A CN112183375A CN 112183375 A CN112183375 A CN 112183375A CN 202011053444 A CN202011053444 A CN 202011053444A CN 112183375 A CN112183375 A CN 112183375A
Authority
CN
China
Prior art keywords
image
electromyographic
visual
gesture
fusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011053444.3A
Other languages
Chinese (zh)
Inventor
印二威
沈瑞
谢良
闫慧炯
邓宝松
罗治国
闫野
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin (binhai) Intelligence Military-Civil Integration Innovation Center
National Defense Technology Innovation Institute PLA Academy of Military Science
Original Assignee
Tianjin (binhai) Intelligence Military-Civil Integration Innovation Center
National Defense Technology Innovation Institute PLA Academy of Military Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin (binhai) Intelligence Military-Civil Integration Innovation Center, National Defense Technology Innovation Institute PLA Academy of Military Science filed Critical Tianjin (binhai) Intelligence Military-Civil Integration Innovation Center
Priority to CN202011053444.3A priority Critical patent/CN112183375A/en
Publication of CN112183375A publication Critical patent/CN112183375A/en
Priority to CN202110773623.2A priority patent/CN113269159A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/107Static hand or arm
    • G06V40/113Recognition of static hand signs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/02Preprocessing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/08Feature extraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a gesture recognition method fusing electromyographic signals and visual images, which comprises the following steps: acquiring an electromyographic signal by using a high-density electromyographic signal acquisition armband, acquiring a visual image by using a camera, and fusing the electromyographic image converted from the electromyographic signal and the visual image into an image; and inputting the fused image of the surface electromyogram and the visual image into an image classification module for identification and classification. According to the invention, two different network structures are not needed for identifying and classifying the surface electromyographic signals and the visual images, and the surface electromyographic signals and the visual images are only needed to be simultaneously transmitted into the network architecture, so that the identification and classification can be carried out by using the same network, the network complexity can be effectively reduced, and limited resources are saved. The invention can improve the form of gesture recognition by using a single input mode in each set of traditional network, and improves the respective limitations of the surface electromyogram signal and the visual image for gesture recognition respectively.

Description

Gesture recognition method fusing electromyographic signals and visual images
Technical Field
The invention relates to the field of gesture recognition, in particular to a gesture recognition technology fusing electromyographic signals and visual images.
Background
The invention is from the problem of gesture recognition of electromyographic signals and visual images. In the daily communication between people, gestures are a widely and frequently used interaction mode, and sign language is an important and convenient mode for the deaf-mute to communicate between people. Over the years, researchers have increasingly explored gesture recognition technology, and gesture recognition has long been one of the hot spots in the field of human-computer interaction.
There are many techniques for implementing gesture recognition, and the vision-based gesture recognition technique is most widely and conveniently researched and applied. The non-contact gesture motion acquisition method does not bring physical discomfort or movement obstruction to the user, has low cost and is favored in practical application. However, when image data for gesture recognition is acquired based on vision, many uncertainties exist, and the image data is susceptible to interference during application, for example, application conditions such as under different backgrounds, under different light, under different angles of an acquisition device (camera), and the like may become a restriction factor influencing gesture recognition accuracy, so that the gesture recognition technology based on vision can obtain a better effect under ideal conditions, but a great improvement space still exists in the recognition effect in practical application. The other technology for gesture recognition is gesture recognition based on electromyographic signals, compared with gesture recognition based on vision, gesture recognition based on the electromyographic signals has the greatest advantage that the gesture recognition based on the electromyographic signals is not influenced by external environments such as background conditions, light and the like, the defect of gesture recognition based on vision is overcome, the electromyographic signals can have differences among different individuals, the arrangement position of the collecting electrode has a large influence on the signals in the mode of collecting the electromyographic signals by using a sparse channel, and the factors increase the difficulty of gesture recognition based on the electromyographic signals.
Considering the advantages and disadvantages of the electromyographic signals and the visual images, better system performance can be obtained by fusing the electromyographic signals and the visual images, and meanwhile, because the data forms of the electromyographic signals and the visual images are different, two sets of networks are generally adopted when the electromyographic signals and the visual images are fused for use, which undoubtedly increases the calculation amount and the storage space. In order to solve the problems, the invention provides a gesture recognition method for fusing an electromyographic signal and a visual image.
Disclosure of Invention
In order to solve the problems of limited application scene of visual image gesture recognition and large difference of electromyographic signals among different individuals and reduce the calculated amount and the storage space, the invention provides a method for gesture recognition based on the electromyographic signals and the visual images by using a neural network. Collecting electromyographic signals by using high-density non-invasive equipment or invasive equipment, converting the electromyographic signals into electromyographic images, simultaneously collecting visual images, respectively carrying out corresponding processing on the two types of images according to a selected image fusion method, fusing the obtained electromyographic images and the visual images into an image, and inputting the fused image into an image classification module for identification and classification.
A gesture recognition method fusing electromyographic signals and visual images comprises the following specific steps:
s1, acquiring an electromyographic signal by using an electromyographic signal acquisition device, acquiring a visual image by using a camera, and converting the electromyographic signal and the visual image into an electromyographic image and a visual image which have the same size and can be used for subsequent fusion respectively;
s2, fusing the myoelectric image and the visual image by adopting an image fusion method to form an image;
and S3, inputting the fused images into an image recognition and classification module, and performing image classification and recognition to obtain a gesture recognition result.
The electromyographic signal acquisition comprises the steps of acquiring a high-density surface electromyographic signal by using a non-invasive device or acquiring a high-density electromyographic signal by using an invasive device;
the electromyographic signal acquisition equipment is high-density electromyographic signal acquisition armband equipment;
the image fusion method adopts a space domain fusion method or a transform domain fusion method;
the image recognition and classification module is a neural network-based classification and identification module.
The step S1 includes the following steps:
s11, setting the size of the image transmitted into the image classification module after fusion, determining the high-density electrode device and the placement position thereof according to the set size of the image, and collecting the electromyographic signals by using the device;
s12, selecting a certain moment, collecting an instantaneous value of the electromyographic signal at the moment, and generating a two-dimensional grid matrix according to the position of the electrode grid array by the instantaneous value;
s13, converting the two-dimensional grid matrix obtained in the step S12 into a gray scale map by utilizing linear transformation, wherein the obtained gray scale map is an electromyogram;
and S14, acquiring a visual image at the same time as the electromyographic signal, using equidistant sampling, reducing the visual image or amplifying the visual image by using an interpolation method, and converting the visual image into a gray map with the same size as the electromyographic image.
The step S2 includes the following steps:
and (4) superposing the two-dimensional images with the same size obtained in the step (S1) according to the dimension direction to obtain a three-dimensional image, namely a fusion image, which is used as the input of the subsequent neural network.
The step S3 of inputting the fused image into an image recognition and classification module for image classification and recognition includes the following steps:
the image identification and classification module comprises a feature extraction module and a classification module, the fused image firstly enters the feature extraction module for feature extraction, and then the output of the feature extraction module is used as the input of the classification module.
S31, firstly, inputting a data set into the feature extraction module to extract feature vectors;
and S32, inputting the feature vectors extracted in the step S31 into an identification module for identification and classification.
The method can automatically extract features by a network by adopting a deep convolutional neural network, and the related calculation of the deep convolutional neural network is realized by specifically comprising the following steps of:
Figure BDA0002710213550000031
wherein,
Figure BDA0002710213550000032
identifying an output of a l-th layer of the convolutional neural network for the gesture; i and j are indices of the rows and columns of the convolutional layer output image; f is a relu activation function; s is the size of the input image of the l-th layer;
Figure BDA0002710213550000033
is the input of the l layer;
Figure BDA0002710213550000034
a convolution kernel which is the l-th layer of convolution; b(l)Is the bias term for the l-th layer.
In the image recognition and classification module, the loss function used is as follows:
Figure BDA0002710213550000035
wherein N is the total number of samples; c is the number of gesture categories; y isi,kA predicted kth true gesture tag value for the ith sample; pi,kThe probability of predicting the kth gesture tag value for the ith sample.
Assuming that the number of gesture categories is C, the output layer classifies the vector with input of C × 1 using the softmax function, and outputs the following:
Figure BDA0002710213550000036
wherein S isjOutputting the jth value of the vector S for softmax; a isjIs the jth value in the C1 vector; a iskIs the kth value in the C x 1 vector.
An image recognition and classification module according to the loss function LlogAnd (Y, P) training the model by adopting a back propagation algorithm.
When predicting, the image recognition and classification module selects an index value with the largest value in a C x 1 vector output by a sample through the softmax layer as a predicted gesture label of the sample.
The myoelectric signal acquisition device is used for acquiring myoelectric signals, the camera is used for acquiring visual images, the myoelectric signals and the visual images are respectively converted into myoelectric images and visual images which have the same size and can be used for subsequent fusion, and the myoelectric image acquisition device specifically comprises:
determining the number of high-density electrodes for acquiring high-density electromyographic signals to be H multiplied by W according to the set resolution of the fused image to be H multiplied by W, and placing the electrodes in a uniform grid mode of H electrodes in each row and W electrodes in each column, wherein the amplitude of the signal acquired by each electrode is subsequently converted into the pixel value of the electromyographic image at the position; acquiring an electromyographic signal by using the high-density electrode, selecting a moment t, and acquiring an instantaneous value of the electromyographic signal at the moment; arranging instantaneous values of the electromyographic signals at the t moment into a two-dimensional H multiplied by W-dimensional grid point matrix according to grids of the electrode array, wherein each grid point in the matrix takes the value of the instantaneous value of the amplitude of the electromyographic signals acquired by a corresponding electrode at the t moment; and converting the obtained two-dimensional grid point matrix into a gray scale map by linear transformation, namely linearly transforming the instantaneous value of the amplitude of the electromyographic signal into the intensity of color to obtain the gray scale map which is the electromyographic image.
The invention has the following beneficial effects:
(1) greatly reduces the quantity of parameters, saves limited resources and improves the efficiency. After the electromyographic signals and the visual images are processed, the fused images are input into an image recognition and classification module, two different networks for respectively recognizing and classifying the electromyographic signals and the visual images are replaced by one network, the parameter quantity is reduced to the parameter quantity in the image recognition and classification module, overfitting caused by excessive parameters is weakened to a certain degree, meanwhile, calculation resources are saved due to the fact that the parameter quantity is greatly reduced, and calculation efficiency is improved.
(2) The method has higher accuracy in the recognition of both dynamic gestures and static gestures. The method integrates the electromyographic signal information and the visual image information, takes the advantages of the electromyographic signal information and the visual image information into account, and enables the recognition of the dynamic gesture and the static gesture to be high in accuracy.
(3) The advantages of the electromyographic signals and the visual images are integrated, and the method has strong adaptability to different external environments. The method integrates the electromyographic signal information and the visual image information, and has strong adaptability to different external environments.
Drawings
Fig. 1 is a flowchart illustrating a gesture recognition method based on fusion of an electromyographic signal and a visual image according to an embodiment of the present invention.
Fig. 2 is a schematic diagram of a high-density electromyographic signal acquisition armband device disclosed in the embodiment of the present invention.
Fig. 3 is a flowchart illustrating the electromyographic signal branch processing and conversion into an electromyographic image according to an embodiment of the present invention.
FIG. 4 is a flowchart illustrating a visual image branching process according to an embodiment of the present invention.
Fig. 5 is a schematic diagram illustrating a fusion mode of an electromyogram and a visual image according to an embodiment of the present invention.
Detailed Description
The embodiments of the present invention are described below with reference to specific embodiments, and other advantages and effects of the present invention will be easily understood by those skilled in the art from the disclosure of the present specification. The invention is capable of other and different embodiments and of being practiced or of being carried out in various ways, and its several details are capable of modification in various respects, all without departing from the spirit and scope of the present invention. It is to be noted that the features in the following embodiments and examples may be combined with each other without conflict.
The invention provides a gesture recognition method fusing an electromyographic signal and a visual image, wherein the processing flow is shown in fig. 1, only one of a plurality of methods for transforming the electromyographic image and fusing the image is selected for detailed explanation in the embodiment, the used high-density electromyographic signal acquisition armband device is shown in fig. 2, 101 in fig. 2 is a high-density electrode array, and 102 in fig. 2 is a fixing band.
Embodiment 1. gesture recognition method fusing electromyographic signals and visual images
A gesture recognition method fusing electromyographic signals and visual images comprises the following specific steps:
s1, acquiring an electromyographic signal by using an electromyographic signal acquisition device, acquiring a visual image by using a camera, and converting the electromyographic signal and the visual image into an electromyographic image and a visual image which have the same size and can be used for subsequent fusion respectively;
s2, fusing the myoelectric image and the visual image by adopting an image fusion method to form an image;
and S3, inputting the fused images into an image recognition and classification module, and performing image classification and recognition to obtain a gesture recognition result.
The electromyographic signal acquisition comprises the steps of acquiring a high-density surface electromyographic signal by using a non-invasive device or acquiring a high-density electromyographic signal by using an invasive device;
the electromyographic signal acquisition equipment is high-density electromyographic signal acquisition armband equipment;
the image fusion method adopts a space domain fusion method or a transform domain fusion method;
the image recognition and classification module is a neural network-based classification and identification module.
The step S1 includes the following steps:
s11, setting the size of the image transmitted into the image classification module after fusion, determining the high-density electrode device and the placement position thereof according to the set size of the image, and collecting the electromyographic signals by using the device;
s12, selecting a certain moment, collecting an instantaneous value of the electromyographic signal at the moment, and generating a two-dimensional grid matrix according to the position of the electrode grid array by the instantaneous value;
s13, converting the two-dimensional grid matrix obtained in the step S12 into a gray scale map by utilizing linear transformation, wherein the obtained gray scale map is an electromyogram;
and S14, acquiring a visual image at the same time as the electromyographic signal, using equidistant sampling, reducing the visual image or amplifying the visual image by using an interpolation method, and converting the visual image into a gray map with the same size as the electromyographic image.
The step S2 includes the following steps:
and (4) superposing the two-dimensional images with the same size obtained in the step (S1) according to the dimension direction to obtain a three-dimensional image, namely the fused image.
The step S3 of inputting the fused image into an image recognition and classification module for image classification and recognition includes the following steps:
the image identification and classification module comprises a feature extraction module and a classification module, the fused image firstly enters the feature extraction module for feature extraction, and then the output of the feature extraction module is used as the input of the classification module.
S31, firstly, inputting a data set into the feature extraction module to extract feature vectors;
and S32, inputting the feature vectors extracted in the step S31 into an identification module for identification and classification.
If the deep convolutional neural network is adopted, the features can be automatically extracted by the network, and the step of manually extracting the features is omitted.
The related calculation implementation process of the deep convolutional neural network is as follows:
the mapping process of convolutional layer forward propagation is expressed as:
Figure BDA0002710213550000061
wherein,
Figure BDA0002710213550000062
identifying an output of a l-th layer of the convolutional neural network for the gesture; i and j are indices of the rows and columns of the convolutional layer output image; f is a relu activation function; s is the size of the input image of the l-th layer;
Figure BDA0002710213550000063
is the input of the l layer;
Figure BDA0002710213550000064
a convolution kernel which is the l-th layer of convolution; b(l)Is the bias term for the l-th layer.
In the image recognition and classification module, the loss function used is as follows:
Figure BDA0002710213550000065
wherein N is the total number of samples; c is the number of gesture categories; y isi,kA predicted kth true gesture tag value for the ith sample; pi,kThe probability of predicting the kth gesture tag value for the ith sample.
Assuming that the number of gesture categories is C, the output layer classifies the vector with input of C × 1 using the softmax function, and outputs the following:
Figure BDA0002710213550000066
wherein S isjOutputting the jth value of the vector S for softmax; a isjIs the jth value in the C1 vector; a iskIs the kth value in the C x 1 vector.
An image recognition and classification module according to the loss function LlogAnd (Y, P) training the model by adopting a back propagation algorithm.
When predicting, the image recognition and classification module selects an index value with the largest value in a C x 1 vector output by a sample through the softmax layer as a predicted gesture label of the sample.
The present invention utilizes the collected electromyographic signals and the visual images to obtain two images with the same size in step S1, and specifically, the flow of obtaining the electromyographic images from the electromyographic signals is shown in fig. 3, and the visual image processing method is shown in fig. 4.
In step 2, the present invention performs fusion according to the two images with the same size obtained in step 1 to obtain an image transmitted to the subsequent image classification module, where the processing manner is shown in fig. 5.
Through the steps 1 and 2, the fusion of the electromyographic signals and the visual images is completed, and the obtained electromyographic-visual images contain the characteristics of the electromyographic signals and the characteristics of the visual images.
The method comprises 2 key steps, namely a method for converting a preprocessed high-density electromyogram signal into an electromyogram image and a method for fusing the electromyogram image and a visual image, wherein the 2 key steps are described in detail.
The high-density electromyographic signals are converted into electromyographic images after filtering and normalization preprocessing. Fig. 3 shows a flow chart of the process of preprocessing the high-density electromyographic signals and converting the preprocessed signals into electromyographic images. In order to improve the influence of the difference of the electrode placement positions on the final recognition result, the electromyographic signals are collected by adopting the high-density electrodes in the example, and the high-density electrode device and the placement positions thereof are determined according to the set image size, and the specific steps are as follows:
according to the set fused image resolution H multiplied by W, determining that the number of high-density electrodes for acquiring high-density electromyographic signals is H multiplied by W, and placing the electrodes in a uniform grid mode of H electrodes in each row and W electrodes in each column, and subsequently converting the signal amplitude acquired by each electrode into the pixel value of the electromyographic image at the position; acquiring an electromyographic signal by using the high-density electrode, selecting a moment t, and acquiring an instantaneous value of the electromyographic signal at the moment; arranging instantaneous values of the electromyographic signals at the t moment into a two-dimensional H multiplied by W-dimensional grid point matrix according to grids of the electrode array, wherein each grid point in the matrix takes the value of the instantaneous value of the amplitude of the electromyographic signals acquired by a corresponding electrode at the t moment; converting the obtained two-dimensional grid point matrix into a gray scale by linear transformation, namely linearly transforming the instantaneous value of the amplitude of the electromyographic signal into the intensity of color to obtain the gray scale which is the electromyographic image;
for the fusion method of the electromyogram and the visual image, fig. 4 and 5 show a flow chart for converting the visual image into the gray scale and a schematic diagram of a fusion mode of the electromyogram and the visual image with the same size. Collecting a visual image at the same time of t moment as the electromyographic signal while collecting the electromyographic signal, and converting the visual image into a gray scale image with H multiplied by W size; and superposing the two obtained two-dimensional images with the same size according to the dimension direction to obtain a three-dimensional image, namely the input of the subsequent neural network.
Such implementations include, but are not limited to, the methods described above. The foregoing embodiments are merely illustrative of the principles and utilities of the present invention and are not intended to limit the invention. Any person skilled in the art can modify or change the above-mentioned embodiments without departing from the spirit and scope of the present invention. Accordingly, it is intended that all equivalent modifications or changes which can be made by those skilled in the art without departing from the spirit and technical spirit of the present invention be covered by the claims of the present invention.

Claims (9)

1. A gesture recognition method fusing an electromyographic signal and a visual image is characterized by comprising the following specific steps:
s1, acquiring an electromyographic signal by using an electromyographic signal acquisition device, acquiring a visual image by using a camera, and converting the electromyographic signal and the visual image into an electromyographic image and a visual image which have the same size and can be used for subsequent fusion respectively;
s2, fusing the myoelectric image and the visual image by adopting an image fusion method to form an image;
and S3, inputting the fused images into an image recognition and classification module, and performing image classification and recognition to obtain a gesture recognition result.
2. The method for gesture recognition based on fusion of electromyographic signals and visual images according to claim 1, wherein the acquiring of the electromyographic signals comprises acquiring high-density surface electromyographic signals by using a non-invasive device or acquiring high-density electromyographic signals by using an invasive device.
3. The method for recognizing the gesture based on the fusion of the electromyographic signal and the visual image according to claim 1, wherein the electromyographic signal acquisition device is a high-density electromyographic signal acquisition armband device.
4. A gesture recognition method according to claim 1, wherein the image fusion method is a spatial domain fusion method or a transform domain fusion method.
5. The method for recognizing the gesture based on the fusion of the electromyographic signal and the visual image according to claim 1, wherein the image recognition and classification module is a neural network-based classification and recognition module.
6. The method for recognizing a gesture based on a fusion of an electromyographic signal and a visual image according to claim 1, wherein the step S1 comprises the following steps:
s11, setting the size of the image transmitted into the image classification module after fusion, determining the high-density electrode device and the placement position thereof according to the set size of the image, and collecting the electromyographic signals by using the device;
s12, selecting a certain moment, collecting an instantaneous value of the electromyographic signal at the moment, and generating a two-dimensional grid matrix according to the position of the electrode grid array by the instantaneous value;
s13, converting the two-dimensional grid matrix obtained in the step S12 into a gray scale map by utilizing linear transformation, wherein the obtained gray scale map is an electromyogram;
and S14, acquiring a visual image at the same time as the electromyographic signal, using equidistant sampling, reducing the visual image or amplifying the visual image by using an interpolation method, and converting the visual image into a gray map with the same size as the electromyographic image.
7. The method for recognizing a gesture based on a fusion of an electromyographic signal and a visual image according to claim 1, wherein the step S2 comprises the following steps:
and (4) superposing the two-dimensional images with the same size obtained in the step (S1) according to the dimension direction to obtain a three-dimensional image, namely a fusion image, which is used as the input of the subsequent neural network.
8. The method for gesture recognition based on fusion of electromyographic signals and visual images according to claim 1, wherein the step S3 is performed by inputting the fused images into an image recognition and classification module for image classification and recognition, and specifically comprises the following steps:
the image identification and classification module comprises a feature extraction module and a classification module, the fused image firstly enters the feature extraction module for feature extraction, and then the output of the feature extraction module is used as the input of the classification module;
s31, firstly, inputting a data set into the feature extraction module to extract feature vectors;
s32, inputting the feature vectors extracted in the step S31 into an identification module for identification and classification;
the method adopts a deep convolutional neural network to autonomously extract features by the network, and the related calculation of the deep convolutional neural network is realized, specifically, the forward propagation mapping process of the convolutional layer is expressed as follows:
Figure FDA0002710213540000021
wherein,
Figure FDA0002710213540000022
identifying an output of a l-th layer of the convolutional neural network for the gesture; i and j are indices of the rows and columns of the convolutional layer output image; f is a relu activation function; s is the size of the input image of the l-th layer;
Figure FDA0002710213540000023
is the input of the l layer;
Figure FDA0002710213540000024
a convolution kernel which is the l-th layer of convolution; b(l)Is the bias term of the l layer;
in the image recognition and classification module, the loss function used is as follows:
Figure FDA0002710213540000025
wherein N is the total number of samples; c is the number of gesture categories; y isi,kA predicted kth true gesture tag value for the ith sample; pi,kPredicting a probability of a kth gesture tag value for the ith sample;
assuming that the number of gesture categories is C, the output layer classifies the vector with input of C × 1 using the softmax function, and outputs the following:
Figure FDA0002710213540000026
wherein S isjOutputting the jth value of the vector S for softmax; a isjIs the jth value in the C1 vector; a iskIs the kth value in the C1 vector;
an image recognition and classification module according to the loss function Llog(Y, P), training the model by adopting a back propagation algorithm;
when predicting, the image recognition and classification module selects an index value with the largest value in a C x 1 vector output by a sample through the softmax layer as a predicted gesture label of the sample.
9. The method for gesture recognition based on fusion of electromyographic signals and visual images according to claim 1, wherein the electromyographic signals are collected by an electromyographic signal collecting device, the visual images are collected by a camera, and the electromyographic signals and the visual images are respectively converted into the same size of electromyographic images and the same size of visual images which can be used for subsequent fusion, specifically comprising:
determining the number of high-density electrodes for acquiring high-density electromyographic signals to be H multiplied by W according to the set resolution of the fused image to be H multiplied by W, and placing the electrodes in a uniform grid mode of H electrodes in each row and W electrodes in each column, wherein the amplitude of the signal acquired by each electrode is subsequently converted into the pixel value of the electromyographic image at the position; acquiring an electromyographic signal by using the high-density electrode, selecting a moment t, and acquiring an instantaneous value of the electromyographic signal at the moment; arranging instantaneous values of the electromyographic signals at the t moment into a two-dimensional H multiplied by W-dimensional grid point matrix according to grids of the electrode array, wherein each grid point in the matrix takes the value of the instantaneous value of the amplitude of the electromyographic signals acquired by a corresponding electrode at the t moment; and converting the obtained two-dimensional grid point matrix into a gray scale map by linear transformation, namely linearly transforming the instantaneous value of the amplitude of the electromyographic signal into the intensity of color to obtain the gray scale map which is the electromyographic image.
CN202011053444.3A 2020-09-29 2020-09-29 Gesture recognition method fusing electromyographic signals and visual images Pending CN112183375A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202011053444.3A CN112183375A (en) 2020-09-29 2020-09-29 Gesture recognition method fusing electromyographic signals and visual images
CN202110773623.2A CN113269159A (en) 2020-09-29 2021-07-08 Gesture recognition method fusing electromyographic signals and visual images

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011053444.3A CN112183375A (en) 2020-09-29 2020-09-29 Gesture recognition method fusing electromyographic signals and visual images

Publications (1)

Publication Number Publication Date
CN112183375A true CN112183375A (en) 2021-01-05

Family

ID=73946697

Family Applications (2)

Application Number Title Priority Date Filing Date
CN202011053444.3A Pending CN112183375A (en) 2020-09-29 2020-09-29 Gesture recognition method fusing electromyographic signals and visual images
CN202110773623.2A Pending CN113269159A (en) 2020-09-29 2021-07-08 Gesture recognition method fusing electromyographic signals and visual images

Family Applications After (1)

Application Number Title Priority Date Filing Date
CN202110773623.2A Pending CN113269159A (en) 2020-09-29 2021-07-08 Gesture recognition method fusing electromyographic signals and visual images

Country Status (1)

Country Link
CN (2) CN112183375A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116449967A (en) * 2023-06-20 2023-07-18 浙江强脑科技有限公司 Bionic hand teaching aid, control method thereof and main control equipment

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9486332B2 (en) * 2011-04-15 2016-11-08 The Johns Hopkins University Multi-modal neural interfacing for prosthetic devices
CN109498362A (en) * 2018-09-10 2019-03-22 南京航空航天大学 A kind of hemiplegic patient's hand movement function device for healing and training and model training method
CN110298286B (en) * 2019-06-24 2021-04-30 中国科学院深圳先进技术研究院 Virtual reality rehabilitation training method and system based on surface myoelectricity and depth image
CN111475024B (en) * 2019-12-25 2024-04-16 山东中科先进技术有限公司 Human motion capturing system and method
CN111476161A (en) * 2020-04-07 2020-07-31 金陵科技学院 Somatosensory dynamic gesture recognition method fusing image and physiological signal dual channels

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116449967A (en) * 2023-06-20 2023-07-18 浙江强脑科技有限公司 Bionic hand teaching aid, control method thereof and main control equipment

Also Published As

Publication number Publication date
CN113269159A (en) 2021-08-17

Similar Documents

Publication Publication Date Title
CN111339903B (en) Multi-person human body posture estimation method
CN109344701B (en) Kinect-based dynamic gesture recognition method
CN111401384B (en) Transformer equipment defect image matching method
CN112949565A (en) Single-sample partially-shielded face recognition method and system based on attention mechanism
CN109146831A (en) Remote sensing image fusion method and system based on double branch deep learning networks
CN111814661A (en) Human behavior identification method based on residual error-recurrent neural network
CN112184734B (en) Animal long-time gesture recognition system based on infrared image and wearable optical fiber
CN112163508A (en) Character recognition method and system based on real scene and OCR terminal
CN112244878A (en) Method for identifying key frequency band image sequence by using parallel multi-module CNN and LSTM
CN110929685A (en) Pedestrian detection network structure based on mixed feature pyramid and mixed expansion convolution
CN113010013A (en) Wasserstein distance-based motor imagery electroencephalogram migration learning method
CN111209873A (en) High-precision face key point positioning method and system based on deep learning
Bose et al. In-situ recognition of hand gesture via Enhanced Xception based single-stage deep convolutional neural network
CN115471757A (en) Hyperspectral image classification method based on convolutional neural network and attention mechanism
Yang et al. DGLT-Fusion: A decoupled global–local infrared and visible image fusion transformer
CN110135435B (en) Saliency detection method and device based on breadth learning system
Bachay et al. Hybrid Deep Learning Model Based on Autoencoder and CNN for Palmprint Authentication.
CN113792807B (en) Skin disease classification model training method, system, medium and electronic equipment
CN112884062B (en) Motor imagery classification method and system based on CNN classification model and generated countermeasure network
CN113269159A (en) Gesture recognition method fusing electromyographic signals and visual images
CN114120359A (en) Method for measuring body size of group-fed pigs based on stacked hourglass network
CN113627391A (en) Cross-mode electroencephalogram signal identification method considering individual difference
CN116580289A (en) Fine granularity image recognition method based on attention
CN116469172A (en) Bone behavior recognition video frame extraction method and system under multiple time scales
CN113963427B (en) Method and system for rapid in-vivo detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20210105

WD01 Invention patent application deemed withdrawn after publication