CN110348288B - Gesture recognition method based on 77GHz millimeter wave radar signal - Google Patents
Gesture recognition method based on 77GHz millimeter wave radar signal Download PDFInfo
- Publication number
- CN110348288B CN110348288B CN201910445702.3A CN201910445702A CN110348288B CN 110348288 B CN110348288 B CN 110348288B CN 201910445702 A CN201910445702 A CN 201910445702A CN 110348288 B CN110348288 B CN 110348288B
- Authority
- CN
- China
- Prior art keywords
- spectrogram
- time
- gesture
- distance
- size
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/02—Preprocessing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/08—Feature extraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2218/00—Aspects of pattern recognition specially adapted for signal processing
- G06F2218/12—Classification; Matching
Abstract
The invention provides a gesture recognition method based on 77GHz millimeter wave radar signals, which comprises the steps of firstly obtaining intermediate frequency signals of different gesture actions through a radar, innovatively utilizing an improved wavelet threshold function to preprocess low-frequency coefficients of the intermediate frequency signals, solving the problem that short-distance gestures cannot be recognized due to an antenna coupling phenomenon, secondly extracting a time-distance spectrogram, a time-speed spectrogram and a time-angle spectrogram from the preprocessed intermediate frequency signals, innovatively splicing the three characteristic spectrograms to obtain a diversified characteristic map, inputting the diversified characteristic map into a convolutional neural network for training, optimizing the problem that the traditional recognition algorithm is incomplete in information expression, simultaneously being beneficial to simplifying a network structure and finally obtaining a better recognition effect.
Description
Technical Field
The invention relates to the technical field of radar signal processing and recognition, in particular to a gesture recognition method based on 77GHz millimeter wave radar signals.
Background
Since the twenty-first century, with the rapid development of computer technology, human-computer interaction technology has become one of the major subject technologies today. At present, a mouse and a keyboard are used as mechanical input devices in common man-machine interaction methods, however, the methods have no way to realize simple, efficient and highly free information interaction between a person and a computer. In the development process of the field of computers and signal processing, the gesture recognition technology has more and more application scenes with the characteristics of lively, vivid, visual and efficient expression, such as an intelligent home system, a sign language real-time teaching system, a gesture control game system and the like. With the rapid development of human-computer interaction technology, gesture recognition technology has become a research hotspot for students at home and abroad.
Most of the traditional gesture recognition technologies are based on videos and images, for example, a motion sensing device Kinect of microsoft corporation forms a scene graph with depth by using a 3D motion sensing camera and an optical coding technology, performs image depth recognition by evaluating the pixel level of a depth image, and captures each part of a human body or gesture actions by combining a human skeleton tracking technology. However, conventional image or video-based gesture recognition methods have certain limitations. Firstly, the recognition accuracy of the traditional gesture recognition technology based on images is very easily influenced by factors such as illumination, weather and working environment. Secondly, the traditional gesture recognition method based on images or videos is also easily affected by occlusion, such as a wall, a bookcase, etc., and when the gesture performer is located behind the wall or at a position in a room that is partially or completely occluded, the method is completely impossible. In addition, the traditional gesture recognition method based on images or videos also has the risk of revealing privacy of users. In the era of high sensitivity of personal information, the privacy disclosure problem caused by the method can have serious consequences. Finally, the traditional gesture recognition method based on images or videos has relatively high requirements on computing resources and energy consumption, and generally requires an independent external power supply system, so that the application scene and the scale of the gesture recognition method are greatly limited.
Compared with the traditional gesture recognition method based on videos or images, the gesture recognition method based on radar signals generally has the characteristics of non-contact, no influence of illumination, weather and working environment and the like, and can effectively solve the problem that the recognition accuracy is influenced by conditions such as insufficient illumination and the like. Meanwhile, the radar also has a certain function of penetrating through shielding propagation, so that the influence of shielding objects such as walls and bookcases can be effectively avoided, and the gesture executor can realize gesture control and interaction under the condition of being completely shielded or partially shielded. And the gesture recognition method based on the radar signals can effectively eliminate the problem of user privacy disclosure caused by video or image shooting, and the advantage of the method protects the user privacy and simultaneously guarantees the safety of the user. In addition, most radar sensors can be integrated on a high-quality chip with low energy consumption, so that the identification cost can be greatly reduced, the calculation complexity can be reduced, and the application scene and scale of the gesture identification technology can be greatly increased. The 77GHz millimeter wave radar is increasingly widely applied due to the characteristics of light weight and small volume. Meanwhile, the 77GHz millimeter wave radar can easily achieve higher spatial resolution, so that the precision of distance measurement, angle measurement and speed measurement is higher. By combining the application of the existing 77GHz millimeter wave radar in the aspects of intelligent driving, intelligent home and the like, the gesture recognition technology based on the 77GHz millimeter wave radar has a very wide application prospect.
The existing gesture recognition method based on radar signals has the following problems. First, for low-frequency signals with large energy generated by coupling of a transmitting antenna and a receiving antenna, the current method intercepts signals from a distance domain and discards close-range information. In fact, for dynamic gesture motions, it is not scientific to remove useful gesture signals altogether when the gesture is closer to the radar. Secondly, the utilization of radar angle information is less at present, useful information provided by a radar is not fully utilized, information fusion of a characteristic spectrogram is lacked when an input data set of the convolutional neural network is constructed, and the design difficulty of the convolutional neural network in the later period is increased.
Disclosure of Invention
The invention aims to provide a method for removing low-frequency signals with larger energy generated by coupling of a transmitting antenna and a receiving antenna, and a method for constructing a radar gesture signal diversified characteristic map. Compared with the traditional gesture recognition technology, on one hand, the short-distance gesture signals can be recognized, and the application scenes of the gesture recognition technology are further increased; on the other hand, the gesture signal feature expression is more complete, the problem of low description of the previous gesture information amount is solved, the design of a convolutional neural network in the later period is facilitated to be simplified, and the accurate classification of various gestures is facilitated to be realized.
The invention relates to a gesture recognition method based on 77GHz millimeter wave radar signals, which comprises the following steps:
designing N gesture actions, taking N =4 as an example, designing four gesture actions of hooking, radially waving hands, clockwise rotating and anticlockwise rotating, and acquiring data of corresponding gestures by different volunteers in a microwave darkroom environment to obtain 4 groups of data of ClassNum;
step two, transmitting a signal S T (t) and a received signal S R (t) obtaining a mixed signal S by means of a mixer MIX (t),S MIX (t) analyzing the intermediate frequency signal x (t) through a low-pass filter, and extracting a signal SN corresponding to a corresponding receiving antenna from the intermediate frequency signal I\Q (t);
Step three, due to the coupling phenomenon of the transmitting antenna and the receiving antenna, a low-frequency signal with large energy exists in the gesture intermediate-frequency signal x (t), and the intermediate-frequency signal SN is processed by utilizing a wavelet threshold method I\Q (t) preprocessing, selecting an improved threshold function, processing only the low-frequency coefficient, and reconstructing to obtain a new intermediate-frequency signal
Step four, centering the intermediate frequency signalProcessing, namely estimating a time-distance spectrogram, a time-velocity spectrogram and a time-angle spectrogram of the gesture action, and performing numerical value normalization processing on the three spectrograms respectively;
step five, splicing the time-distance spectrogram, the time-velocity spectrogram and the time-angle spectrogram after normalization in the step four to construct a diversified characteristic spectrogram A;
step six, respectively carrying out the operations of step two to step six on the collected 4 groups of ClassNum group gesture echo dataThen, 4 sets of ClassNum original gesture signal diversified feature spectrum atlas are obtained
Step seven, the diversified characteristic spectrum atlas obtained in the step six is collectedCarrying out graying processing on each sample to obtain an original diversified feature map set B;
eighthly, removing the mean value of all samples in the original diversified feature atlas B and carrying out scale normalization to obtain a diversified feature atlasLabeling each characteristic graph;
step nine, a diversified feature map setDivided into training sets S according to a certain proportion train Verification set S val And test set S test For example, the training set accounts for 70%, the validation set accounts for 20%, and the test set accounts for 10%;
step ten, adding S train 、S val As input data set C of convolutional neural network together with its corresponding label input And initializing network weights, wherein S train For training the network coefficients, S val After training for a period of time, carrying out network verification, and adjusting the weight of the network through errors;
step eleven, inputting a data set C input Performing convolution pooling operation once, and setting the scale kernel _ size1, convolution step size kernel _ stride1, pooling size pool _ size1 and pooling step size pool _ stride1 of a convolution kernel to obtain a feature map set feature1;
step twelve, further performing convolution pooling on the feature map set feature1, extracting deep features, and setting the size kernel _ size2, convolution step size kernel _ stride2, pooling size pool _ size2 and pooling step size pool _ stride2 of a convolution kernel to obtain a feature map set feature2;
step thirteen, performing convolution pooling operation on the feature map set feature2 again to extract deeper features, and setting the size kernel _ size3, convolution step size kernel _ stride3, pooling size pool _ size3 and pooling step size pool _ stride3 of a convolution kernel to obtain a feature map set feature3;
step fourteen, sequentially passing the feature3 through full connection layers fc4, fc5 and fc6, respectively setting the sizes of fc4, fc5 and fc6 as size4, size5 and size6, and converting the feature map set into a column vector v1 of 1 × size 6;
and step fifteen, outputting the column vector v1 into different gesture categories through a softmax classifier, and obtaining a trained convolutional neural network model NetModel through multiple iterations, wherein the network accuracy and the loss function tend to be stable.
Sixthly, testing the data set S test And loading the gesture classification result y into the NetModel.
The third step comprises the following steps:
3.1 selecting proper wavelet basis function to gesture intermediate frequency signal SN I\Q (t) carrying out N-layer wavelet transform to obtain an approximate coefficient A (i) and detail coefficients D (1, i), D (2, i), \\ 8230;, D (N, i);
3.2 thresholding the low-frequency coefficients A (i), i.e.
A i '=mA i +(1-m)sgn(A i )(|A i |-n)
3.3 utilization ofD (1, i), D (2, i), \8230, and D (N, i) reconstruct the signals to obtain a preprocessed gesture intermediate frequency signal x' (t).
The fourth step comprises the following steps:
4.1 extracting a time-distance spectrogram: firstly, FFT is carried out in a fast time domain, then weighting and averaging are carried out in a slow time domain, and finally interframe accumulation is carried out to obtain a time-distance spectrogram with the size of FFTNum1 x FrameNum;
4.2 extraction time-velocity spectrogram: firstly, carrying out two-dimensional FFT by taking a frame as a unit to obtain a distance-Doppler image, then searching for the maximum value of the distance-Doppler image, extracting a line where the maximum value is positioned, taking speed information of the line as the speed of a current frame signal, and finally carrying out accumulation between frames to obtain a time-speed spectrogram, wherein the size of the time-speed spectrogram is FFTNum2 x FrameNum;
4.3 extracting a time-angle spectrogram: firstly, taking a frame as a unit, respectively carrying out two-dimensional FFT on corresponding frame signals of N receiving antennas to obtain N distance-Doppler graphs. Respectively extracting maximum values from the N distance-Doppler graphs, arranging the maximum values into 1 XN one-dimensional vectors, performing FFT on the one-dimensional vectors to obtain speed information corresponding to the current frame signal, and finally performing accumulation between frames to obtain a time-angle graph with the size of FFTNum3 x FrameNum;
4.4 numerical normalization treatment: firstly, performing dispersion standardization on each characteristic spectrogram, specifically, taking a time-distance spectrogram D as an example, performing numerical scaling on the time-distance spectrogram according to the following formula:
by the above scaling operation, each value in the time-distance spectrum falls within the [0,1] range.
Description of the drawings:
FIG. 1 is a flow chart of the present invention.
FIG. 2 illustrates specific gesture actions designed and identified in the present invention.
FIG. 3 is a simulation diagram of wavelet low-frequency threshold processing for each gesture.
Fig. 4 is a time-distance spectrogram simulation diagram of each gesture.
Fig. 5 is a time-velocity spectrogram simulation diagram of each gesture.
Fig. 6 is a time-angle spectrogram simulation diagram of each gesture.
Fig. 7 is a diversified feature map (normalized result) of each gesture operation.
Fig. 8 is a network architecture of a convolutional neural network.
Fig. 9 is table 1.
Detailed Description
The invention is further described with reference to the following figures and specific embodiments.
Take four gestures as an example:
designing four gesture actions of hooking, radial waving, clockwise rotation and anticlockwise rotation, and firstly configuring relevant parameters of the 77GHz millimeter wave radar as shown in fig. 2. In this patent, set up sampling frequency and be 2000kHz, the frame cycle is 55ms, gathers 100 frame data at every turn, has 128 chirp signals in every frame, and every chirp signal has 64 sampling points. The antennas adopt one transmitting antenna and four receiving antennas, namely one transmitting antenna and four receiving antennas, and corresponding gesture data acquisition is carried out by three volunteers in a microwave darkroom environment;
step two, transmitting a signal S T (t) and the received signal S R (t) mixing by means of a mixer to obtain a mixed signal S MIX (t) and mixing S MIX (t) obtaining the intermediate frequency signal x (t) through a low-pass filter, and extracting the signal SN corresponding to the corresponding receiving antenna from the intermediate frequency signal I\Q (t), the concrete steps are as follows:
the method comprises the following steps: the specific expression of the 2-1 radar emission signal is as follows:
wherein, A T To transmit the amplitude of the signal, f c Is the carrier frequency, T is the sawtooth period, B is the signal bandwidth, f T (T) is the frequency of the transmitted signal during time T;
step 2-2, the radar receiving signal is the result after the transmitting signal is delayed by delta t, and the result is as follows:
wherein A is R To receive the amplitude of the signal, f R (T) is the frequency of the received signal within the time T, and Δ f is the frequency offset;
step 2-3: will transmit signal S T (t) and a received signal S R (t) obtaining a mixed signal S by means of a mixer MIX The expression of (t) is as follows:
S MIX (t)=S T (t)S R (t)
step 2-4: mixing the signal S MIX (t) passing through a low-frequency filter, obtaining an intermediate-frequency signal x (t) with the following expression:
step 2-5: one-transmitting four-receiving is adopted, the four receiving antennas have 8 paths of I \ Q channels in total, the intermediate frequency signal x (t) is assigned to one path of data every 8 points, and the intermediate frequency signal corresponding to the antenna of the missing path can be extractedNumber SN I\Q (t);
Step three, because of the coupling phenomenon of the transmitting antenna and the receiving antenna, a low-frequency signal with large energy exists in the gesture intermediate-frequency signal, and the intermediate-frequency signal SN is subjected to wavelet threshold method I\Q (t) preprocessing, selecting an improved threshold function, processing only the low-frequency coefficient, and reconstructing to obtain a new intermediate-frequency signalThe method comprises the following specific steps:
step 3-1: sym6 is selected as a wavelet basis function to handle gesture intermediate frequency signalsAfter three-layer wavelet decomposition, approximate coefficients A (i), detail coefficients D (1, i), D (2, i) and D (3, i) are obtained,
step 3-2: thresholding its low-frequency coefficients A (i), i.e.
A i '=mA i +(1-m)sgn(A i )(|A i |-n)
Step 3-3: by usingD (1, i), D (2, i), D (3, i) are reconstructedAnd obtaining a preprocessed signal x' (t), wherein the simulation result of wavelet low-frequency threshold processing of each gesture is shown in fig. 3.
Step four, centering the intermediate frequency signalAnd processing, estimating a time-distance spectrogram, a time-velocity spectrogram and a time-angle spectrogram of the gesture action, and performing numerical value normalization processing on the three spectrograms respectively. The method comprises the following specific steps:
4.1 extracting a time-distance spectrogram: firstly, FFT is carried out in a fast time domain, then weighting and averaging are carried out in a slow time domain, and finally interframe accumulation is carried out to obtain a time-distance spectrogram, wherein the size of the time-distance spectrogram is FFTNum1 x FrameNum, and the simulation result of each gesture action time-distance spectrogram is shown in figure 4;
4.2 extraction time-velocity spectrogram: firstly, performing two-dimensional FFT by taking a frame as a unit to obtain a distance-Doppler diagram, searching for the maximum value of the distance-Doppler diagram, extracting a line where the maximum value is located, taking speed information of the line as the speed of a current frame signal, and finally performing accumulation between frames to obtain a time-speed spectrogram, wherein the size of the time-speed spectrogram is FFTNum2 x FrameNum, and the simulation result of each gesture action time-speed spectrogram is shown in FIG. 5;
4.3 extracting a time-angle spectrogram: firstly, taking a frame as a unit, respectively carrying out two-dimensional FFT on corresponding frame signals of N receiving antennas to obtain N distance-Doppler graphs. Maximum values are respectively extracted from the N distance-Doppler graphs, the maximum values are arranged into 1 multiplied by N one-dimensional vectors, FFT is carried out on the one-dimensional vectors, speed information corresponding to the current frame signal can be obtained, finally, accumulation is carried out between frames, a time-angle graph with the size of FFTNum3 x FrameNum can be obtained, and simulation results of time-angle spectrograms of all gesture actions are shown in figure 6;
4.4 numerical normalization treatment: firstly, performing dispersion standardization on each feature spectrogram, specifically, taking a time-distance spectrogram D as an example, performing numerical scaling on the time-distance spectrogram according to the following formula:
by the above scaling operation, each value in the time-distance spectrogram falls within the [0,1] range.
Step five, splicing the time-distance spectrogram, the time-velocity spectrogram and the time-angle spectrogram after normalization in the step four according to columns to construct a diversified characteristic spectrogram A with the size of (FFTNum 1+ FFTNum2+ FFTNum 3) × FrameNum;
step six, respectively carrying out operations of the step two to the step six on the collected 4 groups of ClassNum group gesture echo data to obtain a diversified characteristic spectrum atlas set of the 4 groups of ClassNum group original gesture signals
Step seven, the diversified characteristic spectrum atlas obtained in the step six is subjected toCarrying out graying processing on each sample to obtain an original diversified feature map set B;
eighthly, removing the mean value of all samples in the original diversified feature atlas B and carrying out scale normalization to obtain a diversified feature atlasThe results of the gesture-action diversification feature maps are shown in fig. 7, and each feature map is labeled, the collective steps are as follows:
8.1 averaging all samples in the multivariate feature set B according to
It is worth noting here that all samples in the multivariate data set B are averaged here, rather than just within a class for a certain type of gesture;
8.2 averaging, i.e. averaging, of any sample in the set B of multivariate feature maps
8.3 pairs of diversified feature mapsAnd (3) carrying out scale normalization: normalizing the scale of each feature map in the data set into Hight multiplied by Width, specifically, carrying out down sampling when the size of the original feature map is larger than the Hight multiplied by Width, otherwise carrying out up sampling, and finally obtaining the normalized diversified feature map set->
Step nine, a diversified feature map setDivided into training sets S according to a certain proportion train Verification set S val And test set S test For example, the training set accounts for 70%, the validation set accounts for 20%, and the test set accounts for 10%;
step ten, adding S train 、S val Together with its corresponding label as input data set C of the convolutional neural network input And initializing network weights, wherein S train For training the network coefficients, S val After training for a period of time, carrying out network verification, and adjusting the weight of the network through error back propagation;
step eleven, inputting a data set C input Performing convolution pooling operation once, and setting the scale 11 × 11 of a convolution kernel, the convolution step 4, the pooling size3 × 3 and the pooling step 2 to obtain a feature map set feature1;
step twelve, further performing convolution pooling on the feature map set feature1, extracting deep features, and setting the size of a convolution kernel of 5 × 5, the convolution step size1, the pooling size of 3 × 3 and the pooling step size of 2 to obtain a feature map set feature2;
step thirteen, performing convolution pooling operation on the feature map set feature2 again to extract deeper features, and setting the size of a convolution kernel of 3 × 3, the convolution step size of 1, the pooling size of 3 × 3 and the pooling step size of 2 to obtain a feature map set feature3;
step fourteen, sequentially passing feature3 through full connection layers fc4, fc5 and fc6, respectively setting the sizes of fc4, fc5 and fc6 to 4096, 2048 and 1000, and converting the feature diagram set into a 1 × 1000 column vector v1;
and step fifteen, outputting the column vector v1 into different gesture categories through a softmax classifier, and obtaining a trained convolutional neural network model NetModel through multiple iterations when the network accuracy and the loss function tend to be stable.
Sixthly, testing the data set S test And loading the gesture data into a NetModel to obtain a gesture classification result y, and obtaining a gesture data classification result shown in a table 1.
Claims (2)
1. A gesture recognition method based on 77GHz millimeter wave radar signals is characterized by comprising the following steps:
designing N gesture actions, and acquiring data of corresponding gestures by different volunteers in a microwave darkroom environment to obtain N ClassNum group data in total;
step two, analyzing the radar original data to obtain an intermediate frequency signal x (t), and extracting an intermediate frequency signal SN corresponding to a corresponding receiving antenna from the intermediate frequency signal I\Q (t);
Step three, due to the coupling phenomenon of the transmitting antenna and the receiving antenna, a low-frequency signal with large energy exists in the gesture intermediate-frequency signal x (t), and the intermediate-frequency signal SN is subjected to wavelet threshold method I\Q (t) preprocessing, selecting an improved threshold function, processing only the low-frequency coefficient, and reconstructing to obtain a new intermediate-frequency signal
Step four, centering the intermediate frequency signalProcessing, namely estimating a time-distance spectrogram, a time-velocity spectrogram and a time-angle spectrogram of the gesture action, and performing numerical value normalization processing on the three spectrograms respectively;
step five, splicing the time-distance spectrogram, the time-velocity spectrogram and the time-angle spectrogram after normalization in the step four to construct a diversified characteristic spectrogram A;
step six, respectively carrying out operations of the step two to the step six on the collected N groups of ClassNum gesture echo data to obtain N groups of ClassNum original gesture signal diversification feature spectrum atlas
Step seven, the diversified characteristic spectrum atlas obtained in the step six is collectedCarrying out graying processing on each sample to obtain an original diversified feature map set B;
eighthly, removing the mean value of all samples in the original diversified feature atlas B and carrying out scale normalization to obtain a diversified feature atlasLabeling each characteristic graph;
step nine, a diversified feature map setDivided into training sets S according to a certain proportion train And a verification set S val And test set S test ;
Step ten, adding S train 、S val Together with its corresponding label as input data set C of the convolutional neural network input And initializing network weights, wherein S train For training the network coefficients, S val In training for a period of timeThen, network verification is carried out, and the weight of the network is adjusted through error back propagation;
step eleven, inputting a data set C input Performing convolution pooling operation for one time, and setting the scale kernel _ size1, convolution step kernel _ stride1, pooling size pool _ size1 and pooling step pool _ stride1 of a convolution kernel to obtain a feature map set feature1;
step twelve, further performing convolution pooling on the feature map set feature1, extracting deep features, and setting the size kernel _ size2, convolution step size kernel _ stride2, pooling size pool _ size2 and pooling step size pool _ stride2 of a convolution kernel to obtain a feature map set feature2;
thirteen, performing convolution pooling operation on the feature map set feature2 again to extract deeper features, and setting the size kernel _ size3, convolution step size kernel _ stride3, pooling size pool _ size3 and pooling step size pool _ stride3 of a convolution kernel to obtain a feature map set feature3;
step fourteen, sequentially passing the feature3 through full connection layers fc4, fc5 and fc6, respectively setting the sizes of fc4, fc5 and fc6 as size4, size5 and size6, and converting the feature map set into a column vector v1 of 1 × size 6;
step fifteen, outputting the column vector v1 into different gesture categories through a softmax classifier, and obtaining a trained convolutional neural network model NetModel through multiple iterations when the network accuracy and the loss function tend to be stable;
sixthly, testing the data set S test Loading the gesture classification result y into a NetModel;
the pretreatment method in the third step specifically comprises the following steps:
step 3-1, for low-frequency signals with large energy generated by coupling of the transmitting antenna and the receiving antenna, an improved wavelet threshold function is applied for removal, and the threshold function is as shown in formula 1:
wherein
W j,k (m,n)=mω j,k +(1-m)sgn(ω j,k )(|ω j,k |-n)
The improved threshold function is continuous at lambda and-lambda, and the values of alpha and beta are adjusted to enable the method to adapt to the requirements of various noise scenes;
step 3-2 aiming at the characteristics of the interference signal, the intermediate frequency signal of the gesture is subjected toAfter N layers of wavelet decomposition, approximate coefficients A (i) are obtained, detail coefficients D (1, i), D (2, i),. And D (N, i) are obtained, and only the low-frequency coefficients A (i) are subjected to threshold processing, namely
A i '=mA i +(1-m)sgn(A i )(|A i |-n)
2. The gesture recognition method based on 77GHz millimeter wave radar signals according to claim 1, characterized in that: the construction of the diversified characteristic spectrogram A in the fifth step specifically comprises the following steps:
step 5-1: extracting a time-distance spectrogram: firstly, FFT is carried out in a fast time domain, then weighting and averaging are carried out in a slow time domain, and finally interframe accumulation is carried out to obtain a time-distance spectrogram with the size of FFTNum1 x FrameNum;
step 5-2: extracting a time-velocity spectrogram: firstly, carrying out two-dimensional FFT by taking a frame as a unit to obtain a distance-Doppler diagram, searching for the maximum value of the distance-Doppler diagram, extracting a line where the maximum value is located, taking speed information of the line as the speed of a current frame signal, and finally carrying out accumulation between frames to obtain a time-speed spectrogram with the size of FFTNum2 x FrameNum;
step 5-3: extracting a time-angle spectrogram: firstly, taking a frame as a unit, respectively performing two-dimensional FFT (fast Fourier transform) on corresponding frame signals of N receiving antennas to obtain N distance-Doppler graphs, respectively extracting maximum values from the N distance-Doppler graphs, arranging the maximum values into 1 xN one-dimensional vectors, performing FFT (fast Fourier transform) on the one-dimensional vectors to obtain speed information corresponding to a current frame signal, and finally accumulating the speed information between frames to obtain a time-angle graph with the size of FFTNum3 x FrameNum;
step 5-4: constructing a diversified characteristic spectrogram A: the three feature spectrograms extracted by the method have the same column number, after the three spectrograms are respectively normalized, the normalized feature spectrograms are spliced by rows to obtain a diversified feature spectrogram, and the dimension of the diversified feature spectrogram is (FFTNum 1+ FFTNum2+ FFTNum 3) × FrameNum.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910445702.3A CN110348288B (en) | 2019-05-27 | 2019-05-27 | Gesture recognition method based on 77GHz millimeter wave radar signal |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910445702.3A CN110348288B (en) | 2019-05-27 | 2019-05-27 | Gesture recognition method based on 77GHz millimeter wave radar signal |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110348288A CN110348288A (en) | 2019-10-18 |
CN110348288B true CN110348288B (en) | 2023-04-07 |
Family
ID=68174720
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910445702.3A Active CN110348288B (en) | 2019-05-27 | 2019-05-27 | Gesture recognition method based on 77GHz millimeter wave radar signal |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110348288B (en) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110687816A (en) * | 2019-10-31 | 2020-01-14 | 复旦大学 | Intelligent household control system and method based on millimeter wave radar |
CN110988863A (en) * | 2019-12-20 | 2020-04-10 | 北京工业大学 | Novel millimeter wave radar gesture signal processing method |
CN113408328B (en) * | 2020-03-16 | 2023-06-23 | 哈尔滨工业大学(威海) | Gesture segmentation and recognition algorithm based on millimeter wave radar |
CN111461037B (en) * | 2020-04-07 | 2023-04-07 | 电子科技大学 | End-to-end gesture recognition method based on FMCW radar |
CN111929674B (en) * | 2020-07-10 | 2022-10-04 | 西安电子科技大学 | Intelligent amplitude comparison angle measurement method based on neural network, storage medium and equipment |
CN112183279B (en) * | 2020-09-21 | 2022-06-10 | 中国人民解放军国防科技大学 | Communication radiation source individual identification method based on IQ graph characteristics |
CN112148128B (en) * | 2020-10-16 | 2022-11-25 | 哈尔滨工业大学 | Real-time gesture recognition method and device and man-machine interaction system |
CN112446348B (en) * | 2020-12-08 | 2022-05-31 | 电子科技大学 | Behavior identification method based on characteristic spectrum flow |
CN112198966B (en) * | 2020-12-08 | 2021-03-16 | 中南大学 | Stroke identification method and system based on FMCW radar system |
CN113406610B (en) * | 2021-06-16 | 2023-06-23 | 深圳大学 | Target detection method, device, equipment and storage medium |
CN113567950B (en) * | 2021-09-24 | 2021-12-17 | 巍泰技术(武汉)有限公司 | Millimeter wave radar distance and velocity spectrum estimation method and system |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106055089A (en) * | 2016-04-27 | 2016-10-26 | 深圳市前海万象智慧科技有限公司 | Control system for gesture recognition based on man-machine interaction equipment and control method for same |
CN106295684A (en) * | 2016-08-02 | 2017-01-04 | 清华大学 | A kind of the most continuous based on micro-Doppler feature/discontinuous gesture recognition methods |
CN108509910A (en) * | 2018-04-02 | 2018-09-07 | 重庆邮电大学 | Deep learning gesture identification method based on fmcw radar signal |
CN109271838A (en) * | 2018-07-19 | 2019-01-25 | 重庆邮电大学 | A kind of three parameter attributes fusion gesture identification method based on fmcw radar |
-
2019
- 2019-05-27 CN CN201910445702.3A patent/CN110348288B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106055089A (en) * | 2016-04-27 | 2016-10-26 | 深圳市前海万象智慧科技有限公司 | Control system for gesture recognition based on man-machine interaction equipment and control method for same |
CN106295684A (en) * | 2016-08-02 | 2017-01-04 | 清华大学 | A kind of the most continuous based on micro-Doppler feature/discontinuous gesture recognition methods |
CN108509910A (en) * | 2018-04-02 | 2018-09-07 | 重庆邮电大学 | Deep learning gesture identification method based on fmcw radar signal |
CN109271838A (en) * | 2018-07-19 | 2019-01-25 | 重庆邮电大学 | A kind of three parameter attributes fusion gesture identification method based on fmcw radar |
Non-Patent Citations (2)
Title |
---|
Zhanfeng Zhao.Feature Extraction and Identification of Pipeline Intrusion Based on Phase-Sensitive Optical Time Domain Reflectometer.《Web of science》 .2019,全文. * |
张佳俊.基于5GHz频段雷达的手势识别系统研究.《CNKI》 .2018,全文. * |
Also Published As
Publication number | Publication date |
---|---|
CN110348288A (en) | 2019-10-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110348288B (en) | Gesture recognition method based on 77GHz millimeter wave radar signal | |
Wang et al. | TS-I3D based hand gesture recognition method with radar sensor | |
Liu et al. | Deep learning and recognition of radar jamming based on CNN | |
CN107024685A (en) | A kind of gesture identification method based on apart from velocity characteristic | |
CN112001306A (en) | Electroencephalogram signal decoding method for generating neural network based on deep convolution countermeasure | |
Zhao et al. | Cubelearn: End-to-end learning for human motion recognition from raw mmwave radar signals | |
CN111784560A (en) | SAR and optical image bidirectional translation method for generating countermeasure network based on cascade residual errors | |
Ni et al. | LPI radar waveform recognition based on multi-resolution deep feature fusion | |
CN103473559A (en) | SAR image change detection method based on NSCT domain synthetic kernels | |
CN113376600A (en) | Pedestrian radar echo denoising method based on RSDNet | |
CN113408328A (en) | Gesture segmentation and recognition algorithm based on millimeter wave radar | |
CN115343704A (en) | Gesture recognition method of FMCW millimeter wave radar based on multi-task learning | |
Li et al. | Image reflection removal using end‐to‐end convolutional neural network | |
CN113449587B (en) | Human behavior recognition and identity authentication method and device and electronic equipment | |
CN113111706B (en) | SAR target feature unwrapping and identifying method for azimuth continuous deletion | |
CN113901931A (en) | Knowledge distillation model-based behavior recognition method for infrared and visible light videos | |
CN114037891A (en) | High-resolution remote sensing image building extraction method and device based on U-shaped attention control network | |
CN112241001A (en) | Radar human body action recognition method and device, electronic equipment and storage medium | |
Yang et al. | Extraction and denoising of human signature on radio frequency spectrums | |
Yang et al. | Design of airborne target tracking accelerator based on KCF | |
CN115345216A (en) | FMCW radar interference elimination method fusing prior information | |
Zheng et al. | Unsupervised human contour extraction from through-wall radar images using dual UNet | |
CN114580468A (en) | Interference signal identification method based on time-frequency waterfall graph and convolutional neural network | |
CN113421281A (en) | Pedestrian micromotion part separation method based on segmentation theory | |
Wang et al. | Hand gesture recognition scheme based on millimeter-wave radar with convolutional neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |