CN114595729A - Communication signal modulation identification method based on residual error neural network and meta-learning fusion - Google Patents

Communication signal modulation identification method based on residual error neural network and meta-learning fusion Download PDF

Info

Publication number
CN114595729A
CN114595729A CN202210336623.0A CN202210336623A CN114595729A CN 114595729 A CN114595729 A CN 114595729A CN 202210336623 A CN202210336623 A CN 202210336623A CN 114595729 A CN114595729 A CN 114595729A
Authority
CN
China
Prior art keywords
neural network
residual
residual error
error neural
signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210336623.0A
Other languages
Chinese (zh)
Inventor
孙晓东
刘禹震
孙思瑶
于晓辉
李新波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jilin University
Original Assignee
Jilin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jilin University filed Critical Jilin University
Priority to CN202210336623.0A priority Critical patent/CN114595729A/en
Publication of CN114595729A publication Critical patent/CN114595729A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Digital Transmission Methods That Use Modulated Carrier Waves (AREA)

Abstract

The invention provides a communication signal modulation identification method based on residual error neural network and meta-learning fusion, belonging to the technical field of communication and comprising the following steps: acquiring a communication modulation signal; dividing subtasks; constructing a residual error neural network; extracting a sample; updating residual error neural network parameters; updating a meta-learning loss function; updating initial parameters of a residual error neural network; iteration is carried out; determining residual neural network parameters for the test signal; and inputting the final test communication signal into a residual error neural network to obtain a final classification result. The method integrates the residual error neural network with the meta-learning method, can effectively learn the good characteristics of the signals, solves the problem of weak generalization capability of a network model, and can complete the identification of the signals only by a small amount of samples.

Description

Communication signal modulation identification method based on residual error neural network and meta-learning fusion
Technical Field
The invention belongs to the technical field of communication, and particularly relates to a communication signal modulation identification method based on residual error neural network and meta-learning fusion.
Background
Wireless signal identification (WSR) has broad prospects in military and civilian applications, which may include signal reconnaissance and interception, interference resistance, and device identification. In general, WSR mainly includes Modulation Recognition (MR) and Wireless Technology Recognition (WTR). MR, also known as Automatic Modulation Classification (AMC), is widely used first in the military field and then extended to the civilian field. MR classifies radio signals by identifying modulation patterns, which helps in evaluating wireless transmission schemes and device types. Also, MR is able to extract digital baseband information even in situations where a priori information is limited.
Conventional modulation recognition algorithms can be mainly classified into two types: likelihood-based (LB) and feature-based (FB) methods. The LB method is based on a hypothesis testing theory; the performance based on decision theory is optimal, but the computational complexity is high. Therefore, feature-based methods have been developed as sub-optimal classifiers for practical applications. However, the feature-based method requires manual feature extraction, and is complex to implement. In recent years, deep learning and machine learning have been attracting attention for applications in the field of communications, because it is excellent in computer vision, machine translation, speech recognition, and natural language processing.
The machine learning algorithm in the communication field has the characteristics of no need of special feature extraction, flexible model, end-to-end learning and the like. However, when the communication signal is greatly affected by noise, the recognition rate of the communication signal may be seriously affected. When the network depth of the deep learning model is increased to improve the recognition effect, a 'degeneration' phenomenon often occurs, that is, the model accuracy rate is greatly reduced without signs along with the continuous increase of the network depth. How to ensure the accuracy of the model while ensuring the deepening of the network depth is a problem to be solved. Secondly, the machine learning model is a data-driven model, a large amount of data is needed to ensure the realization effect of the model, and when available data is scarce, the modern deep neural network often cannot achieve the desired effect. Therefore, efficient learning of features of data through a small number of samples is also an issue to be solved. Thirdly, the current machine learning model mostly adopts a random initialization method to initialize parameters, and how to initialize the parameters by using a certain priori knowledge, so that the model knows the characteristics of data, and the specific data is identified and distinguished in a targeted manner. Finally, current machine learning algorithms are greatly affected by data quality in the case of pure data driving, and are difficult to maintain robustness in the presence of imperfect data (e.g., missing or noisy values, outliers, etc.), which results in poor model generalization capability. Machine learning algorithms need to be generalized to improve their generalization capability to result in efficient implementation of the algorithms in practical applications.
Disclosure of Invention
The invention aims to overcome the defects that the prior art has low recognition rate under the condition that signals are influenced by noise, a network model is easy to degrade, the generalization capability of the network model is weak, the influence of data quantity is large and the like, and provides a communication signal modulation recognition method based on the fusion of a residual error neural network and meta-learning, wherein the special structure of the residual error neural network is used, so that the problem of degradation is effectively solved, and the signal recognition rate is effectively improved; in addition, the model can self-adapt the initialization parameters by using different noise schools by using a meta-learning method. By effectively initializing the network parameters, the model can be quickly adapted to the modulation identification of the communication signal. The method can effectively learn the good characteristics of the signals, solve the problem of weak generalization capability of the network model, and can complete the identification of the signals only by a small amount of samples.
The technical scheme adopted by the invention for realizing the purpose is as follows: a communication signal modulation identification method based on residual error neural network and meta-learning fusion comprises the following steps:
step 1, obtaining communication modulation signal
Acquiring various communication modulation signals with different signal-to-noise ratios under additive white Gaussian noise as a training data set, wherein the expression of the communication modulation signals with different signal-to-noise ratios under the additive white Gaussian noise is as follows:
Figure BDA0003576839230000021
wherein, s [ t ]]And r [ t ]]Respectively representing the transmitted and received signals of the t-th time period, gamma t]For communication channel gain, ω0To shift the carrier frequency, θ0Is the phase offset between the sender and receiver, v [ t ]]Is additive white gaussian noise, and j is an imaginary unit.
The expression for the signal-to-noise ratio is:
Figure BDA0003576839230000022
wherein, PrIs the power of the signal, PvIs the power of the noise.
Step 2, subtask division
Dividing the training data set into a plurality of subtask data according to the number of noise-to-noise ratios, identifying a signal under one signal-to-noise ratio as one subtask, wherein the number of the noise-to-noise ratios is N, the number of the subtasks is N, numbering the subtasks according to the magnitude of the signal-to-noise ratios, the greater the number of the signal-to-noise ratios is, the subtask data contains r communication signals, and the types of the communication signals contained in the subtask data are the same. The number of the communication signal types contained in each subtask data is the same as that of the communication modulation signal types used for final testing, and the number of samples in each subtask is the same.
Step 3, constructing a residual error neural network
Constructing a residual neural network structure for signal classification, wherein the residual neural network consists of 6 residual blocks, 2 full connection layers, 1 Dropout layer and 1 Softmax layer, and the full connection layers integrate the automatically extracted high-level features into local feature information;
the Softmax layer is used for classifying the multiple classes, and the expression is as follows:
Figure BDA0003576839230000031
wherein, ak'Is the k' th input signal of the output layer, yk'Is the output of the kth' neuron, C is the number of output nodes, i.e., the number of classes classified, acRepresenting the c-th input signal of the output layer.
Each residual block adopts a network structure of a residual block I or a residual block II, each residual block I and the residual block II consist of a convolution layer, two sub-residual blocks and a pooling layer, each sub-residual block comprises 2 convolution layers, all the convolution layers comprise 32 convolution kernels (including the convolution layers in the sub-residual blocks and the convolution layers outside the sub-residual blocks), the step length is 1, and the padding is 1; the sizes of convolution kernels of sub residual blocks of the first residual block and the second residual block are different, the size of the convolution kernel of the sub residual block of the first residual block is 3 multiplied by 2, and the size of the convolution kernel of the sub residual block of the second residual block is 3 multiplied by 1; the convolutional layer located outside the sub-residual block is activated using the Relu function; the output of each sub-residual block is also activated by the Relu function and maximally pooled, wherein the convolution layer performs convolution operation on the input to extract the characteristics of the signal, and the expression of the convolution operation is as follows:
xl′=f(wl′*xl′-1+bl′)
wherein l' represents the number of network layers (the network layers including input layers and convolutional layers), wl′Representing the convolution kernel parameters in the current convolution layer, bl′For the bias vector in the current convolutional layer, f (-) denotes the activation function, xl′-1Output, x, representing the last network layer (the network layer comprising the input layer and the convolutional layer)l′An output representing a current convolutional layer;
the pooling layer carries out down-sampling on the convolution layer, and the expression of the pooling layer is as follows:
xl′=down(xl′-1)
wherein down (·) represents a pooling function; x is the number ofl′-1Representing the last winding layerOutput, xl′An output representing a current pooling layer;
the Relu function expression is:
Figure BDA0003576839230000041
where f (x) is a function of the maximum value of Relu, and when the input is negative, the output is 0, and the neuron is not activated, x is the input to the function.
Step 4, extracting samples
And randomly extracting samples from the subtask data, and dividing the samples into a training set and a testing set as the input of the residual error neural network. The number of samples randomly extracted from the subtask data at each time is h + m, the samples include r communication signals, wherein h samples are used as a training set for extracting the samples, and m samples are used as a test set for extracting the samples.
Input data of the residual error neural network is nin×l×dim×1,ninThe number of input samples of the residual error neural network is, and l is the signal sampling length. The signal is in complex form, taking the real and imaginary parts of the signal superimposed together, so dim is 2.
Step 5, updating residual error neural network parameters
And initializing initial parameters of the residual error neural network randomly, inputting the extracted sample training set into the residual error neural network, and quickly iterating to update the parameters of the residual error neural network.
The step of updating the parameters of the residual error neural network comprises the following steps:
1) random initialization residual neural network parameter phi0
2) Inputting the extracted sample training set into a residual error neural network;
3) defining a cross entropy loss function l (phi) by using the extracted sample label and a network output result;
the function is gradiented and the parameters are updated,
Figure BDA0003576839230000042
wherein k is the number of samples takenThe number of times, alpha, is the inner loop learning rate, phikFor the residual neural network parameter at the kth sample draw, l (φ)k) As a cross-entropy loss function of the network parameters,
Figure BDA0003576839230000051
and the residual error neural network parameters after gradient updating.
The cross entropy loss function is defined as:
l(φ)=[y(k)logg(x(k))+(1-y(k))log(1-g(x(k)))]
wherein x is(k)For the data of the k-th training set of samples, y(k)For the kth sample of the data of the training set, g represents the residual neural network structure, and l (φ) is the cross entropy loss function.
The iterative update parameters are to ensure that the residual neural network performs well on the extracted data training set. The conditions for judging whether the performance is good are as follows: and when the cross entropy loss function l (phi) is reduced for a period of time, continuing iteration, and keeping l (phi) stable, so that the residual error neural network is proved to have good performance on the extracted data training set.
Step 6, updating the meta learning loss function
And inputting the extracted sample test set into the updated network to obtain output data corresponding to the test set. Updating the expression of the meta-learning loss function according to the label corresponding to the test set and the output data corresponding to the test set as follows:
Figure BDA0003576839230000052
wherein,
Figure BDA0003576839230000053
for the kth sample test set data draw,
Figure BDA0003576839230000054
for the kth sample test set data label,
Figure BDA0003576839230000055
for the updated meta learning loss function after k-1 samples are taken,
Figure BDA0003576839230000056
step 7, updating the initial parameters of the residual error neural network
And solving a gradient of the meta-learning loss function, and updating initial parameters of the residual neural network by using the gradient. The expression for updating the initial parameters of the residual neural network by using the gradient is as follows:
Figure BDA0003576839230000057
where β is the outer loop learning rate. The inner loop learning rate α is larger than the outer loop learning rate β.
Figure BDA0003576839230000058
And initial parameters of the updated residual error neural network.
Step 8, iteration
Repeating the steps 4 to 7 for a plurality of times. When the step 5 is repeated, the parameters of the residual error neural network are not initialized randomly any more, but are initialized to the network parameters updated by the last iteration
Figure BDA0003576839230000059
Step 9, determining residual error neural network parameters for testing signals
And (5) inputting the communication modulation signal with the same property as the final test signal into the residual error neural network, repeating the step 5, and adjusting the initial parameters of the residual error neural network. When the step 5 is repeated, the parameters of the residual error neural network are not initialized randomly any more, but are initialized to the network parameters updated by the last iteration
Figure BDA0003576839230000061
Step 10, outputting the result
And inputting the final test communication signal into a residual error neural network to obtain a final classification result.
Through the design scheme, the invention can bring the following beneficial effects:
1. the invention provides a communication signal modulation recognition method based on a residual error neural network, which learns the difference value between a target value and the input of a residual error block, namely a residual error, by utilizing a special structure (shortcut connection mode) of the residual error block, so that the accuracy rate is not reduced any more as the network is deepened, the problem of 'degradation' when the network depth is increased can be solved, and the recognition accuracy rate of communication signal modulation is improved.
2. The invention also provides a communication signal modulation identification method based on meta-learning, which is characterized in that the priori knowledge of the communication signal is utilized to learn and initialize the initial parameters of the neural network, the initial parameters of the neural network contain the high-quality characteristics of the signal after learning, and the effective characteristics of the signal can be learned without a large amount of data and only a few samples.
3. The communication signal modulation identification method based on meta-learning provided by the invention utilizes the signal-to-noise ratio and the priori knowledge of the communication signals, and learns to initialize the communication signal classification model by extracting the signals under different signal-to-noise ratios, so that the model can adapt to the signals under different signal-to-noise ratios, can adapt to different communication signal identifications under different noises, and improves the generalization capability of the classification model.
Drawings
FIG. 1 is a block diagram of a communication signal modulation recognition method based on the fusion of residual neural network and meta-learning according to the present invention;
FIG. 2 is a signal time domain waveform diagram of 4ASK signal-5 dB in the embodiment of the communication signal modulation identification method based on residual error neural network and meta-learning fusion of the present invention;
FIG. 3 is a time domain waveform diagram of a 4FSK signal of-5 dB according to an embodiment of a communication signal modulation identification method based on a residual error neural network and meta-learning fusion in the present invention;
FIG. 4 is a signal time domain waveform diagram of-5 dB of a 4PSK signal in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion in the present invention;
FIG. 5 is a signal time domain waveform diagram of 4QAM signal-5 dB in an embodiment of the communication signal modulation identification method based on residual error neural network and meta-learning fusion of the present invention;
FIG. 6 is a signal time domain waveform diagram of 0dB of a 4ASK signal in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion in the present invention;
FIG. 7 is a time domain waveform diagram of a 4 dB FSK signal according to an embodiment of the present invention based on a communication signal modulation and identification method with a residual error neural network and meta-learning fusion;
FIG. 8 is a signal time domain waveform diagram of 0dB of a 4PSK signal in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion in the present invention;
FIG. 9 is a signal time domain waveform diagram of 0dB of a 4QAM signal in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion in the present invention;
FIG. 10 is a signal time domain waveform diagram of 5dB of a 4ASK signal in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion according to the present invention;
fig. 11 is a signal time domain waveform diagram of a 4FSK signal of 5dB in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion according to the present invention;
fig. 12 is a signal time domain waveform diagram of 5dB of a 4PSK signal in an embodiment of a communication signal modulation identification method based on residual error neural network and meta-learning fusion according to the present invention;
fig. 13 is a signal time domain waveform diagram of 5dB of a 4QAM signal in an embodiment of the method for communication signal modulation identification based on residual neural network and meta-learning fusion in accordance with the present invention;
FIG. 14 is a flowchart illustrating decomposition of a training subtask in an embodiment of a communication signal modulation recognition method based on fusion of a residual neural network and meta-learning according to the present invention;
FIG. 15 is a diagram illustrating a residual neural network structure in an embodiment of a communication signal modulation recognition method based on a fusion of a residual neural network and meta-learning according to the present invention;
FIG. 16 is a block diagram of a residual block in an embodiment of a communication signal modulation recognition method based on a fusion of a residual neural network and meta-learning according to the present invention;
FIG. 17 is a diagram of a second residual block structure in an embodiment of a communication signal modulation recognition method based on the fusion of a residual neural network and meta-learning according to the present invention;
FIG. 18 is a flowchart illustrating updating parameters of a residual neural network in an embodiment of a communication signal modulation recognition method based on fusion of the residual neural network and meta-learning according to the present invention;
FIG. 19 is a flowchart illustrating updating of initial parameters of a residual neural network in an embodiment of a communication signal modulation identification method based on fusion of a residual neural network and meta-learning according to the present invention;
fig. 20 is a flowchart of a test task in an embodiment of the communication signal modulation identification method based on the fusion of the residual neural network and the meta-learning according to the present invention.
Detailed Description
In order to more clearly illustrate the invention, the invention is further described below with reference to preferred embodiments and the accompanying drawings. It is to be understood by persons skilled in the art that the following detailed description is illustrative and not restrictive, and is not to be taken as limiting the scope of the invention. Unless defined otherwise, technical or scientific terms used herein shall have the ordinary meaning as understood by one of ordinary skill in the art to which this invention belongs.
As shown in fig. 1, the present invention relates to a communication signal modulation identification method based on residual error neural network and meta-learning fusion, firstly, acquiring communication modulation signals with different signal-to-noise ratios under additive white gaussian noise as training objects, dividing the whole data set into a plurality of subtask data sets according to different signal-to-noise ratios of the signals, wherein the identification of the signal under one signal-to-noise ratio corresponds to one subtask, and numbering the subtasks accordingly; secondly, constructing a residual error neural network for signal classification; then, by utilizing a meta-learning theory, obtaining initial parameters of a residual error neural network through learning: taking random sampling data in each subtask data as input of a residual error neural network, and sampling each subtask until all subtasks are used, so that the network learns how to initialize parameters by using the characteristics of signals; thirdly, acquiring a part of communication modulation signals of known labels, which have the same properties as the final test signal, inputting the communication modulation signals into the residual error neural network for training, and completing parameter updating of the residual error neural network; and finally, acquiring a communication modulation signal used for final testing, and inputting the communication modulation signal into the residual error neural network to complete a communication signal modulation identification task. The invention integrates the residual error neural network and the meta-learning method, avoids the phenomenon of 'degeneration' in the training process, improves the generalization capability of the neural network and improves the accuracy of modulation recognition.
The specific implementation process is as follows:
firstly, a training task and test task data set is generated by utilizing Matlab software simulation, the data set comprises 4ASK, 4FSK, 4PSK and 4QAM signals, the signal-to-noise ratio range is-5 dB to 5dB, noise is additive white Gaussian noise, and the expression of communication modulation signals under the additive white Gaussian noise with different signal-to-noise ratios is as follows:
Figure BDA0003576839230000081
wherein, s [ t ]]And r [ t ]]Respectively representing the transmitted and received signals of the t-th time period, gamma t]For communication channel gain, omega0To shift the carrier frequency, θ0Is the phase offset between the sender and receiver, v [ t ]]Is additive white gaussian noise, and j is an imaginary unit.
The expression for the signal-to-noise ratio is:
Figure BDA0003576839230000091
wherein, PrIs the power of the signal, PvIs the power of the noise. Fig. 2-5 are graphs of simulated-5 dB time domain waveforms, fig. 6-9 are graphs of simulated 0dB time domain waveforms, and fig. 10-13 are graphs of simulated 5dB time domain waveforms.
And dividing the training task into 11 subtasks according to the number of the noise signal-to-noise ratios, wherein each subtask comprises four generated communication signals. An exploded view of the training subtask is shown in FIG. 14.
Secondly, on the premise that the neural network can be converged, as the depth of the network increases, the performance of the network gradually increases to saturation and then rapidly decreases. This is a "degradation" phenomenon, and this problem makes it difficult to improve the recognition rate of classification problems. And the residual error neural network utilizes the special structure thereof, so that when the neural network propagates forwards, the input signal can directly propagate from any lower layer to a higher layer. The network degradation problem can be solved to a certain extent due to the fact that the network degradation problem is contained in the network degradation mapping method.
Constructing a residual neural network structure for signal classification, wherein the residual neural network consists of 6 residual blocks, 2 full connection layers, 1 Dropout layer and 1 Softmax layer, and the full connection layers integrate the automatically extracted high-level features into local feature information;
the Softmax layer is used for classifying the multiple classes, and the expression is as follows:
Figure BDA0003576839230000092
wherein, ak'Is the k' th input signal of the output layer, yk'Is the output of the kth' neuron, C is the number of output nodes, i.e., the number of classes classified, acRepresenting the c-th input signal of the output layer.
Each residual block adopts a network structure of a residual block I or a residual block II, and as shown in fig. 15, 6 residual blocks respectively adopt a residual module I, a residual module II and a residual module II in sequence; the first residual block and the second residual block are composed of a convolution layer, two sub-residual blocks and a pooling layer, each sub-residual block comprises 2 convolution layers, all convolution layers comprise 32 convolution kernels, the step length is 1, and the padding is 1; the sizes of convolution kernels of sub residual blocks of the first residual block and the second residual block are different, the size of the convolution kernel of the sub residual block of the first residual block is 3 multiplied by 2, and the size of the convolution kernel of the sub residual block of the second residual block is 3 multiplied by 1; the convolutional layer located outside the sub-residual block is activated using the Relu function; the output of each sub-residual block is also activated by the Relu function and maximally pooled, wherein the convolution layer performs convolution operation on the input to extract the characteristics of the signal, and the expression of the convolution operation is as follows:
xl′=f(wl′*xl′-1+bl′)
wherein l' represents the number of network layers (the network layers including input layers and convolutional layers), wl′Representing the convolution kernel parameters in the current convolution layer, bl′For the bias vector in the current convolutional layer, f (-) denotes the activation function, xl′-1Output, x, representing the last network layer (the network layer comprising the input layer and the convolutional layer)l′Representing the output of the current convolutional layer.
The pooling layer carries out down-sampling on the convolution layer, and the expression of the pooling layer is as follows:
xl′=down(xl′-1)
where down (·) represents the pooling function. x is the number ofl′-1Representing the output, x, of the previous convolution layerl′Representing the output of the current pooling layer.
The Relu function expression is:
Figure BDA0003576839230000101
where f (x) is a function of the maximum value of Relu, and when the input is negative, the output is 0, and the neuron is not activated, x is the input to the function.
The structure of the constructed residual neural network is shown in fig. 15. The two residual block structures in the residual neural network are shown in fig. 16 and 17.
And thirdly, randomly extracting samples from the subtasks and dividing the samples into a training set and a testing set to be used as input of the residual error neural network.
And initializing initial parameters of the residual error neural network randomly, inputting the extracted sample training set into the residual error neural network, and quickly iterating to update the parameters of the residual error neural network.
The step of updating the residual error neural network parameters comprises the following steps:
1) random initialization residual neural network parameter phi0
2) Inputting the extracted sample training set into a residual error neural network;
3) defining a cross entropy loss function l (phi) by using the extracted sample label and a network output result;
the function is gradiented and the parameters are updated,
Figure BDA0003576839230000111
where k is the number of times the sample is taken, α is the inner loop learning rate, φkFor the residual neural network parameter at the kth sample draw, l (φ)k) As a cross-entropy loss function of the network parameters,
Figure BDA0003576839230000112
and the residual error neural network parameters after gradient updating.
The cross entropy loss function is defined as:
l(φ)=[y(k)logg(x(k))+(1-y(k))log(1-g(x(k)))]
wherein x is(k)For the data of the k-th training set of samples, y(k)The label of the data of the training set of samples is extracted for the first time k, and g represents the residual neural network structure.
The iterative update parameters are to ensure that the residual neural network performs well on the extracted data training set. The conditions for judging whether the performance is good are as follows: when cross entropy loss function
Figure BDA0003576839230000113
After a period of time of falling, the iteration is continued,
Figure BDA0003576839230000114
and keeping the stability, and proving that the residual error neural network has good performance on the extracted data training set at the moment.
A flow chart of the residual neural network updating parameters is shown in fig. 18.
And fourthly, inputting the extracted sample test set into the updated residual error neural network to obtain output data corresponding to the test set. Updating the expression of the meta-learning loss function according to the label corresponding to the test set and the output data corresponding to the test set as follows:
Figure BDA0003576839230000115
wherein,
Figure BDA0003576839230000116
for the kth sample test set data,
Figure BDA0003576839230000117
for the kth sample test set data label,
Figure BDA0003576839230000118
for the updated meta learning loss function after k-1 samples are taken,
Figure BDA0003576839230000119
and solving a gradient of the meta-learning loss function, and updating initial parameters of the residual neural network by using the gradient. The expression for updating the initial parameters of the residual neural network by using the gradient is as follows:
Figure BDA00035768392300001110
wherein, beta is the outer ring learning rate, and the inner ring learning rate alpha is larger than the outer ring learning rate beta;
Figure BDA00035768392300001111
and initial parameters of the updated residual error neural network.
And randomly taking samples for multiple times, and repeating the steps. A flowchart for updating the initial parameters of the residual neural network is shown in fig. 19. And extracting sample data of the subtasks, and dividing the data into a subtask training set and a subtask test set. And inputting the subtask training set into a residual error neural network, and updating parameters of the residual error neural network. And inputting the subtask test set into the residual error neural network after updating the parameters, updating a meta-learning loss function by utilizing the subtask test set label and the output of the subtask test set input into the residual error neural network, and updating the initial parameters of the residual error neural network by calculating the gradient of the meta-learning loss function. And (4) finishing the initialization of the residual error neural network parameters by extracting the data of the subtasks for multiple times. The initial parameters of the residual error neural network are updated, so that the residual error neural network can adapt to signals under different signal-to-noise ratios, effectively learn key characteristics of the signals, and inhibit noise interference to a certain extent. When a new task occurs, the residual neural network can adjust its parameters according to the characteristics of the new task, so that it performs well on the task.
The test flow chart is shown in fig. 20. Firstly, inputting a signal with the same property as the test signal into the residual error neural network, and updating the parameters of the residual error neural network so that the residual error neural network is suitable for the test data. And finally, inputting the test communication signal into the residual error neural network with updated parameters to obtain a final classification result.

Claims (10)

1. The communication signal modulation identification method based on the fusion of the residual error neural network and the meta-learning is characterized by comprising the following steps of:
step 1: acquiring various communication modulation signals with different signal-to-noise ratios under additive white Gaussian noise as a training data set;
step 2: dividing the training data set into N subtask data according to the number of noise-to-noise ratios, wherein the identification of a signal under one signal-to-noise ratio corresponds to one subtask, and the subtasks are numbered according to the magnitude of the signal-to-noise ratio, and each subtask comprises a plurality of communication signals if the number of the signal-to-noise ratio is larger;
and 3, step 3: constructing a residual error neural network structure for signal classification;
and 4, step 4: randomly extracting a small amount of samples which are lower than 10% of subtask data from the lowest-number subtask data in the currently unused subtasks, and dividing the samples into a training set and a test set to be used as input of a residual error neural network;
and 5: initializing initial parameters of a residual error neural network randomly, inputting an extracted sample training set into the residual error neural network, quickly iterating, and updating parameters of the residual error neural network;
step 6: inputting the extracted sample test set into the updated residual error neural network, and updating a meta-learning loss function according to the test set label and the actual output;
and 7: solving a gradient of the meta-learning loss function, and updating the residual neural network parameters by using the gradient;
and 8: repeating the steps 4 to 7 for multiple times until the extraction of the N subtask data is completed;
and step 9: inputting the communication modulation signal with the same property as the final test signal into the residual error neural network, repeating the step 5, and finishing the parameter updating of the residual error neural network;
step 10: and inputting the final test signal into a residual error neural network to obtain a final classification result.
2. The communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 1, characterized in that: in step 1, the expression of the communication modulation signals with different signal-to-noise ratios under additive white gaussian noise is as follows:
Figure FDA0003576839220000011
wherein, s [ t ]]And r [ t ]]Respectively representing the transmitted and received signals of the t-th time period, gamma t]For communication channel gain, omega0To shift the carrier frequency, θ0Is the phase offset between the sender and receiver, v [ t ]]Is additive white Gaussian noise, and j is an imaginary number unit;
the expression for the signal-to-noise ratio is:
Figure FDA0003576839220000021
wherein, PrIs the power of the signal, PvIs the power of the noise.
3. The communication signal modulation identification method based on the fusion of the residual neural network and the meta-learning according to claim 1, characterized in that: in step 2, the number of the noise-to-noise ratios is N, the number of the subtasks is N, each subtask data includes r types of communication signals, each subtask data includes the same type of communication signal, the number of the types of communication signals included in each subtask is the same as the number of the types of communication modulation signals used for final testing, and the number of samples in each subtask is the same.
4. The communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 1, characterized in that: in step 3, the residual neural network consists of 6 residual blocks, 2 full-link layers, 1 Dropout layer and 1 Softmax layer, and the full-link layers integrate the automatically extracted high-level features into local feature information;
the Softmax layer is used for classifying the multiple classes, and the expression is as follows:
Figure FDA0003576839220000022
wherein, ak'Is the k' th input signal of the output layer, yk'Is the output of the kth' neuron, C is the number of output nodes, i.e., the number of classes classified, acRepresenting the c-th input signal of the output layer.
5. The communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 4, characterized in that: each residual block adopts a network structure of a residual block I or a residual block II, the residual block I and the residual block II consist of a convolution layer, two sub-residual blocks and a pooling layer, each sub-residual block comprises 2 convolution layers, all the convolution layers comprise 32 convolution kernels, the step length is 1, and the padding is 1; the sizes of convolution kernels of sub residual blocks of the first residual block and the second residual block are different, the size of the convolution kernel of the sub residual block of the first residual block is 3 multiplied by 2, and the size of the convolution kernel of the sub residual block of the second residual block is 3 multiplied by 1; the convolutional layer located outside the sub-residual block is activated using the Relu function; the output of each sub-residual block is also activated by the Relu function and maximally pooled, wherein the convolution layer performs convolution operation on the input to extract the characteristics of the signal, and the expression of the convolution operation is as follows:
xl′=f(wl′*xl′-1+bl′)
wherein l' represents the number of network layers including input layers and convolutional layers, wl′Representing the convolution kernel parameters in the current convolution layer, bl′For the bias vector in the current convolutional layer, f (-) denotes the activation function, xl′-1Representing the output of the last network layer, xl′An output representing a current convolutional layer;
the pooling layer carries out down-sampling on the convolution layer, and the expression of the pooling layer is as follows:
xl′=down(xl′-1)
wherein down (·) represents a pooling function; x is the number ofl′-1Representing the output, x, of the previous convolution layerl′An output representing a current pooling layer;
relu function expression is:
Figure FDA0003576839220000031
where f (x) is a function of the maximum value of Relu, and when the input is negative, the output is 0, and the neuron is not activated, x is the input to the function.
6. The residual neural network and meta-learning fusion based communication signal modulation identification method of claim 1, whichIs characterized in that: in step 4, the number of samples randomly extracted from the subtask data each time is h + m, the samples include r communication signals, wherein h samples are used as a training set of the extracted samples, and m samples are used as a test set of the extracted samples; input data of the residual error neural network is nin×l×dim×1,ninThe number of input samples of the residual neural network is l, which is the signal sampling length, the signal is in a complex form, and the real part and the imaginary part of the signal are added together, so dim is 2.
7. The communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 1, characterized in that: in step 5, the step of updating the residual error neural network parameters is as follows:
1) random initialization residual neural network parameter phi0
2) Inputting the extracted sample training set into a residual error neural network;
3) defining a cross entropy loss function l (phi) by using the extracted sample label and a network output result;
the function is gradiented and the parameters are updated,
Figure FDA0003576839220000032
wherein k is the number of times of sample extraction, alpha is the inner loop learning rate, phikFor the residual neural network parameter at the kth sample draw, l (φ)k) As a cross-entropy loss function of the network parameters,
Figure FDA0003576839220000033
the residual error neural network parameters after gradient updating are obtained;
the cross entropy loss function is defined as:
l(φ)=[y(k)logg(x(k))+(1-y(k))log(1-g(x(k)))]
wherein x is(k)For the data of the k-th training set of samples, y(k)For the kth sample training set data label, g represents the residual neural network structure, l (φ) is the cross entropy lossA function.
8. The communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 1, characterized in that: in step 6, updating the expression of the meta-learning loss function according to the label corresponding to the test set and the output data corresponding to the test set as follows:
Figure FDA0003576839220000041
wherein,
Figure FDA0003576839220000042
for the kth sample test set data draw,
Figure FDA0003576839220000043
for the kth sample test set data label,
Figure FDA0003576839220000044
for the updated meta learning loss function after k-1 samples are taken,
Figure FDA0003576839220000045
9. the communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 1, characterized in that: step 7, solving a gradient of the meta-learning loss function, and updating initial parameters of the residual error neural network by using the gradient; the expression for updating the initial parameters of the residual neural network by using the gradient is as follows:
Figure FDA0003576839220000046
wherein, beta is the outer ring learning rate, and the inner ring learning rate alpha is larger than the outer ring learning rate beta;
Figure FDA0003576839220000047
and initial parameters of the updated residual error neural network.
10. The communication signal modulation identification method based on residual error neural network and meta-learning fusion of claim 1, characterized in that: in step 8, when step 5 is repeated, the parameters of the residual error neural network are not initialized randomly any more, but are initialized to the initial parameters of the network updated by the last iteration
Figure FDA0003576839220000048
In step 9, when step 5 is repeated, the parameters of the residual error neural network are not initialized randomly any more, but are initialized to the initial parameters of the network updated by the last iteration
Figure FDA0003576839220000049
CN202210336623.0A 2022-04-01 2022-04-01 Communication signal modulation identification method based on residual error neural network and meta-learning fusion Pending CN114595729A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210336623.0A CN114595729A (en) 2022-04-01 2022-04-01 Communication signal modulation identification method based on residual error neural network and meta-learning fusion

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210336623.0A CN114595729A (en) 2022-04-01 2022-04-01 Communication signal modulation identification method based on residual error neural network and meta-learning fusion

Publications (1)

Publication Number Publication Date
CN114595729A true CN114595729A (en) 2022-06-07

Family

ID=81811858

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210336623.0A Pending CN114595729A (en) 2022-04-01 2022-04-01 Communication signal modulation identification method based on residual error neural network and meta-learning fusion

Country Status (1)

Country Link
CN (1) CN114595729A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115994303A (en) * 2023-03-24 2023-04-21 中国人民解放军军事科学院国防科技创新研究院 Residual neural network model and signal modulation recognition method thereof
CN117131416A (en) * 2023-08-21 2023-11-28 四川轻化工大学 Small sample modulation identification method, system, electronic equipment and storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115994303A (en) * 2023-03-24 2023-04-21 中国人民解放军军事科学院国防科技创新研究院 Residual neural network model and signal modulation recognition method thereof
CN117131416A (en) * 2023-08-21 2023-11-28 四川轻化工大学 Small sample modulation identification method, system, electronic equipment and storage medium
CN117131416B (en) * 2023-08-21 2024-06-04 四川轻化工大学 Small sample modulation identification method, system, electronic equipment and storage medium

Similar Documents

Publication Publication Date Title
CN112418014B (en) Modulated signal identification method based on wavelet transformation and convolution long-term and short-term memory neural network
CN111314257B (en) Modulation mode identification method based on complex value neural network
CN110120926B (en) Communication signal modulation mode identification method based on evolution BP neural network
Hu et al. Robust modulation classification under uncertain noise condition using recurrent neural network
Chen et al. Automatic modulation classification scheme based on LSTM with random erasing and attention mechanism
CN114595729A (en) Communication signal modulation identification method based on residual error neural network and meta-learning fusion
Zhang et al. A data preprocessing method for automatic modulation classification based on CNN
CN110113288B (en) Design and demodulation method of OFDM demodulator based on machine learning
CN112039820B (en) Communication signal modulation and identification method for quantum image group mechanism evolution BP neural network
Zhang et al. Automatic modulation classification using a deep multi-stream neural network
CN114881092A (en) Signal modulation identification method based on feature fusion
CN110659684A (en) Convolutional neural network-based STBC signal identification method
CN114726692B (en) SERESESESENet-LSTM-based radiation source modulation mode identification method
Zhang et al. Lightweight automatic modulation classification via progressive differentiable architecture search
CN110233810A (en) Msk signal demodulation method based on deep learning under a kind of mixed noise
CN114615118B (en) Modulation identification method based on multi-terminal convolution neural network
Sang et al. Application of novel architectures for modulation recognition
CN113298031A (en) Signal modulation identification method considering signal physical and time sequence characteristics and application
Ma et al. Modulation classification method based on deep learning under non-Gaussian noise
CN113902095A (en) Automatic modulation identification method, device and system for wireless communication
CN114070684B (en) Modulation signal identification method based on two-dimensional convolutional neural network
CN113239788A (en) Mask R-CNN-based wireless communication modulation mode identification method
Kalade et al. Using sequence to sequence learning for digital bpsk and qpsk demodulation
CN117376066A (en) Asynchronous sampling-oriented digital signal modulation mode identification method
CN112348165A (en) Underwater acoustic communication signal classification and identification method and system based on hybrid cycle network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination