CN116016071A - Modulation signal identification method based on double-flow fusion CNN-BiLSTM network - Google Patents
Modulation signal identification method based on double-flow fusion CNN-BiLSTM network Download PDFInfo
- Publication number
- CN116016071A CN116016071A CN202211579515.2A CN202211579515A CN116016071A CN 116016071 A CN116016071 A CN 116016071A CN 202211579515 A CN202211579515 A CN 202211579515A CN 116016071 A CN116016071 A CN 116016071A
- Authority
- CN
- China
- Prior art keywords
- network
- cnn
- bilstm
- stream
- fusion
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 44
- 230000004927 fusion Effects 0.000 title claims abstract description 40
- 238000012549 training Methods 0.000 claims abstract description 19
- 230000003993 interaction Effects 0.000 claims abstract description 9
- 230000006870 function Effects 0.000 claims description 16
- 239000013598 vector Substances 0.000 claims description 16
- 238000007781 pre-processing Methods 0.000 claims description 11
- 238000012360 testing method Methods 0.000 claims description 8
- 238000005070 sampling Methods 0.000 claims description 7
- 230000008569 process Effects 0.000 claims description 6
- 230000004913 activation Effects 0.000 claims description 5
- 239000000284 extract Substances 0.000 claims description 5
- 230000017105 transposition Effects 0.000 claims description 5
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 4
- 230000015572 biosynthetic process Effects 0.000 claims description 4
- 230000002708 enhancing effect Effects 0.000 claims description 4
- 238000003786 synthesis reaction Methods 0.000 claims description 4
- 230000009977 dual effect Effects 0.000 claims description 3
- 238000013527 convolutional neural network Methods 0.000 description 11
- 238000004891 communication Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 4
- 238000000605 extraction Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D30/00—Reducing energy consumption in communication networks
- Y02D30/70—Reducing energy consumption in communication networks in wireless communication networks
Landscapes
- Digital Transmission Methods That Use Modulated Carrier Waves (AREA)
Abstract
A modulating signal identification method based on double-flow fusion CNN-BiLSTM is realized by using the existing public data set radio ML2016.10a, firstly, the data set is preprocessed, the data is converted into an I/Q format and an A/P format, and the information diversity is increased; in order to extract the space-time characteristics of signals at the same time, each stream consists of CNN and BiLSTM, the BiLSTM network layer carries out forward and backward learning, the recognition performance is improved, the network carries out double-stream fusion, the feature interaction is effectively carried out, and more effective classification features are obtained. The invention can greatly reduce training time and improve accuracy of signal identification performance.
Description
Technical Field
The invention relates to signal modulation identification classification, in particular to a modulation signal identification method based on a double-flow fusion CNN-BiLSTM network.
Background
With the rapid development of communication technology, electronic technology and signal processing technology, new modulation modes of communication signals are continuously emerging, the types of modulation signals are more and more, and the communication environment is more and more complex. The communication signal modulation identification classification refers to that the modulation type of a signal is judged after the received signal without priori knowledge is analyzed and processed, and the communication signal modulation identification classification is used as an important link between signal detection and demodulation in the civil and military fields such as frequency spectrum resource management, illegal radio station interference, electronic countermeasures and the like.
The debug mode identification of the conventional communication signal is largely classified into a Likelihood (LB) based method and a Feature (FB) based method. The communication signal modulation mode identification based on the likelihood method is equivalent to the multiple composite hypothesis test problem, different hypotheses are made for different modulation modes, likelihood functions of received signals under different hypotheses are calculated, and then the values of the likelihood functions are compared with a predefined threshold value to obtain possible categories of the received signals. Although likelihood-based methods are optimal in a bayesian sense, they are often accompanied by considerable computational complexity or require complete prior knowledge and sensitivity to unknown channel condition scenarios.
The feature-based method mainly comprises the steps of extracting feature parameters corresponding to a modulation signal by using a machine learning or deep learning algorithm, and selecting a classification criterion meeting requirements based on the difference between the feature parameters so as to make a signal type judgment. The method is generally divided into three parts, namely signal preprocessing, feature extraction and classification recognition. The actual process of signal preprocessing is mainly determined by the characteristic parameters of the usage data. Feature extraction is the extraction of intrinsic attribute information of the dataset signal. The classification identifies, i.e. uses the extracted feature information to make decisions on the signal type, such as wavelet transforms, cyclostationary features and higher order cumulants, etc.
The FB method has proven to provide a suboptimal solution compared to the LB method, which can use higher quality functions to obtain strong performance with less complexity, extracting some appropriate features from the received signal to distinguish the modulation types. However, FB methods generally rely on the extraction of manual features that require a great deal of domain knowledge and engineering techniques, and neural network-based methods are relatively easier to operate. However, the neural network method is too dependent on data, and the time complexity of training is increased due to too large data volume, so that various characteristics of signals are required to be utilized to improve accuracy. Therefore, in order to increase the diversity of the features, the chapter adopts a double-flow network structure, the first flow inputs I/Q data, the second flow inputs A/P data, a CNN network is used for extracting the spatial characteristics of signals, and a Bi-LSTM network is used for bidirectionally extracting the time characteristics of the signals to judge the modulated signals.
Disclosure of Invention
In order to overcome the defect of low accuracy of the existing low signal-to-noise ratio signal identification, the invention provides a modulation signal identification method based on double-flow fusion BiLSTM, which is more stable and has higher accuracy than the traditional network structure, and can effectively classify the modulation signals with low signal-to-noise ratio.
The technical scheme adopted by the invention is as follows:
a modulation signal identification method based on a double-flow fusion CNN-BiLSTM network comprises the following steps:
step 1: preparing a data set;
step 2: data preprocessing, converting signals into a time inphase/quadrature (I/Q) format and an amplitude/phase (a/P) representation, forming two data sets;
step 3: building a double-flow fusion CNN-BiLSTM network structure and setting network basic parameters;
step 4: the training model is used for inputting the training set into a double-flow fusion CNN-BiLSTM network for training;
step 5: inputting the test set into a trained double-flow fusion CNN-BiLSTM network structure, automatically identifying the type of the modulation signal, and outputting the accuracy of signal identification.
Further, in the step 1, the data set is a publicly available data set Radio ml2016.10a generated using GNU Radio synthesis, including training data and test data, and the types thereof include BPSK, QPSK, 8PSK, 16QAM, 64QAM, BFSK, CPFSK and PAM48 digital modulations and WB-FM, AM-SSB and AM-DSB3 analog modulation signals.
Still further, in the step 2, the data preprocessing process is as follows:
(2.1) the modulated signal is an I/Q signal consisting of two sets of sequences, I and Q, which are converted into I/Q vectors;
j represents the number of data set samples, the maximum value is 220000, N represents the number of sampling points, and the maximum value is 128;
(2.2) calculating amplitude and phase vectors:
(2.3) constructing an AP dataset consisting of amplitude components and phase components and converting to a/P vectors:
j represents dataThe number of samples is set to 220000, n represents the number of samples, and 128 is the maximum.
Further, in the step 3, the dual-flow fusion CNN-BiLSTM network structure and network basic parameter setting are described as follows:
(3.1) one stream in the double-stream fusion network extracts local original time characteristics from an original I/Q signal, the other stream learns knowledge from an A/P signal, namely the amplitude and phase information after preprocessing, and the characteristics learned from the two streams are interacted in pairs by adopting double-stream information interaction, so that more effective classification characteristics can be acquired, the characteristic diversity is increased, and the performance is further improved;
(3.2) the specific CNN-BiLSTM network model uses CNN and Bi-LSTM network splicing structures, the CNN network structure of each stream is used for extracting the spatial characteristics of data, and the Bi-LSTM network extracts the time characteristics of the data in a two-way long and short time memory mode; each stream network model comprises three zero filling layers, three two-dimensional LSTM layers and a full connection layer, wherein a two-dimensional I/Q signal data set is processed into a three-bit 1 x 2 x 128 size input model, the three convolution layers respectively comprise 256 convolution kernels, 128 and 256 convolution kernels, the first convolution kernel is 2*3, the second convolution kernel is 1*3, the third convolution kernel is 1*3, and finally the full connection layer outputs 11-bit feature vectors which are expressed as prediction probabilities of 11 signals, wherein both the CNN layer and the LSTM layer sample a dropout method, and the dropout rate is p=0.5 so as to avoid overfitting;
(3.3) per-layer convolution and per-layer Bi-LSTM of the network structure using ReLU as the activation function, the network structure using class cross entropy error as the loss function:
wherein y is i Representing a predicted value, N representing the number of training samples;
(3.4) the network model uses a random gradient descent algorithm Adam as an optimizer:
setting the learning rate mu as 0.0001, setting the loss function to be minimized, and updating network parameters in an iterative way;
and (3.5) the fusion of the two-stream network uses a feature interaction mode of feature transposition multiplication, and the feature fusion is used for enhancing the performance, so that different from the common weighted fusion, more features can be extracted, the diversity of the attributes is enriched, and the performance is further enhanced.
The beneficial effects of the invention are as follows:
1. according to the invention, a network structure combining CNN and BiLSTM is adopted, so that the time and space characteristics of a modulation signal are automatically extracted, and ReLU is used as an activation function in each layer, so that the characteristic information is reduced, the calculation process is quickened, and more characteristic information is reserved to the maximum extent;
2. according to the invention, a CNN-BiLSTM network structure with double-flow fusion is adopted, a first-flow network directly inputs I/Q form signals, a first-flow network directly inputs A/P form signals, and a feature matrix transposition multiplication feature fusion mode is adopted, so that more feature information is extracted, and the recognition accuracy is improved;
drawings
FIG. 1 is a workflow diagram of a modulated signal identification method based on a dual-flow converged CNN-BiLSTM network;
FIG. 2 is a schematic diagram of the operation of a network model;
fig. 3 is an identification performance (snr= -2) of different modulated signals;
fig. 4 is an identification performance (snr=10) of different modulated signals;
FIG. 5 is a performance diagram of average classification accuracy;
fig. 6 is a schematic diagram of the operation of the classification accuracy of each modulated signal.
Detailed Description
The following detailed description of the preferred embodiments of the invention, taken in conjunction with the accompanying drawings, is provided to enable those skilled in the art to more simply and quickly understand the advantages and features of the invention, thereby making clear and defining the scope of the invention.
Referring to fig. 1 to 6, a modulation signal identification method based on a dual-flow fusion CNN-BiLSTM network includes the steps of:
step 1: preparing a data set;
the data set is a public available data set Radio ML2016.10a generated by using GNU Radio synthesis, and comprises training data and test data, wherein the types of the data set comprise 8 digital modulation signals such as BPSK, QPSK, 8PSK, 16QAM, 64QAM, BFSK, CPFSK and PAM4 and 3 analog modulation signals such as WB-FM, AM-SSB and AM-DSB;
step 2: data preprocessing, converting signals into a time inphase/quadrature (I/Q) format and an amplitude/phase (a/P) representation, forming two data sets;
the data preprocessing process comprises the following steps:
(2.1) the modulated signal is an I/Q signal consisting of two sets of sequences, I and Q, which are converted into I/Q vectors;
j represents the number of data set samples, the maximum value is 220000, N represents the number of sampling points, and the maximum value is 128;
(2.2) calculating amplitude and phase vectors:
(2.3) constructing an AP dataset consisting of amplitude components and phase components and converting to a/P vectors:
j represents the number of data set samples, the maximum value is 220000, N represents the number of sampling points, and the maximum value is 128;
step 3: building a double-flow fusion CNN-BiLSTM network structure and setting network basic parameters;
the double-flow fusion CNN-BiLSTM network structure and network basic parameter setting are described as follows:
(3.1) one stream in a dual stream fusion network is to extract local raw time features from the raw I/Q signal, and the other stream is to learn knowledge from the a/P signal, i.e. the pre-processed amplitude and phase information. The double-flow information interaction is adopted, the features learned from the two flows are interacted in pairs, so that more effective classification features can be obtained, the feature diversity is increased, and the performance is further improved;
(3.2) the specific CNN-BiLSTM network model uses CNN and Bi-LSTM network splicing structures, the CNN network structure of each stream is used for extracting the spatial characteristics of data, and the Bi-directional long-short-term memory (Bi-LSTM) network extracts the temporal characteristics of data; each flow network model comprises three zero filling layers, three two-dimensional LSTM layers and a full connection layer; in a size input model in which a two-dimensional I/Q signal data set is processed into three bits of 1×2×128, three convolution layers respectively include 256, 128 and 256 convolution kernels, wherein the size of the first convolution layer convolution kernel is 2*3, the size of the second convolution layer convolution kernel is 1*3, the size of the third convolution layer convolution kernel is 1*3, and finally, a full-connection layer outputs 11-bit feature vectors which are represented as prediction probabilities of 11 signals, wherein both CNN layers and LSTM layers sample a dropout method, and the dropout rate is p=0.5 so as to avoid overfitting;
(3.3) per-layer convolution and per-layer Bi-LSTM of the network structure using ReLU as the activation function, the network structure using class cross entropy error as the loss function:
wherein y is i Represents the predicted value and N represents the number of training samples.
(3.4) the network model uses a random gradient descent algorithm Adam as an optimizer:
wherein the learning rate mu is set to 0.0001, the loss function is set to be minimized, and the network parameters are updated iteratively.
And (3.5) the fusion of the two-stream network uses a feature interaction mode of feature transposition multiplication, and the feature fusion is used for enhancing the performance, so that different from the common weighted fusion, more features can be extracted, the diversity of the attributes is enriched, and the performance is further enhanced.
Step 4: the training model is used for inputting the training set into a double-flow fusion CNN-BiLSTM network for training;
step 5: inputting the test set into a trained double-flow fusion CNN-BiLSTM network structure, automatically identifying the type of the modulation signal, and outputting the accuracy of signal identification.
The processing procedure of this embodiment is as follows:
step 1: preparing a data set;
the data set is a public available data set Radio ML2016.10a generated by using GNU Radio synthesis, and comprises training data and test data, wherein the types of the data set comprise 8 digital modulation signals such as BPSK, QPSK, 8PSK, 16QAM, 64QAM, BFSK, CPFSK and PAM4 and 3 analog modulation signals such as WB-FM, AM-SSB and AM-DSB.
Step 2: data preprocessing, converting signals into a time inphase/quadrature (I/Q) format and an amplitude/phase (a/P) representation, forming two data sets;
the specific method for preprocessing the data comprises the following steps:
(2.1) the modulated signal is an I/Q signal consisting of two sets of sequences, I and Q, which are converted into I/Q vectors;
j represents the number of data set samples, the maximum value is 220000, N represents the number of sampling points, and the maximum value is 128;
(2.2) calculating amplitude and phase vectors:
(2.3) constructing an AP dataset consisting of amplitude components and phase components and converting to a/P vectors:
j represents the data set sampleThe number is 220000 at maximum, N represents the number of sampling points, and the maximum is 128;
step 3: building a double-flow fusion CNN-BiLSTM network structure and setting network basic parameters;
the double-flow fusion CNN-BiLSTM network structure and network basic parameter setting are specifically described as follows:
(3.1) one stream in a dual stream fusion network is to extract local raw time features from the raw I/Q signal, and the other stream is to learn knowledge from the a/P signal, i.e. the pre-processed amplitude and phase information. The double-flow information interaction is adopted, the features learned from the two flows are interacted in pairs, so that more effective classification features can be obtained, the feature diversity is increased, and the performance is further improved;
(3.2) the specific CNN-BiLSTM network model uses CNN and Bi-LSTM network splicing structures, the CNN network structure of each stream is used for extracting the spatial characteristics of data, and the Bi-directional long-short-term memory (Bi-LSTM) network extracts the temporal characteristics of data; each flow network model contains three zero-fill layers, three two-dimensional LSTM layers, and one full link layer. In a size input model in which a two-dimensional I/Q signal data set is processed into three bits 1×2×128, three convolution layers respectively include 256, 128, 256 convolution kernels, a first convolution kernel is 2*3 in size, a second convolution kernel is 1*3 in size, a third convolution kernel is 1*3 in size, and finally a full-connection layer outputs 11-bit feature vectors, which are expressed as prediction probabilities of 11 signals. Wherein both CNN layer and LSTM layer sample the dropout method with a dropout rate of p=0.5 to avoid overfitting.
(3.3) per-layer convolution and per-layer Bi-LSTM of the network structure using ReLU as the activation function, the network structure using class cross entropy error as the loss function:
wherein y is i Represents the predicted value and N represents the number of training samples.
(3.4) the network model uses a random gradient descent algorithm Adam as an optimizer:
wherein the learning rate mu is set to 0.0001, the loss function is set to be minimized, and the network parameters are updated iteratively.
And (3.5) the fusion of the two-stream network uses a feature interaction mode of feature transposition multiplication, and the feature fusion is used for enhancing the performance, so that different from the common weighted fusion, more features can be extracted, the diversity of the attributes is enriched, and the performance is further enhanced.
The foregoing description is only illustrative of the present invention and is not intended to limit the scope of the invention, and all equivalent structures or equivalent processes or direct or indirect application in other related technical fields are included in the scope of the present invention.
Claims (4)
1. The modulation signal identification method based on the double-flow fusion CNN-BiLSTM network is characterized by comprising the following steps:
step 1: preparing a data set;
step 2: data preprocessing, converting signals into a time inphase/quadrature (I/Q) format and an amplitude/phase (a/P) representation, forming two data sets;
step 3: building a double-flow fusion CNN-BiLSTM network structure and setting network basic parameters;
step 4: the training model is used for inputting the training set into a double-flow fusion CNN-BiLSTM network for training;
step 5: inputting the test set into a trained double-flow fusion CNN-BiLSTM network structure, automatically identifying the type of the modulation signal, and outputting the accuracy of signal identification.
2. The method for identifying modulated signals based on a dual-stream converged CNN-BiLSTM network according to claim 1, wherein in the step 1, the data set is a publicly available data set Radio ml2016.10a generated by using GNU Radio synthesis, and includes training data and test data, and the types of the data set include BPSK, QPSK, 8PSK, 16QAM, 64QAM, BFSK, CPFSK and PAM48 digital modulations and WB-FM, AM-SSB and AM-DSB3 analog modulated signals.
3. The method for identifying modulated signals based on a dual-stream converged CNN-BiLSTM network according to claim 1 or 2, wherein in the step 2, the data preprocessing process is as follows:
(2.1) the modulated signal is an I/Q signal consisting of two sets of sequences, I and Q, which are first converted into I/Q vectors;
j represents the number of samples of the data set, the maximum value is 220000, N represents the number of sampling points, and the maximum value is 128;
(2.2) calculating amplitude and phase vectors:
(2.3) constructing an AP dataset consisting of amplitude components and phase components and converting to a/P vectors:
4. The method for identifying modulated signals based on a dual-stream converged CNN-BiLSTM network according to claim 3, wherein in the step 3, the dual-stream converged CNN-BiLSTM network structure and the network basic parameter setting are described as follows:
(3.1) one stream in the dual stream fusion network is to extract local raw time features from the raw I/Q signal, and the other stream is to learn knowledge from the a/P signal, i.e. the preprocessed amplitude and phase information; the double-flow information interaction is adopted, the features learned from the two flows are interacted in pairs, so that more effective classification features can be obtained, the feature diversity is increased, and the performance is further improved;
(3.2) the specific CNN-BiLSTM network model uses CNN and Bi-LSTM network splicing structures, the CNN network structure of each stream is used for extracting the spatial characteristics of data, and the Bi-LSTM network extracts the time characteristics of the data in a two-way long and short time memory mode; each flow network model comprises three zero filling layers, three two-dimensional LSTM layers and a full connection layer; in a size input model in which a two-dimensional I/Q signal data set is processed into three bits 1×2×128, three convolution layers respectively comprise 256, 128 and 256 convolution kernels, wherein the size of the first convolution layer convolution kernel is 2*3, the size of the second convolution layer convolution kernel is 1*3, the size of the third convolution layer convolution kernel is 1*3, and finally, a full-connection layer outputs 11-bit feature vectors which are expressed as prediction probabilities of 11 signals; the CNN layer and the LSTM layer both sample a dropout method, where the dropout rate is p=0.5 to avoid overfitting;
(3.3) per-layer convolution and per-layer Bi-LSTM of the network structure using ReLU as the activation function, the network structure using class cross entropy error as the loss function:
wherein y is i Representing a predicted value, N representing the number of training samples;
(3.4) the network model uses a random gradient descent algorithm Adam as an optimizer:
setting the learning rate mu as 0.0001, setting the loss function to be minimized, and updating network parameters in an iterative way;
and (3.5) the fusion of the two-stream network uses a feature interaction mode of feature transposition multiplication, and the feature fusion is used for enhancing the performance, so that different from the common weighted fusion, more features can be extracted, the diversity of the attributes is enriched, and the performance is further enhanced.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211579515.2A CN116016071A (en) | 2022-12-09 | 2022-12-09 | Modulation signal identification method based on double-flow fusion CNN-BiLSTM network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211579515.2A CN116016071A (en) | 2022-12-09 | 2022-12-09 | Modulation signal identification method based on double-flow fusion CNN-BiLSTM network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116016071A true CN116016071A (en) | 2023-04-25 |
Family
ID=86025702
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211579515.2A Pending CN116016071A (en) | 2022-12-09 | 2022-12-09 | Modulation signal identification method based on double-flow fusion CNN-BiLSTM network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116016071A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117932312A (en) * | 2024-03-22 | 2024-04-26 | 山东大学 | Radio positioning recognition system based on space-time attention network and contrast loss |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111832417A (en) * | 2020-06-16 | 2020-10-27 | 杭州电子科技大学 | Signal modulation pattern recognition method based on CNN-LSTM model and transfer learning |
CN113298031A (en) * | 2021-06-16 | 2021-08-24 | 中国人民解放军国防科技大学 | Signal modulation identification method considering signal physical and time sequence characteristics and application |
CN113486724A (en) * | 2021-06-10 | 2021-10-08 | 重庆邮电大学 | Modulation identification model based on CNN-LSTM multi-tributary structure and multiple signal representations |
CN114091541A (en) * | 2021-11-19 | 2022-02-25 | 嘉兴学院 | Multi-mode ensemble learning automatic modulation recognition method for offshore complex environment |
-
2022
- 2022-12-09 CN CN202211579515.2A patent/CN116016071A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111832417A (en) * | 2020-06-16 | 2020-10-27 | 杭州电子科技大学 | Signal modulation pattern recognition method based on CNN-LSTM model and transfer learning |
CN113486724A (en) * | 2021-06-10 | 2021-10-08 | 重庆邮电大学 | Modulation identification model based on CNN-LSTM multi-tributary structure and multiple signal representations |
CN113298031A (en) * | 2021-06-16 | 2021-08-24 | 中国人民解放军国防科技大学 | Signal modulation identification method considering signal physical and time sequence characteristics and application |
CN114091541A (en) * | 2021-11-19 | 2022-02-25 | 嘉兴学院 | Multi-mode ensemble learning automatic modulation recognition method for offshore complex environment |
Non-Patent Citations (1)
Title |
---|
彭岑昕;程伟;李晓柏;张永利;: "一种基于STFT-BiLSTM的通信信号调制方式识别方法", 空军预警学院学报, no. 01, 15 February 2020 (2020-02-15) * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117932312A (en) * | 2024-03-22 | 2024-04-26 | 山东大学 | Radio positioning recognition system based on space-time attention network and contrast loss |
CN117932312B (en) * | 2024-03-22 | 2024-06-04 | 山东大学 | Radio positioning recognition system based on space-time attention network and contrast loss |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110855591B (en) | QAM and PSK signal intra-class modulation classification method based on convolutional neural network structure | |
CN108234370B (en) | Communication signal modulation mode identification method based on convolutional neural network | |
Zhang et al. | Automatic modulation classification using CNN-LSTM based dual-stream structure | |
CN112702294B (en) | Modulation recognition method for multi-level feature extraction based on deep learning | |
CN108718288B (en) | Method for recognizing digital signal modulation mode based on convolutional neural network | |
CN109887047B (en) | Signal-image translation method based on generation type countermeasure network | |
CN113014524B (en) | Digital signal modulation identification method based on deep learning | |
CN114881092A (en) | Signal modulation identification method based on feature fusion | |
CN114553648A (en) | Wireless communication modulation mode identification method based on space-time diagram convolutional neural network | |
CN116016071A (en) | Modulation signal identification method based on double-flow fusion CNN-BiLSTM network | |
CN114896887A (en) | Frequency-using equipment radio frequency fingerprint identification method based on deep learning | |
Lin et al. | A real-time modulation recognition system based on software-defined radio and multi-skip residual neural network | |
CN114548201B (en) | Automatic modulation identification method and device for wireless signal, storage medium and equipment | |
CN116070136A (en) | Multi-mode fusion wireless signal automatic modulation recognition method based on deep learning | |
CN116628566A (en) | Communication signal modulation classification method based on aggregated residual transformation network | |
Wang et al. | A spatiotemporal multi-stream learning framework based on attention mechanism for automatic modulation recognition | |
CN114595729A (en) | Communication signal modulation identification method based on residual error neural network and meta-learning fusion | |
Perenda et al. | Contrastive learning with self-reconstruction for channel-resilient modulation classification | |
WO2024012185A1 (en) | Signal detection method and device, and storage medium | |
Cai et al. | The performance evaluation of big data-driven modulation classification in complex environment | |
Sang et al. | Deep learning based predictive power allocation for V2X communication | |
WO2021262052A1 (en) | A context aware data receiver for communication signals based on machine learning | |
CN114826850B (en) | Modulation identification method, device and equipment based on time-frequency diagram and deep learning | |
CN112580598B (en) | Radio signal classification method based on multichannel Diffpool | |
CN115955375A (en) | Modulated signal identification method and system based on CNN-GRU and CA-VGG feature fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |