Next Article in Journal
Textile Electrodes: Influence of Knitting Construction and Pressure on the Contact Impedance
Previous Article in Journal
Review of Wearable Sensor-Based Health Monitoring Glove Devices for Rheumatoid Arthritis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Modulation Recognition Based on a DCN-BiLSTM Network

School of Communication and Information Engineering, Shanghai University, Shanghai 200444, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(5), 1577; https://doi.org/10.3390/s21051577
Submission received: 17 January 2021 / Revised: 15 February 2021 / Accepted: 19 February 2021 / Published: 24 February 2021
(This article belongs to the Section Sensor Networks)

Abstract

:
Automatic modulation recognition (AMR) is a significant technology in noncooperative wireless communication systems. This paper proposes a deep complex network that cascades the bidirectional long short-term memory network (DCN-BiLSTM) for AMR. In view of the fact that the convolution operation of the traditional convolutional neural network (CNN) loses the partial phase information of the modulated signal, resulting in low recognition accuracy, we first apply a deep complex network (DCN) to extract the features of the modulated signal containing phase and amplitude information. Then, we cascade bidirectional long short-term memory (BiLSTM) layers to build a bidirectional long short-term memory model according to the extracted features. The BiLSTM layers can extract the contextual information of signals well and address the long-term dependence problems. Next, we feed the features into a fully connected layer. Finally, a softmax classifier is used to perform classification. Simulation experiments show that the performance of our proposed algorithm is better than that of other neural network recognition algorithms. When the signal-to-noise ratio (SNR) exceeds 4 dB, our model’s recognition rate for the 11 modulation signals can reach 90%.

1. Introduction

As a significant technology for noncooperative wireless communication systems, automatic modulation recognition (AMR) plays an important role in practical civil and military applications, such as cognitive radio, interference recognition and spectrum monitoring [1]. In the absence of prior knowledge, it can identify the modulation type of an intercepted signal, providing parameter information for subsequent demodulation [2].
Traditional AMR algorithms can be divided into two categories. One is based on maximum likelihood (ML) theory [3], and the other is a feature-based (FB) method [4]. The first approach uses probability theory, hypothesis test theory and an appropriate decision strategy to solve the AMR problem. The feature-based approach first extracts the modulated signal characteristics and then completes the recognition using classifiers. In feature-based methods, the number of selected features influences the recognition performance. The main features used for modulation signal identification include instantaneous amplitude, phase, frequency, high-order cumulant [5], cyclic spectrum [6] and wavelet characteristics [7]. Many of the classifiers used are based on machine learning algorithms; these include decision trees, support vector machines (SVMs) [8] and artificial neural networks (ANNs) [9].
In recent years, deep learning (DL), which is a powerful machine learning approach, has achieved great success in diverse fields such as image classification [10] and speech recognition [11]. The concept of DL comes from the research of ANNs. A multilayer perceptron with multiple hidden layers is a DL structure. DL forms a more abstract high-level representation attribute category or feature by combining low-level features to discover distributed feature representations of data. Its goal is to allow machines to have the ability to analyze and learn similar to humans and recognize data such as text, images and sounds. DL solves nonlinear classification problems by using activation functions and uses regularization to improve the robustness of the model [12]. The DL-based methods cascade multilayer nonlinear processing units to extract features. This approach can automatically optimize the extracted features to minimize later classification errors.
Many scholars have applied DL to the field of AMR in recent years. For example, Lee et al. [13] proposed a new method that calculated 28 statistical characteristics of five modulation signals; these were then sent to a fully connected feedforward network to conduct classification. Wang Yu at al. [14] trained CNN on samples composed of in-phase and quadrature component signals to distinguish modulation patterns that are relatively easy to identify. At the same time, a CNN based on the constellation map was designed to identify modulation modes that were difficult to distinguish in previous CNNs and improved the ability to classify QAM signals under low SNRs. Li et al. [12] studied the AMR method based on the original IQ signal under the parameter estimation error. First, the influence of parameter estimation error on the performance of CNN classifier is analyzed. Then, an AMR method based on spatial transformation network (STN) is proposed, which improves the robustness under parameter estimation errors. Li et al. [15] proposed a deep joint learning algorithm based on CNN and kernel collaborative representation and discriminative projection (KCRDP), including deep learning and kernel dictionary learning, which improves the adaptability of small samples and reduces the computational complexity without prior knowledge and feature enhancement processing.
In 2016, O’Shea et al. [16] generated a public dataset of modulated signals, named RML2016.10a, using GNU Radio software; then, they used a convolutional neural network (CNN) to identify the modulated signals. Subsequent studies have also adopted this dataset for AMR research. That same year, O’Shea [17] proposed a method to optimize the CNN structure. West et al. [18] applied a CNN, a residual network (Resnet), a convolutional long short-term memory deep neural network (CLDNN) and an Inception network to the modulated signal identification task and compared their respective recognition performances. The results show that the modulation signal recognition performance is not solely dependent on the network depth. Zhang et al. [19] proposed a preprocessing signal representation that combined the in-phase, quadrature and fourth-order statistics of the modulated signals. Omar S. Mossad et al. [20] proposed a CNN that used a multitask learning scheme (MTL-CNN) to reduce the confusion between similar classes. Kumar et al. [21] introduced a signal distortion correction module (CM) to estimate the carrier frequency offset and phase noise of the received signal to improve modulation recognition accuracy of deep learning schemes. Liu [22] proposed a group lasso based lightweight DNN for AMR which can learn to prune via automatically removing the neurons in hidden layers.
In 2020, Jakob et al. [23] used a linear combination to enable the DL architecture to perform complex convolutions (CVCs) and learn the characteristics of the real and imaginary parts of modulated signals. Wu et al. [24] constructed a CNN followed by a long short-term memory (LSTM) model as the classifier (CNN-LSTM) to efficiently explore the temporal and spatial correlation. However, to the best of our knowledge, none of the currently available methods fully consider the phase and contextual information of the modulated signals simultaneously.
In view of the current problems, this paper proposes a deep complex network that cascades bidirectional long short-term memory network (DCN-BiLSTM). First, deep complex network (DCN) layers with convolution kernels of different scales are connected; then, the bidirectional long short- term memory (BiLSTM) layers are cascaded. Finally, a softmax classifier is used to classify 11 kinds of digitally modulated signals. The main contributions of this paper are as follows:
(1)
To the best of our knowledge, this study is the first to cascade DCN and BiLSTM models and apply them in the AMR field.
(2)
We demonstrate the effectiveness of the DCN-BiLSTM network through experiments. The experimental results show that the performance of the proposed algorithm is better than that of other neural network recognition algorithms. When the signal-to-noise ratio (SNR) exceeds 4 dB, the recognition rate of our proposed model on the 11 modulation signals can reach 90%.
The remainder of this paper is organized as follows. Section 2 shows the signal model. Section 3 proposes the DCN-BiLSTM network model and introduces its components. In Section 4, we report the results of an experiment conducted to evaluate the proposed method and provide the optimal parameter configuration for the DCN-BiLSTM model. In addition, we use a cross-validation method to evaluate the network. Finally, Section 5 concludes this paper.

2. Signal Model

This paper uses the open dataset named RML2016.10a [16]. Figure 1 illustrates the dataset generation technique used in [16]. For the Rayleigh fading channel, the received signal can be expressed as
r ( n ) = l = 1 L h l ( n ) s ( n n l ) e j 2 π f d l n T 0 + ω ( n )
where s ( n ) is the modulated signal sent by the communication transmitter and L is the number of multipaths; h l ( n ) is the Rayleigh fading factor of the lth path; n l is the delay of the lth path; f d l represents the Doppler frequency; and ω ( n ) is additive white Gaussian noise. In addition, to ensure that the channel model is similar to a real channel, the channel model of this dataset includes the sampling rate and carrier rate offsets. The specific modulation types and parameters are shown in Table 1.
For a received signal, if we know the modulation type range, sampling frequency, sampling rate offset range, carrier rate offset range and the other signal parameters included in Table 1, we can use the recognition system model proposed in Section 3.

3. The Proposed Algorithm

3.1. DCN-BiLSTM Network Model

As shown in Figure 2, we propose the DCN-BiLSTM network model for AMR. First, we preprocess the received signal and divide it into I and Q components. Then, we send I and Q to the DCN-BiLSTM network, which is designed for identification. Finally, we obtain the identified signal type. The network has four parts: an input layer, DCN layers, BiLSTM layers and a fully connected layer.
In the DCN layers, the I-channel data of the signal are convoluted with the I-channel convolution kernel of the complex-valued convolution kernel, while the Q-channel data are convoluted with the Q-channel convolution kernel. After convolution, the real features and imaginary features are output. The activation function for the complex-valued convolution is a rectified linear unit (ReLU) function, which is defined as follows:
Re L U ( x ) = x , x > 0 0 , x 0 ,
where x is the input. When x > 0 , the activation function has a linear relationship with the input.
The BiLSTM layers connect the contextual information among signals and build a bidirectional long short-term memory model for the extracted features. The fully connected layer uses the softmax activation function to output the predicted probability of the modulation information.
y i = S ( z ) i = e z i j = 1 C e z j , i = 1 , , C ,
where z is the output of the previous layer and eventually forms the input to the fully connected layer; C is the input dimension and the number of modulation types; and y i is the probability of an unknown signal being predicted as category i.
The algorithm uses the cross-entropy loss function to calculate the gradient in reverse to update the bias and weight values. The back-propagation update process is as follows
θ n + 1 = θ n η φ ( y , f ( x , θ ) ) θ n ,
where θ n is the bias or weight of the last moment; η is the learning rate; and φ is the loss function.

3.2. Deep Complex-Valued Network Module (DCN)

The DCN [25] layers are composed of many complex-valued convolution kernels. These complex-valued convolution kernels of different scales are stacked together to perform a hierarchical convolution operation on the input signal.
In the complex-valued convolution operation, the real and imaginary parts are convolved separately. In Cartesian notation, the complex input matrix is defined as M = M R + i M I . Similarly, the complex-valued convolution kernel matrix is defined as K = K R + i K I . These parameters, including M R , M I , K R , K I , are all real-valued matrices. The complex-valued convolution expression is
M = M K = ( M R + i M I ) ( K R + i K I ) .
where * is the operation of convolution. The above formula can be expanded to
M = { M R K R M I K I } + i { M R K I + M I K R } .
Figure 3 shows a schematic diagram of the complex-valued convolution operation. The real and imaginary convolutions of the complex-valued signal are expressed as follows:
R e { M K } I m { M K } = K R K I K I K R M R M I ,
where Re { M K } is the real part of the signal and Im { M K } is the imaginary part of the signal.
The outputs of the DCN layers will carry phase information and are used as the input to the next layer.

3.3. Bidirectional Long Short-Term Memory Module (BiLSTM)

We cascade a BiLSTM behind the DCN to facilitate the extraction of contextual information of features. The BiLSTM network is composed of both forward and reverse LSTM networks [26]. As shown in Figure 4, the LSTM network contains many LSTM memory cells [27] that each includes three control units, namely, an input gate, a forget gate and an output gate. Figure 5 shows a diagram of the LSTM memory cells [27].
In Figure 5, t represents the current moment. The input feature sequence x t and the output sequence of the previous time h t 1 are input to the memory cell. The forgetting factor f t is obtained via the forgetting gate and is expressed as follows:
f t = σ ( W f · [ h t 1 , x t ] + b f ) ,
where W f is the connection matrix of x t , h t 1 . b f is the offset matrix, and σ is the sigmoid activation function, which is used to control the information-passing rate. The expression is
σ ( x ) = 1 1 + e x ,
where the output value of σ is between 0 and 1. The input gate and memory status update information are
i t = σ ( W i · [ h t 1 , x t ] + b i ) C ˜ t = tanh ( W c · [ h t 1 , x t ] + b c ) C t = f t C t 1 + i t C ˜ t
where i t is the output of the input gate and t a n h is an activation function that generates candidate values C ˜ t . In addition, C ˜ t participates in the calculation to obtain the memory state C t .
Among these various components, the memory state C t is the most important because it can allow information to flow through the entire link under the condition that it must remain unchanged, ensuring the integrity of the information for a long time. The output gate control factor o t determines whether to output information h t and is expressed as follows:
o t = σ ( W o · [ h t 1 , x t ] + b o ) h t = o t tanh ( C t )
where W o is the output gate weight matrix and b o is the offset matrix. Compared with C t , h t contains more information about the current moment. Therefore, h t represents short-term memory, while C t represents long-term memory.
However, a classical LSTM considers only information from the previous moment. To consider both the former moment and the next moment together, the BiLSTM [26] adds reverse operations based on the LSTM model in [28,29]. Figure 6 shows a structural operation graph of the BiLSTM.
As shown in Figure 6, the BiLSTM reverses the input sequence and calculates the output again in the same way as an LSTM. The final result is a stack of the forward LSTM and the reverse LSTM, which achieves the goal of considering the contextual information. The final outputs of the BiLSTM are h t , where t = 1 , 2 , , n , can be expressed as shown in Figure 7. The expression of h t is as follows:
h t = [ h t f , h t b ] .
The output features of the BiLSTM are mapped into a sparse space by the fully connected layer. After the network is trained, the algorithm outputs the classification probability of the corresponding modulation modes.

4. Experiment Results and Discussions

The relevant platform and software settings for this experiment are shown in Table 2. There are 1000 samples of each modulated signal for each SNR comprising a total of samples is 220,000 samples. The ratio of training sets to test sets is 8:2. We use the np.random.choice function to implement the proportional selection of the dataset to obtain the training sets and the test sets. In this experiment, we performed the following steps.
Step 1 Initialize the DCN-BiLSTM network randomly, extract a specified number of samples in the training sets and input them into the network for training.
Step 2 Compare the classification result obtained in the last layer of the network with the actual type; use the cross-entropy function to calculate the network loss value; and adjust the network weight value through the optimization algorithm.
Step 3 Before the next training starts, use the loss value as a standard to measure the network performance. When it does not drop within 10 iterations, the training is stopped.
Step 4 Repeat Steps 2–4 until the maximum number of training is reached or the conditions for premature termination of training are met. The maximum number of training in this article is 150. After training, the weights are saved and the classification model is output.
Step 5 Input the test sets into the trained model to obtain the recognition result.

4.1. Algorithm Performance Comparison

We selected the recognition algorithms based on a CNN [17], Resnet [18], Inception [18], CLDNN [18], MTL-CNN [20], CVC [23] and CNN-LSTM [24] as benchmark models. A performance comparison chart for these eight recognition algorithms is shown in Figure 8.
Figure 8 shows that, from −2 to 18 dB, the accuracy of the DCN-BiLSTM network is substantially higher than the accuracy of the other seven recognition algorithms. When the SNR exceeds 4 dB, the recognition accuracy of the DCN-BiLSTM network for the 11 modulation signals can reach 90%.
Figure 9 shows the values of the cross-entropy loss function during DCN-BiLSTM network model training and testing. As the number of iterations increases, the training and test loss values continue to decrease, indicating that the real label and the predicted label are constantly converging. When the number of iterations reaches 23, the training and testing loss values no longer decrease, and the model obtained at this time is the best.
Figure 10 shows confusion matrices for the CNN, DCN and DCN-BiLSTM models when the SNR is 18 dB. According to Figure 10a, it is difficult for the CNN to distinguish QPSK (Quadrature Phase Shift Keying) and 8PSK (8 Phase Shift Keying) because the CNN does not fully extract the phase information: the amplitudes of QPSK and 8PSK are the same and the difference between them is the phase. In addition, 16QAM (16 Quadrature Amplitude Modulation) is often misrecognized as 64QAM (64 Quadrature Amplitude Modulation) because the constellation points of 16QAM can be found in the constellation points of 64QAM.
To fully consider the phase information, we replaced the CNN with a DCN. The accuracy confusion matrix is shown in Figure 10b. Clearly, the accuracies on the QPSK and 8PSK modulation types are much better than in Figure 10a, and the accuracy on the 16QAM and 64QAM modulation types has improved as well. However, the accuracy on the 16QAM and 64QAM types is still not good enough for practical applications; therefore, we still need to improve their recognition accuracy.
Considering the connections among data points, we cascaded the BiLSTM after the DCN to extract the contextual information of signals. An accuracy confusion matrix for the DCN-BiLSTM is shown in Figure 10c, showing that the recognition accuracy of 16QAM and 64QAM is greatly improved compared with the results in Figure 10a,b. This result indicates that the BiLSTM is useful for extracting the contextual features of signals.
As shown in Figure 10a–c, it is quite difficult to recognize wide band frequency modulation (WBFM). The reason is that the dataset uses voice signals to generate analog signals, and people’s voices have silent periods during speaking, leaving only a single carrier during the silent period. Thus, the WBFM signals can easily be misclassified as AM-DSB (amplitude modulation—double side band modulation) signals.
To fully compare the dataset recognition capabilities of the above several networks, we used an online platform that can perform statistical analysis [30]. First, we uploaded the file representing the recognition result in csv format, as shown in Table 3. Table 3 shows the recognition error rate of five types of datasets. The error rates of the first four datasets correspond to the corresponding modulated signals in the fifth dataset. According to Rodríguez-Fdez et al. [30] and the test situation in this paper, we selected Friedman [31] as the test type to be applied. We chose Holm [32], which is widely used, as the post-hoc with control method. At the same time, we set the significance level α to 0.05.
After experiments, the algorithm rankings obtained are shown in Table 4. As we can see, DCN-BiLSTM has the highest ranking and its performance is better than the other seven networks. Moreover, the ranking in Table 4 is consistent with the ranking of the recognition effect in Figure 8, which also shows the correctness of the experiment.
Table 5 summarizes the comparison between DCN-BiLSTM and the other seven algorithms by using post-hoc with control methods. By comparing p-value with α , it can be seen that the DCN-BiLSTM network is significantly different from Inception, CNN and Resnet, indicating that the proposed network has significant progress compared with them. At the same time, there are no significant differences between DCN-BiLSTM and CLDNN, CVC, CNN-LSTM and MTL-CNN, which means that the proposed algorithm inherits the excellent performance of the four networks and can replace them in the field of modulation recognition.

4.2. Performance Comparison When Using Different Parameters for the DCN-BiLSTM Network

To obtain the best parameter configuration for the DCN-BiLSTM network, this section studies the influences of each parameter configuration on the algorithm’s performance.
First, we change only the number of DCN layers to find the best number of DCN layers. The different recognition results are shown in Figure 11.
As shown in Figure 11, the overall recognition rate is the highest with six DCN layers when the SNR is greater than 2 dB. With fewer than six layers, the network’s ability to extract phase features is not strong enough. With more than six layers, the network extracts redundant features, and the recognition rate no longer improves, which wastes memory. Therefore, the best number of DCN layers for this algorithm is six.
Changing the BiLSTM layer number also affects the recognition performance. Figure 12 shows a comparison of the recognition performance under different numbers of BiLSTM layers.
With fewer than two BiLSTM layers, the algorithm’s ability to process feature information is poor. With more than two layers, while the accuracy rate is equivalent to that of a two-layer BiLSTM network, the added layers cause a speed reduction and waste memory. Therefore, it is best to set the number of BiLSTM layers to two.
Based on the previous analysis, the final DCN-BiLSTM network parameters are shown in Table 6.

4.3. Five-Fold Cross Validation

To evaluate the performance of the network proposed in this article, we used the five-fold cross-validation method to train and test the network in this part. We divide the dataset into five parts equally, and use one part as the test sets each time and the remaining four parts as the training sets. Finally we obtained five training and test results, as shown in Figure 13. The average recognition rate curve of these five times is shown in Figure 14. As shown in Figure 13 and Figure 14, when SNR is greater than 4 dB, the recognition accuracy of the proposed network is more than 90% under five trainings, covering the entire dataset, which illustrates the rationality and stability of the designed network.

5. Conclusions

In this paper, we propose a classification algorithm based on the DCN-BiLSTM network that achieves direct recognition of 11 different types of modulated signals. First, DCN layers are used to extract the phase features of the modulation signal. Then, BiLSTM layers are used to extract the contextual information and construct a bidirectional long short-term memory model for the features. Compared with previous network recognition algorithms based on CNN, Inception, Resnet, CLDNN, MTL-CNN, CVC and CNN-LSTM, the recognition accuracy of the DCN-BiLSTM network is significantly higher under high SNR. However, even when the SNR is as low as 4 dB, the recognition accuracy rate of the DCN-BiLSTM network can still reach 90%.
However, the DCN-BiLSTM network has a slow training speed, and the method works satisfactorily only for signals whose frequency offset and sampling frequency offset are within a certain range. In addition, the identified signal must belong to one of the 11 specified types. In future work, we plan to modify or optimize these issues. In particular, on the one hand, for the training speed issue, we can use other GPUs with stronger computing capabilities to speed up training. On the other hand, in terms of network structure, each LSTM unit in the BiLSTM layers contains three gate functions, resulting in more parameters, which is the main reason for the slow network training speed. Therefore, we can try to optimize the LSTM unit to reduce the number of parameters and increase the training speed, such as reducing or simplifying the gate function.
In addition to the phase information and contextual information mentioned in this paper, many other characteristics of modulated signals could be considered, such as time–frequency domain characteristics and constellation characteristics. In future work, we plan to use other features in combination with the DCN-BiLSTM network to improve the modulation signal identification performance.

Author Contributions

Conceptualization, K.L. and W.G.; Data curation, W.G.; Formal analysis, K.L.; Funding acquisition, Q.H.; Investigation, K.L. and W.G.; Methodology, K.L.; Software, W.G.; Supervision, K.L.; Validation, W.G.; Writing—original draft, W.G.; and Writing— review editing, K.L. and W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 61571279.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AMRAutomatic modulation recognition
DCN-BiLSTMDeep complex network-bidirectional long short-term memory network
DCNDeep complex network
BiLSTMBidirectional long short-term memory
CNNConvolutional neural network
SNRSignal-to-noise ratio
MLMaximum likelihood
FBFeature-based
SVMsSupport vector machines
ANNsArtificial neural networks
DLDeep learning
QAMQuadrature Amplitude Modulation
STNSpatial transformation network
KCRDPKernel collaborative representation and discriminative projection
ResnetResidual network
CLDNNLong short-term memory deep neural network
MTL-CNNCNN that uses a multitask learning scheme
CMCorrection module
DNNDeep neural network
CVCsComplex convolutions
LSTMLong short-term memory
ReLURectified linear unit
AM-DSBAmplitude modulation-Double sideband modulation
AM-SSBAmplitude modulation-Single sideband modulation
WBFMWide band frequency modulation
BPSKBinary Phase Shift Keying
8PSK8 Phase Shift Keying
QPSKQuadrature Phase Shift Keying
CPFSKContinuous Phase Frequency Shift Keying
GFSKGauss Frequency Shift Keying
PAM4Pulse Amplitude Modulation 4
16QAM16 Quadrature Amplitude Modulation
64QAM64 Quadrature Amplitude Modulation
GPUGraphic Processing Unit

References

  1. O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-air deep learning based radio signal classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef] [Green Version]
  2. Zeng, Y.; Zhang, M.; Han, F.; Gong, Y.; Zhang, J. Spectrum analysis and convolutional neural network for automatic modulation recognition. IEEE Wirel. Commun. Lett. 2019, 8, 929–932. [Google Scholar] [CrossRef]
  3. Xu, J.L.; Su, W.; Zhou, M. Likelihood-ratio approaches to automatic modulation classification. IEEE Trans. Syst. Man Cybern. Part C 2011, 41, 455–469. [Google Scholar] [CrossRef]
  4. Wu, Z.; Zhou, S.; Yin, Z.; Ma, B.; Yang, Z. Robust automatic modulation classification under varying noise conditions. IEEE Access 2017, 5, 733–741. [Google Scholar] [CrossRef]
  5. Wu, H.; Saquib, M.; Yun, Z. Novel automatic modulation classification using cumulant features for communications via multipath channels. IEEE Trans. Wirel. Commun. 2008, 7, 3098–3105. [Google Scholar]
  6. Yan, X.; Liu, G.; Wu, H.; Feng, G. New automatic modulation classifier using cyclic-spectrum graphs with optimal training features. IEEE Commun. Lett. 2018, 22, 1204–1207. [Google Scholar] [CrossRef]
  7. Ta, N.P. A wavelet packet approach to radio signal modulation classification. Proc. ICCS 1994, 1, 210–214. [Google Scholar]
  8. Park, C.; Choi, J.; Nah, S.; Jang, W.; Kim, D.Y. Automatic modulation recognition of digital signals using wavelet features and svm. Int. Conf. Adv. Commun. Technol. 2008, 1, 387–390. [Google Scholar]
  9. Zhao, Y.; Ren, G.; Wan, X.; Wu, Z.; Gu, X. Automatic digital modulation recognition using artificial neural networks. Int. Conf. Neural Netw. Signal Process. 2003, 1, 257–260. [Google Scholar]
  10. Machhour, A.; Mallahi, M.E.; Zouhri, A.; Chenouni, D. Image classification using shifted legendre-fourier moments and deep learning. In Proceedings of the 2019 7th Mediterranean Congress of Telecommunications (CMT), Fes, Morocco, 24–25 October 2019; pp. 1–6. [Google Scholar]
  11. Lakkhanawannakun, P.; Noyunsan, C. Speech recognition using deep learning. In Proceedings of the 2019 34th International Technical Conference on Circuits/Systems, Computers and Communications (ITC-CSCC), Jeju, Korea, 23–26 June 2019; pp. 1–4. [Google Scholar]
  12. Li, M.; Li, O.; Zhang, C. An automatic modulation recognition method with low parameter estimation dependence based on spatial transformer networks. Appl. Sci. 2019, 9, 5. [Google Scholar] [CrossRef] [Green Version]
  13. Lee, J.H.; Kim, J.; Kim, B.; Yoon, D.; Choi, J.W. Robust automatic modulation classification technique for fading channels via deep neural network. Entropy 2017, 19, 9. [Google Scholar]
  14. Wang, Y.; Liu, M.; Yang, J.; Gui, G. Data-driven deep learning for automatic modulation recognition in cognitive radios. IEEE Trans. Veh. Technol. 2019, 68, 4074–4077. [Google Scholar] [CrossRef]
  15. Li, D.; Yang, R.; Li, X.; Zhu, S. Radar signal modulation recognition based on deep joint learning. IEEE Access 2020, 8, 515–528. [Google Scholar] [CrossRef]
  16. O’Shea, T.; West, N. Radio machine learning dataset generation with gnu radio. Proc. GNU Radio Conf. 2016, 1, 1–6. [Google Scholar]
  17. O’Shea, T.J.; Corgan, J.; Clancy, T.C. Convolutional Radio Modulation Recognition Networks. In International Conference on Engineering Applications of Neural Networks; Springer: Berlin/Heidelberg, Germany, 2016; pp. 213–226. [Google Scholar]
  18. West, N.E.; O’Shea, T.J. Deep architectures for modulation recognition. In Proceedings of the 2017 IEEE International Symposium on Dynamic Spectrum Access Networks (DySPAN), Baltimore, MD, USA, 6–9 March 2017. [Google Scholar]
  19. Zhang, M.; Zeng, Y.; Han, Z.; Gong, Y. Automatic modulation recognition using deep learning architectures. In Proceedings of the 2018 IEEE 19th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Kalamata, Greece, 25–28 June 2018; pp. 1–5. [Google Scholar]
  20. Mossad, O.S.; ElNainay, M.; Torki, M. Deep convolutional neural network with multi-task learning scheme for modulations recognition. In Proceedings of the 2019 15th International Wireless Communications Mobile Computing Conference (IWCMC), Tangier, Morocco, 24–28 June 2019; pp. 1644–1649. [Google Scholar]
  21. Yashashwi, K.; Sethi, A.; Chaporkar, P. A learnable distortion correction module for modulation recognition. IEEE Wirel. Commun. Lett. 2019, 8, 77–80. [Google Scholar] [CrossRef] [Green Version]
  22. Liu, X.; Wang, Q.; Wang, H. A two-fold group lasso based lightweight deep neural network for automatic modulation classification. In Proceedings of the 2020 IEEE International Conference on Communications Workshops (ICC Workshops), Dublin, Ireland, 7–11 June 2020; pp. 1–6. [Google Scholar]
  23. Krzyston, J.; Bhattacharjea, R.; Stark, A. Complex-valued convolutions for modulation recognition using deep learning. In Proceedings of the 2020 IEEE International Conference on Communications Workshops (ICC Workshops), Kansas City, MO, USA, 20–24 May 2020; pp. 1–6. [Google Scholar]
  24. Wu, Y.; Li, X.; Fang, J. A deep learning approach for modulation recognition via exploiting temporal correlations. In Proceedings of the 2018 IEEE 19th International Workshop on Signal Processing Advances in Wireless Communications (SPAWC), Kalamata, Greece, 25–28 June 2018; pp. 1–5. [Google Scholar]
  25. Trabelsi, C.; Bilaniuk, O.; Serdyuk, D.; Subramanian, S.; Santos, J.F.; Mehri, S.; Rostamzadeh, N.; Bengio, Y.; Pal, C.J. Deep complex networks. arXiv 2017, arXiv:1705.09792. [Google Scholar]
  26. Huang, Z.; Xu, W.; Yu, K. Bidirectional lstm-crf models for sequence tagging. arXiv 2015, arXiv:1508.01991. [Google Scholar]
  27. Hajiaghayi, M.; Vahedi, E. Code failure prediction and pattern extraction using lstm networks. In Proceedings of the 2019 IEEE Fifth International Conference on Big Data Computing Service and Applications (BigDataService), Newark, CA, USA, 4–9 April 2019; pp. 55–62. [Google Scholar]
  28. Ma, W.; Yu, H.; Zhao, K.; Zhao, D.; Yang, J.; Ma, J. Tibetan location name recognition based on bilstm-crf model. In Proceedings of the 2019 International Conference on Cyber-Enabled Distributed Computing and Knowledge Discovery (CyberC), Guilin, China, 17–19 October 2019; pp. 412–416. [Google Scholar]
  29. Dyer, C.; Ballesteros, M.; Ling, W.; Matthews, A.; Smith, N.A. Transition-based dependency parsing with stack long short-term memory. arXiv 2015, arXiv:1505.08075. [Google Scholar]
  30. Rodríguez-Fdez, I.; Canosa, A.; Mucientes, M.; Bugarín, A. STAC: A web platform for the comparison of algorithms using statistical tests. In Proceedings of the 2015 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), Istanbul, Turkey, 2–5 August 2015. [Google Scholar]
  31. Friedman, M. The use of ranks to avoid the assumption of normality implicit in the analysis of variance. J. Am. Stat. Assoc. 1937, 32, 675–701. [Google Scholar] [CrossRef]
  32. Holm, S. A simple sequentially rejective multiple test procedure. Scand. J. Stat. 1979, 6, 65–70. [Google Scholar]
Figure 1. Data generation scheme in [16].
Figure 1. Data generation scheme in [16].
Sensors 21 01577 g001
Figure 2. DCN-BiLSTM network identification system model.
Figure 2. DCN-BiLSTM network identification system model.
Sensors 21 01577 g002
Figure 3. Schematic diagram of the complex-valued convolution operation.
Figure 3. Schematic diagram of the complex-valued convolution operation.
Sensors 21 01577 g003
Figure 4. LSTM network.
Figure 4. LSTM network.
Sensors 21 01577 g004
Figure 5. An LSTM memory cell.
Figure 5. An LSTM memory cell.
Sensors 21 01577 g005
Figure 6. The structural operation graph of BiLSTM.
Figure 6. The structural operation graph of BiLSTM.
Sensors 21 01577 g006
Figure 7. The relationship between h t and h t f , h t b .
Figure 7. The relationship between h t and h t f , h t b .
Sensors 21 01577 g007
Figure 8. AMR rates of different networks under different signal-to-noise ratios.
Figure 8. AMR rates of different networks under different signal-to-noise ratios.
Sensors 21 01577 g008
Figure 9. Loss value of DCN-BiLSTM algorithm.
Figure 9. Loss value of DCN-BiLSTM algorithm.
Sensors 21 01577 g009
Figure 10. Accuracy confusion matrices of CNN, DCN and DCN-BiLSTM algorithms when SNR = 18 dB.
Figure 10. Accuracy confusion matrices of CNN, DCN and DCN-BiLSTM algorithms when SNR = 18 dB.
Sensors 21 01577 g010
Figure 11. Algorithm recognition accuracy under different DCN layers.
Figure 11. Algorithm recognition accuracy under different DCN layers.
Sensors 21 01577 g011
Figure 12. Algorithm recognition accuracy under different BiLSTM layers.
Figure 12. Algorithm recognition accuracy under different BiLSTM layers.
Sensors 21 01577 g012
Figure 13. Recognition rate curve for five-fold cross-validation.
Figure 13. Recognition rate curve for five-fold cross-validation.
Sensors 21 01577 g013
Figure 14. Average recognition rate curve of five-fold cross-validation.
Figure 14. Average recognition rate curve of five-fold cross-validation.
Sensors 21 01577 g014
Table 1. The specific dataset parameters.
Table 1. The specific dataset parameters.
Data SourceRML2016.10a
Modulation typesAM-DSB, AM-SSB, WBFM, BPSK, 8PSK, QPSK, CPFSK,
GFSK, PAM4, 16QAM, 64QAM
Data length128
Data dimension2 × 128
Sampling frequency200 kHz
Sampling rate offset standard deviation0.01 Hz
Maximum sampling rate offset50 Hz
Carrier frequency offset standard deviation0.01 Hz
Maximum carrier rate offset500 Hz
Number of sinusoids used in frequency selective fading8
Maximum doppler frequency used in fading1
Fading modelRician
Rician K-factor4
Fractional sample delays for the power delay profile[0.0, 0.9, 1.7]
Magnitudes corresponding to each delay time[1, 0.8, 0.3]
Filter length to interpolate the power delay profile8
Standard deviation of the AWGN process 10 S N R 10
SNR(dB)−20:2:18
Table 2. The relevant experimental platform and software settings.
Table 2. The relevant experimental platform and software settings.
Signal generation software platformGNU Radio
Deep learning simulation platformTensorFlow 1.7.0
Deep learning libraryKeras
Hardware acceleration platformNVIDIA GTX1080Ti
Table 3. The recognition error rate of different networks for different datasets.
Table 3. The recognition error rate of different networks for different datasets.
DatasetsDCN-BiLSTMCNNResnetInceptionCLDNNMTL-CNNCVCCNN-LSTM
8PSK0.4420.6320.5140.5740.5230.4830.4610.498
16QAM0.4220.8760.7580.7940.7730.6210.8260.78
64QAM0.320.3810.5170.5530.4740.6140.4250.506
QPSK0.4290.7830.6980.7220.6010.5180.5830.499
RML2016.10a0.3790.4920.4760.4980.4580.4450.4530.445
Table 4. Algorithm rankings.
Table 4. Algorithm rankings.
RankAlgorithm
1.00000DCN-BiLSTM
3.70000CNN-LSTM
3.70000MTL-CNN
4.00000CVC
4.80000CLDNN
5.20000Resnet
6.60000CNN
7.00000Inception
Table 5. The comparison of DCN-BiLSTM with other algorithms using the post-hoc with control methods.
Table 5. The comparison of DCN-BiLSTM with other algorithms using the post-hoc with control methods.
ComparisonStatisticAdjusted p-ValueResult
DCN-BiLSTM vs Inception3.872980.00075H0 is rejected
DCN-BiLSTM vs CNN3.614780.00180H0 is rejected
DCN-BiLSTM vs Resnet2.711090.03353H0 is rejected
DCN-BiLSTM vs CLDNN2.452890.05669H0 is accepted
DCN-BiLSTM vs CVC1.936490.15842H0 is accepted
DCN-BiLSTM vs CNN-LSTM1.742840.16272H0 is accepted
DCN-BiLSTM vs MTL-CNN1.742840.16272H0 is accepted
Table 6. The DCN-BiLSTM network parameters.
Table 6. The DCN-BiLSTM network parameters.
LayerActivation FunctionOutput Dimensions
Input/(128, 2)
DCN layer L1ReLU(None, 64, 32)
DCN layer L2ReLU(None, 64, 128)
DCN layer L3ReLU(None, 64, 128)
DCN layer L4ReLU(None, 64, 128)
DCN layer L5ReLU(None, 64, 128)
DCN layer L6ReLU(None, 64, 256)
BiLSTM layer B1/(None, 64, 512)
BiLSTM layer B2/(None, 256)
DenseSoftmax(None, 11)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, K.; Gao, W.; Huang, Q. Automatic Modulation Recognition Based on a DCN-BiLSTM Network. Sensors 2021, 21, 1577. https://doi.org/10.3390/s21051577

AMA Style

Liu K, Gao W, Huang Q. Automatic Modulation Recognition Based on a DCN-BiLSTM Network. Sensors. 2021; 21(5):1577. https://doi.org/10.3390/s21051577

Chicago/Turabian Style

Liu, Kai, Wanjun Gao, and Qinghua Huang. 2021. "Automatic Modulation Recognition Based on a DCN-BiLSTM Network" Sensors 21, no. 5: 1577. https://doi.org/10.3390/s21051577

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop