A_Modulation_Classification_Algorithm
A_Modulation_Classification_Algorithm
This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000.
Digital Object Identifier 10.1109/ACCESS.2023.DOI
ABSTRACT Deep-learning is widely used in modulation classification to reduce labor and improve
the efficiency. Graph convolutional network (GCN) is a type of feature extraction network for graph
data. Considering the signals as graph nodes and the similarity of each signal as an edge, the GCN
propagates node information to similar nodes along the edges. GCN extracts more features and achieves
better classification results, particularly for characterless examples. In this paper, we propose a modu-
lation classification algorithm based on a feature-embedding GCN (FE-GCN). It comprises three parts:
feature-embedding network (FEN), similarity adjacent matrix calculation network (SAMCN), and graph
convolutional classification network (GCCN). The FEN embeds the signal data into a one-dimensional
feature vector. The SAMCN calculates the similarity of all signal feature vectors to a matrix using a
single convolutional neural network (CNN). The GCCN is used to extract the final features and classify the
signals in a graph. Simulation results on the public dataset RML2016.10A show that the FE-GCN performs
effectively and outperforms a series of advanced deep-learning methods.
VOLUME 4, 2016 1
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
ing strategy was used for pre-training. The ResNeXt model age recognition rates were 62.93% and 64.58%,respectively,
achieved the highest recognition accuracy of 96.10% on the which were higher than those of the other methods.
RadioML2016.10B dataset and 99.70% for the 10 modula- Currently, GCN are not widely used in modulation classi-
tion modes of the RadioML2018.01A dataset with a high fication, and they mainly adopt few-shot methods, which are
signal-to-noise ratio (SNR). In the same year, Zhang [15] complicated to calculate. The motivation of this paper is to
extracted both temporal and spatial features of modulation apply a feature-embedding GCN (FE-GCN) for modulation
signals using a CNN and a bidirectional long short-term recognition. It embeds signals using a simple CNN or LSTM
memory (Bi-LSTM) network. Combined with the correlation to simplify the inputs, and computes the similarity matrix
between the radio signal channels, they managed to improve using another CNN to construct a graph. The GCN extracts
the recognition accuracy to 92.68% with a high SNR. In features from both the signal and other similar signals to
particular, it reduced the difficulty o f identification of multi- improve the classification accuracy. The main contributions
ple Quadrature Amplitude Modulation (MQAM) signals, and of this paper are as follows:
significantly improved the recognition accuracy of QAM16 1) This novel GCN model for modulation classification
and QAM64 signals. Simultaneously, Weng proposed a deep combines different embedding networks to fully ex-
cascade network architecture (DCNA) [16] to address the tract the signal features and compensate for the clas-
difficulty of AMC under different SNRs. It includes an SNR sification efficiency of the GCN.
Estimation Network (SEN) to identify the SNR level of the 2) A similarity matrix is used to transfer the characteristic
samples and a Modulation Recognition Clustering Network information between signals to determine characterless
(MRCN) that contains several subnetworks for further mod- examples.
ulation recognition under different SNR Settings. Notably, 3) The FE-GCN simplifies the GCN application on AMC
DCNA does not exploit specific network structures and can and does not require the few-shot method or label infor-
be generalized to various network models through improve- mation in feature extraction and classification, which
ments. generalizes the GCN to universal scenarios.
In recent years, Graph Convolutional Networks (GCNs)
The remainder of this paper is organized as follows. Sec-
have gained widespread popularity for processing graph
tion II briefly describes the principle of graph convolution,
data by transmitting feature information between adjacent
and Section III presents the structure of the proposed FE-
nodes through a relation matrix. They have been success-
GCN model. Experiments on EF-GCN utilization for modu-
fully applied to various domains such as text classifica-
lation classification are discussed in Section IV, along with
tion, relation extraction, and image classification [17]. In
the simulation results and performance analysis. Finally,
2023, Zhao [18] proposed a learnable Graph Convolutional
Section V concludes the paper.
Network based on Feature Fusion (LGCN-FF) to learn the
underlying features and explore more discriminative graph
II. PRINCIPLE OF GRAPH CONVOLUTIONAL NEURAL
fusion techniques. LGCN-FF outperformed several state-of-
NETWORK
the-art methods in multiview semi-supervised classification
For a continuous function, convolution is a mathematical
tasks. Xiao [19] introduced a Dual Fusion-Propagation Graph
sum [23] h of functions f and g. It is modulated by the
Neural Network (DFP-GNN) for deep multiview clustering
signal processing, and the convolved signal is smoother than
tasks, achieving significant results compared with several
the original data. For discrete data, the convolution function
state-of-the-art algorithms on popular databases. Wu [20]
h(x, y) is the sum of the multiplication of the discrete data
employed an interpretable multiview Graph Convolutional
f (x, y) and local filters g(m, n) with a size of (m, n). It is
Network (IMvGCN) for a multiview semi-supervised learn-
primarily used in image processing and is expressed as
ing task, demonstrating its superiority over other state-of-the-
art methods through comprehensive experiments.
X
In 2020, Liu [21] originally applied a GCN to signal- h(x, y) = f (x, y)∗g(m, n) = f (x − m, y − n)g(m, n).
modulation classification. A feature extraction network m,n
(FECNN) was used to extract the signal features, and a graph (1)
mapping network (GMCNN) was used to concatenate the The convolution theorem applies to data in Euclidean
extracted features and label information. An adjacency ma- space, such as images, but not to graph structure data in a
trix was then constructed by calculating the distance between non-Euclidean space. A graph G(V, E) with N nodes consists
the concatenated features to create a graph. It achieved a of a node feature set V = {vi } and an edge set E =
higher recognition accuracy than the CNN and K-nearest {eij }, i, j = 1, 2, 3, ...,N. Edges represent the relationships
neighbor (KNN) algorithms, particularly for a low SNR. In between the nodes. The purpose of graph convolution is to
2021, Xuan [22] extracted signal information in a time series find a convolution method to deal with graph data that is sim-
using an Adaptive Visibility Graph (AVG) proposed to map ilar to convolution of images. According to the principle of
each signal into a graph, and used the GNN model for feature signal processing, early graph convolution transforms graph
extraction to achieve end-to-end signal recognition. For the data into the frequency domain by Fourier transform, and
open datasets RML2016.10A and RML2016.10b, the aver- then a convolution operation is performed, namely, spectral
2 VOLUME 4, 2016
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
Step 1: The training signals were divided into M subsets of size N to build
the graphs.
where φ is the symmetric function used to calculate edge Step 2: For epoch = 1, ..., EPOCH do
value. In this paper, edge set E was constructed by calculat- for m = 1, ..., M do
ing the similarities of the nodes as Embed all the signals of a graph into a one-dimensional
eigenvector of length L, Fm l = {F l , ..., F l }, and
m1 mN
φ(xi , xj ) = I − SAM CN (dif (xi , xj )), (6) Fm0 = {F 0 , ..., F 0 }.
m1 mN
for l = 1, ..., GCN_Layer do
where I is the identity matrix of size N×N and dif (xi , xj ) •Compute the adjacent matrix Ãlm using SAMCN;
is the absolute difference between node vectors i and j. The •Create a graph convolutional operation by
initial difference vector matrix A is mapped in numerical l = F l ⊙ Ãl .
F̃m m m
matrix à using a similarity adjacent matrix calculation net- •Aggregate similar features using
work(SAMCN). The SAMCN is a four-layer CNN, and its Fml+1
= Fm l + F̃ l .
m
parameters are adjusted by feedback during training. φ is end for
the result of combining the absolute difference between two Return loss and training accuracy, and training the model
eigenvectors nonlinearly, such that φ(xi , xj ) = φ(xj , xi ) parameters.
and φ(xi , xi ) = 1. The process of constructing the adjacency end for
matrix with N nodes is shown in Fig. 4. Validate with validation sets.
After feature embedding and adjacency matrix construc- Adjust the policy of learning parameters changing.
tion, a graph of N nodes can be built, whose nodes are signal Set the conditions for termination of training.
feature vectors and whose edges are adjacency matrix Ã. The end for
process of constructing a graph is shown in Fig. 5. Step 3: Save the trained model and parameters.
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
FIGURE 7. Accuracy of FE-GCN(Embedding with CNN). FIGURE 8. Accuracy of FE-GCN(Embedding with GRU).
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
D. T-SNE OF FE-GCN
FIGURE 10. Confusion matrix of FE-GCN(Embedding with GRU). The t-Distributed Stochastic Neighbor Embedding(t-SNE)
[31] is a tool to visualize high-dimensional data. It converts
similarities between data points to joint probabilities and tries
to minimize the Kullback-Leibler divergence between the
joint probabilities of the low-dimensional embedding and the
high-dimensional data. To better illustrate the classification
principle of EF-GCN, we took random samples with a SNR
of 10db and plotted t-SNEs when training. The t-SNEs of
EF-GCN(Embedding with CNN) at different training times
as shown in the Fig. 14. EF-GCN(Embedding with GRU)
is similar to EF-GCN(Embedding with CNN), signals are
mapped to a distribution with a rotation center. It is the mul-
tiplication of features and matrices that makes the difference
between GCN and CNN in the principle of feature extraction.
V. CONCLUSIONS
We proposed a general signal-modulation classification
FIGURE 11. Test accuracy on different graph size. model based on the GCN. In contrast to previous methods,
our model does not require the use of few-shot methods
VOLUME 4, 2016 7
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
8 VOLUME 4, 2016
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3385663
Communications Letters, vol. 25, no. 4, pp. 1206-1210, April 2021, doi: YUE ZHANG received the B.S. degree in elec-
10.1109/LCOMM.2020.3044755. tronic information science and technology from
[28] D. Hong, Z. Zhang and X. Xu, "Automatic modulation classification the University of Electronic Science and Tech-
using recurrent neural networks," 2017 3rd IEEE International Confer- nology of China, Chengdu, China, in 2012, and
ence on Computer and Communications (ICCC), 2017, pp. 695-700, doi: the M.S degree in computational intelligence from
10.1109/CompComm.2017.8322633. Sheffield University, Sheffield, United Kingdom,
[29] M. M. Bronstein, J. Bruna, Y. LeCun, A. Szlam and P. Vandergheynst, in 2013, and the Ph.D. degree in communication
"Geometric Deep Learning: Going beyond Euclidean data," in IEEE
and information systems from Xidian University
Signal Processing Magazine, vol. 34, no. 4, pp. 18-42, July 2017, doi:
in 2018. He was an associated professor in Un-
10.1109/MSP.2017.2693418.
[30] X. Chi, S. Huang and J. Li, "Handwriting Recognition Based on Resnet- manned System Research Institute, Northwestern
18," 2021 2nd International Conference on Big Data , Artificial Intel- Polytechnical University and currently a postdoc in Air Force Engineering
ligence and Software Engineering (ICBASE), 2021, pp. 456-459, doi: University. His research interests include machine learning, multi-agent re-
10.1109/ICBASE53849.2021.00091. inforcement learning, deep reinforcement learning, game theory, the Internet
[31] A. C. Belkina, C. O. Ciccolella, R. Anno, et al, "Automated optimized of Things, intelligent transportation systems, and big data.
parameters for T-distributed stochastic neighbor embedding improve vi-
sualization and analysis of large datasets," in Nature Communications,
5415(2019), DOI:10.1038/s41467-019-13055-y.
VOLUME 4, 2016 9
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4