CN118035885A - Multi-feature fusion-based aliasing signal modulation identification method - Google Patents
Multi-feature fusion-based aliasing signal modulation identification method Download PDFInfo
- Publication number
- CN118035885A CN118035885A CN202410116526.XA CN202410116526A CN118035885A CN 118035885 A CN118035885 A CN 118035885A CN 202410116526 A CN202410116526 A CN 202410116526A CN 118035885 A CN118035885 A CN 118035885A
- Authority
- CN
- China
- Prior art keywords
- capsule
- layer
- features
- feature
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 230000004927 fusion Effects 0.000 title claims abstract description 22
- 230000007246 mechanism Effects 0.000 claims abstract description 16
- 238000012545 processing Methods 0.000 claims abstract description 14
- 238000000605 extraction Methods 0.000 claims abstract description 10
- 238000013528 artificial neural network Methods 0.000 claims abstract description 6
- 238000007781 pre-processing Methods 0.000 claims abstract description 6
- 239000002775 capsule Substances 0.000 claims description 61
- 239000013598 vector Substances 0.000 claims description 32
- 230000006870 function Effects 0.000 claims description 28
- 230000004913 activation Effects 0.000 claims description 14
- 238000001228 spectrum Methods 0.000 claims description 14
- 108010076504 Protein Sorting Signals Proteins 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 8
- 238000011176 pooling Methods 0.000 claims description 7
- 235000009854 Cucurbita moschata Nutrition 0.000 claims description 6
- 240000001980 Cucurbita pepo Species 0.000 claims description 6
- 235000009852 Cucurbita pepo Nutrition 0.000 claims description 6
- 238000004364 calculation method Methods 0.000 claims description 6
- 235000020354 squash Nutrition 0.000 claims description 6
- 239000011159 matrix material Substances 0.000 claims description 5
- 210000002569 neuron Anatomy 0.000 claims description 3
- 238000010183 spectrum analysis Methods 0.000 claims description 3
- 238000003491 array Methods 0.000 claims description 2
- 238000010606 normalization Methods 0.000 claims description 2
- 230000001052 transient effect Effects 0.000 claims description 2
- 238000007405 data analysis Methods 0.000 abstract description 2
- 238000013135 deep learning Methods 0.000 abstract description 2
- 230000003595 spectral effect Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 5
- 230000006872 improvement Effects 0.000 description 3
- 238000010801 machine learning Methods 0.000 description 3
- 238000009825 accumulation Methods 0.000 description 2
- 238000013136 deep learning model Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000002759 z-score normalization Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003090 exacerbative effect Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/243—Classification techniques relating to the number of classes
- G06F18/2431—Multiple classes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/10—Pre-processing; Data cleansing
- G06F18/15—Statistical pre-processing, e.g. techniques for normalisation or restoring missing data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2131—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on a transform domain processing, e.g. wavelet transform
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
- G06N3/0442—Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04B—TRANSMISSION
- H04B17/00—Monitoring; Testing
- H04B17/30—Monitoring; Testing of propagation channels
- H04B17/309—Measuring or estimating channel quality parameters
- H04B17/345—Interference values
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D30/00—Reducing energy consumption in communication networks
- Y02D30/70—Reducing energy consumption in communication networks in wireless communication networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Electromagnetism (AREA)
- Image Analysis (AREA)
- Radar Systems Or Details Thereof (AREA)
Abstract
The invention belongs to the technical field of electromagnetic signal processing, and particularly relates to an electromagnetic data analysis method based on deep learning. An aliasing signal modulation identification method based on multi-feature fusion comprises the following steps: A. preprocessing data; B. extracting artificial features; C. automatic feature extraction; D. feature fusion; E. classifying; according to the method, the characteristics that the electromagnetic signals generated by a low-rail electromagnetic system are more and mostly aliased signals are combined, an aliased signal identification task is converted into a multi-label classification task, the characteristics are fused by combining a manual characteristic extraction method with deep characteristics of the extracted signals of a deep neural network, corresponding weights are given to different characteristics by an attention mechanism, the most reliable classification result is obtained according to different signal to noise ratios, and the modulation type identification of the aliased signals is completed. Compared with the traditional signal modulation recognition scheme, the method is more suitable for practical application scenes and can obtain higher recognition accuracy.
Description
Technical Field
The invention belongs to the technical field of electromagnetic signal processing, and particularly relates to an electromagnetic data analysis method based on deep learning.
Background
In recent years, with the development of communication technology, global communication networks have become more complex. The high density of device deployment and the tension in spectrum resources make the problem of aliasing interference of signals within the same channel unavoidable. In this case, the signals in the co-channel interfere with each other, so that the receiver cannot accurately separate and identify the single signal during demodulation. Such interference results in time-frequency overlapping and spatial interleaving in spatial communications, while further exacerbating the complexity of the electromagnetic environment, thereby presenting significant challenges for the design and optimization of the communication system. Modulation identification is critical to efficiently managing such complex environments. In addition, modulation identification has important applications in radio spectrum management, scout monitoring, radio wave interference analysis, and the like. Thus, modulation identification of aliased electromagnetic signals is an important issue in wireless communications. The traditional method performs poorly when facing complex electromagnetic environment, and the artificial intelligent algorithm is used for signal processing at present, so that the method is an effective solution to the problems of low recognition efficiency, long calculation time and the like.
Disclosure of Invention
The purpose of the invention is that: based on the technical framework of a low-rail electromagnetic system and combining the characteristics of electromagnetic data, the method suitable for processing complex aliasing signals is provided.
The technical scheme of the invention is as follows: an aliasing signal modulation identification method based on multi-feature fusion comprises the following steps:
A. Preprocessing data;
A1. for data tags: the modulation mode is represented by multi-heat coding;
A2. For data values: converting the original data into standard normal distribution by adopting Z-score standardization;
B. Extracting artificial features;
selecting instantaneous features, power spectrum features and high-order cumulant features of signals from time domain features, frequency domain features and statistical domain features as manual extraction features; wherein:
expressing the instantaneous characteristics of the signal in three aspects of instantaneous amplitude, instantaneous frequency and instantaneous phase;
Obtaining the spectrum characteristics of the signals through Welch spectrum analysis;
Extracting a fourth-order cumulant of the signal from the high-order cumulant feature;
C. Automatic feature extraction;
for the original data and amplitude phase, extracting shallow automatic features by adopting a convolution layer, and extracting advanced features by an attention mechanism, a capsule layer and BiLSTM;
D. Feature fusion;
D1. manually extracting feature combination;
B, converting the instantaneous characteristic, the power spectrum characteristic and the high-order cumulant characteristic obtained in the step B into three characteristic vectors, namely Feature1, feature2 and Feature3, splicing the three characteristic vectors, and extracting the three characteristic vectors from the three characteristic vectors by adopting a deep neural network to obtain F 2;
D2. Automatic feature combination of the original data and the amplitude phase;
b, expanding the shallow automatic features of the original data and the amplitude phase extracted by the convolution layer into a one-dimensional array, and then splicing to form F 0;
Splicing the high-level features extracted from the original data and the amplitude phase in the step C, inputting the high-level features into a capsule network and a BiLSTM network to extract higher-level features so as to form a high-dimensional automatic feature F 1;
D3. Multi-mode feature fusion;
Splicing and fusing the extracted artificial feature F 2, the shallow automatic feature F 0 and the high-dimensional automatic feature F 1 into F C;
D4. Learning a corresponding weight vector Q c by adopting an attention mechanism to fuse F C to form F a;
E. Classifying;
And obtaining the probability of each category through a sigmoid classifier, selecting and outputting the optimal threshold value according to F1-score under different signal to noise ratios, and finally outputting the final classification result.
The beneficial effects are that: according to the invention, by combining the characteristics that the electromagnetic signals generated by a low-rail electromagnetic system are more and mostly aliased signals, the aliased signal identification task is converted into a multi-label classification task, the characteristics are fused by combining a manual extraction characteristic method with deep characteristics of the extracted signals of a deep neural network, corresponding weights are given to different characteristics by an attention mechanism, the most reliable classification result is obtained according to different signal-to-noise ratios, and the modulation type identification of the aliased signals is completed. Under the condition of signal aliasing, the modulation type identification of the signal is realized, and compared with the traditional signal modulation identification scheme, the method is more suitable for practical application scenes and can obtain higher identification accuracy.
Drawings
FIG. 1 is a schematic flow chart of the present invention;
FIG. 2 is a flow chart of step D1 of the present invention;
FIG. 3 is a flow chart of step D2 of the present invention;
FIG. 4 is a flow chart of step D3 of the present invention;
FIG. 5 is a flow chart of step D4 of the present invention;
fig. 6-8 are graphs comparing performance at various signal-to-noise ratios using the present method in the examples.
Detailed Description
Example 1: referring to fig. 1, an aliasing signal modulation identification method based on multi-feature fusion comprises the following steps:
A. Preprocessing data;
A1. For data tags: assuming that there are n modulation methods in total, the set tag is a one-dimensional array composed of 0 and 1, each index corresponds to one modulation method, and if the data contains a certain modulation method, the corresponding index value is 1. For example, there are 11 total modulation schemes for the received data, with possible values including ("BPSK", "QPSK", "8 PSK", "16 QAM"). In this case, multiple thermal codes are used to represent these modulation schemes, as in [1,0,0,0,1,0,1,0,0,1].
A2. For data values: preprocessing of data values is an important step in data mining and machine learning processes, the purpose of which is to convert raw data into a format more suitable for algorithmic processing. Since it is necessary to extract a plurality of features of a signal, there is a difference between the dimensions and values of the respective features. In the process of feature fusion, if the scale difference of the features is too large, the features with large scale may have an excessive influence on the result of the model. In this example, the Z-score normalization method was used during the data preprocessing.
Z-score Standardization (Standard): the Z-score normalization converts the raw data to a standard normal distribution, i.e., 0 as the mean and 1 as the standard deviation.
B. Extracting artificial features;
Selecting instantaneous features, power spectrum features and high-order cumulant features of signals from time domain features, frequency domain features and statistical domain features as manual extraction features; the specific implementation mode is as follows:
Given a signal sequence a (i), i=1, 2., N, N s is the number of samples,
Average value of signal sequence:
Normalization of signal sequences: a n(i)=a(i)/ma of the total number of the components,
Centering or shifting the already normalized signal sequence: a cn(i)=an (i) -1.
(1) Transient characteristics
The scheme expresses the instantaneous characteristics of the signal from three aspects of instantaneous amplitude, instantaneous frequency and instantaneous phase, and an expression formula of the instantaneous characteristics is given below.
① Normalized center instantaneous amplitude absolute standard deviation:
② Instantaneous phase center nonlinear component absolute value criterion:
Wherein, Is a nonlinear component of the instantaneous phase, C is/>A t is the determination threshold of the weak signal.
③ Nonlinear-centered absolute standard deviation of instantaneous frequency:
where f N(i)=(f(i)-E[f(i)])/Rs,Rs is the symbol rate.
(2) High order cumulant features
The higher order cumulant is a measure of non-gaussian in the data set and is mainly used to extract the non-gaussian and non-linear characteristics of the signal in signal processing. Since the calculation amount of the excessively high-order accumulation amount is too large, the present scheme considers the fourth-order accumulation amount of the extracted signal, and its formula is as follows:
Wherein,
(3) Power spectral features
The power spectral signature is a spectral signature representation commonly used in signal processing to analyze the energy distribution of a signal at different frequencies, describing the energy density of the signal in the frequency domain. Advantages of the power spectral features include: the power spectrum characteristics reflect the distribution characteristics of the signals in the frequency domain, and can provide frequency component information of the signals; compared with the original signal, the power spectrum characteristic is more stable and is not easily influenced by instantaneous fluctuation; the power spectral features may emphasize the strong frequency components in the signal, i.e. those with larger energy, which is advantageous for distinguishing between different types of signals.
The welch spectrum is a commonly used power spectrum estimation method for calculating the power spectral density of a signal. It estimates the energy distribution of a signal in the frequency domain by segmenting the signal and fourier transforming each segment. Through the welch spectrum analysis, information such as the intensity of frequency components, the spectral shape, the band energy distribution, etc. of the signal can be obtained.
C. Automatic feature extraction;
and selecting a proper deep learning model to extract the characteristics of the electromagnetic signals according to the specific requirements of the task and the characteristics of the data. For the original data and amplitude phase, the convolution layer is adopted to extract the shallow automatic characteristics in the example, and then the advanced characteristics are extracted through the attention mechanism, the capsule layer and BiLSTM; the specific procedure is given below:
The convolution layer is used for extracting features in the data, the first layer of convolution layer takes the preprocessed data as input, and convolution operation is carried out between the input and the weight set. Let the convolution layer be l, assume that the input three-dimensional data d l[x,y,c].Kl is a convolution kernel of l layers, represented by four-dimensional data [ k x,ky,cl, c ], and 0.ltoreq.k x≤kx-1,0≤ky≤ky-1. The relationship of the output d l+1 of the convolutional layer to the input data can be expressed as:
Attention mechanisms are used to focus on important parts of the sequence data, where the present scheme uses channel attention and spatial attention in combination to focus on the channel and spatial dimensions of the data. Channel attention attempts to find the most important channel. Channel attention calculates the average a avg[c] and maximum a max[c] for each channel by global average pooling (Global Average Pooling, GAP) and global maximum pooling (Global Max Pooling, GMP), respectively:
amax[c]=max(∑x∑yd1+1[x,y,c])
Next, the results of GAP and GMP are concatenated and processed through the full concatenation layer and activation function to get channel attention weights. Assuming that the weights and biases of the fully connected layers are W 1,W2,b1,b2, the activation functions are ReLU and sigmoid, respectively, the process can be expressed as:
FCCon=concat(aavg,amax)
FC1=ReLU(W1*FCCon+b1)
FC2=σ(W2*FC1+b2)
where concat represents a join operation, x represents a matrix multiplication, + represents a matrix addition, and σ represents a sigmoid function.
Finally, the channel attention weight obtained by FC2 is applied to the input data d l+1, resulting in an attention weighted output f c [ x, y, c ]:
fc[x,y,c]=FC2*dl+1[x,y,c]。
For the spatial attention mechanism, first, the scheme requires calculating the average value AP [ x, y ] and the maximum value MP [ x, y ] of the input feature map f c[x,y,c ] along the channel direction (i.e., the c-dimension):
Where C is the number of channels, Indicating that the maximum is taken in the c-dimension. Then, the AP and MP are superimposed together to obtain M [ x, y ]:
M[x,y]=AP[x,y]+MP[x,y]
And processing M through a convolution layer and an activation function to obtain the spatial attention weight. The size of the convolution layer is 1x1, the convolution kernel is W s, and the bias term is b s:
as[x,y]=σ(Ws*M[x,y]+bs)
* Representing convolution operations
The input feature map is weighted by using the spatial attention weight, and an output fs [x,y,c] after spatial attention processing is obtained:
fs[x,y,c]=as[x,y]×fc[x,y,c]
X represents the multiplication operation at the element level.
And f [x,y,c] is obtained after the output processed by the attention mechanism passes through a layer of convolution layer, the feature extracted by the convolution layer is converted into a vector by the capsule layer, so that the information loss caused by pooling operation in the traditional convolution neural network is avoided, and the high-dimensional spatial feature is further learned.
First, a set of vector elements of size 1 XC 1 is obtained by the primary capsule layer, specifically, the output vector for the capsule i of the first layerAnd weight c ij can be expressed as:
Wherein, Is the input vector for capsule i of layer i, b ij is the linear combination of capsule i and capsule i of layer i+1, the formula is squash, which is a nonlinear activation function.
The capsule is recorded as:
W u represents the primary capsule layer convolution kernel weight parameter, N1 represents the number of capsules output by the primary capsule layer, C 1 is 16, and the squaring function is a nonlinear activation function of the capsules. Since each capsule is composed of a set of vector neurons, it has the ability to represent more characteristic information of an object.
And (3) carrying out weighted fusion on the primary capsules to finally obtain N groups of capsules, wherein the length of each capsule represents the probability that the target belongs to different categories. The input is primary capsule ui, the output is N capsules:
vj=[v1,v2,…,vN],j=1,2,…,N,
The advanced capsule is obtained from the primary capsule, and is mainly determined by a dynamic routing algorithm, which is also the core of a capsule network, and the specific flow is as follows:
each advanced capsule v i is a weighted sum of the output vectors u i of all primary capsules, where the weight c ij represents the contribution of the ith primary capsule to the jth advanced capsule:
S i is then converted to v i by a squash function to ensure that the length of v i does not exceed 1:
vj=squash(sj)
The weight c ij is dynamically determined by a "routing" process. Initially, all weights c ij are initialized to equal values. These weights are then updated by a number of iterations, based on the consistency of the predictions for each primary capsule with the actual output of each advanced capsule. Specifically, the present scheme calculates a "consistency" score for each primary capsule i and each advanced capsule j:
aij=ui·vj
where·represents the dot product of the vector.
The present solution then uses a "softmax" function to update the weights c ij:
This process may be iterated multiple times until the weights c ij converge, increasing the number of iterations may slightly improve performance, but the improvement in effect is not significant. Thus, the present scheme selects 5 iterations to find a balance between maintaining reasonable computational complexity and model performance.
The output N capsules also contain rich information, while LSTM can be used to handle long-term dependencies in the sequence data. BiLSTM can be expressed as:
Forward propagation: for the previous output capsule v j=[v1,v2,…,vN, forward LSTM calculates the hidden state at each time step t (F represents forward):
Wherein LSTM forward is a forward LSTM computation function whose inputs are input v t at the current time and the hidden state at the previous time
Back propagation: reverse LSTM calculates hidden states at each time step t(B represents backward):
Wherein LSTM backward is a calculation function of the reverse LSTM whose inputs are the input v t at the current time and the hidden state at the next time
Output calculation: finally, the forward and reverse hidden states are spliced to obtain a final output y t:
here [; and represents concatenation of vectors.
Aiming at the manually extracted characteristics, the scheme continues to learn the characteristics further by adopting a full connection layer.
D. Feature fusion;
The multi-modal feature fusion based on the attention mechanism is a method for fusing a plurality of feature sources and weighting by using the attention mechanism so as to realize more accurate classification tasks.
D1. and (5) manually extracting feature combination.
As shown in fig. 2, the instantaneous Feature, the power spectrum Feature and the high-order cumulant Feature obtained in the step B are converted into three Feature vectors Feature1, feature2 and Feature3, and the three Feature vectors are spliced together and extracted from the three Feature vectors by using a deep neural network to obtain F 2.
D2. Automatic feature combining of raw data and amplitude phase.
And B, expanding the shallow automatic features of the original data and the amplitude phase extracted by the convolution layer into a one-dimensional array, and then splicing to form F 0.
As shown in fig. 3, the advanced features extracted from the raw data and amplitude phase in step C are stitched together and input into the capsule network and BiLSTM network to extract the higher-order features to form a high-dimensional automatic feature F 1.
D3. Multi-mode feature fusion;
The purpose of feature fusion is to combine information from different sources or different modalities together in order to obtain a more comprehensive representation. Stitching is a common way of connecting different features together in sequence, preserving the complete information of each feature, and providing more ways of feature combination. The splicing fusion has flexibility and simplicity, can be combined with various machine learning and deep learning models, and simultaneously considers multi-mode information. Therefore, in this scheme, a splice fusion scheme is used.
As shown in fig. 4, the extracted artificial feature F 2, the shallow automatic feature F 0, and the high-dimensional automatic feature F 1 are splice-fused into F C.
D4. Not all features in Fc contribute to classification, however, in a few scenarios, there may be some immature and unfavorable features due to insufficient learning. Therefore, the scheme adopts the attention mechanism to learn the corresponding weight vector Q c, the weight vector Q c is obtained by two fully-connected layers, the number of neurons of the layer structure is the same as the number of features in Fc, and the activation function is Tanh, and the formula is as follows.
P1=Tanh(W1Fc+b1)
P2=Tanh(W2P1+b2)
Representing the corresponding multiplication of two one-dimensional arrays;
As shown in fig. 5, F C is fused by learning a corresponding weight vector Q c using an attention mechanism to form F a.
E. Classifying;
And obtaining the probability of each category through a sigmoid classifier, selecting and outputting the optimal threshold value according to F1-score under different signal to noise ratios, and finally outputting the final classification result. The specific process is as follows:
(1) The probability of each category is obtained through a sigmoid classifier: the sigmoid function is an activation function whose output ranges between 0 and 1. This makes it an ideal activation function in a classification task, as it can represent the probability that a class exists. The result is a predictive probability for each category.
(2) Selecting the threshold with the best output under different signal to noise ratios according to F1-score: is an index for evaluating the performance of classified tasks, and is a harmonic mean of accuracy and recall. Ranging from 0 to 1,1 representing a perfect classification and 0 representing a completely inaccurate classification. The model's F1-score is tested over different SNR's to determine the best decision threshold at each SNR level.
(3) And finally, outputting a final classification result: based on the selected optimal threshold, the model will output a final class label for each class. For example, if the predicted probability of a class is greater than a determined threshold, the model may mark that class as present (or 1), otherwise it is marked as absent (or 0).
Example 2, the procedure described in example 1 was verified:
an aliased signal dataset was created based on rml 2016.10a. Rml2016.10a is a public data set widely used in the fields of machine learning and digital signal processing, and a total of 220000 data samples include 11 modulation modes, specifically 8 digital modulation modes of 8PSK, BPSK, CPFSK, GFSK, PAM, QAM16, QAM64 and QPSK and 3 analog modulation modes of WBFM, AM-DSB and AM-SSB. Each modulation scheme has 20 signal-to-noise ratio (SNR) levels (from-20 dB to 18 dB), each level has 1000 IQ samples, each IQ sample is composed of two channels (real and imaginary parts), and there are 128 sampling points in total.
In this scenario, the scheme sets 3 independent transmitters based on each signal-to-noise ratio of the dataset, each transmitter being capable of generating signals corresponding to the signals in the RML2016.10A dataset, and these signals being independently selectable and controllable. Furthermore, the solution provides a receiver whose task is to receive the signals emitted by these transmitters. The signals actually received by the receivers tend to be a superposition of the signals sent by these transmitters, taking into account the effects of the physical environment and the propagation characteristics of the electromagnetic waves. However, due to physical distance between the transmitter and the receiver, speed differences between the transmitter and the receiver, etc., the signal from each transmitter may have some time delay and frequency delay when reaching the receiver. Finally, overlapping different signals with time delay and frequency delay, wherein 100 ten thousand groups of data are obtained, and the groups of data are divided into a training set, a verification set and a test set according to the proportion of 6:2:2.
The data processing is carried out on the group of data through the method, and the recognition result is obtained as follows:
The performance of this example at each snr, i.e., F1-score, accuracy and recall at each snr optimum threshold, is shown in fig. 6. At high signal-to-noise ratio, the total recognition accuracy of each modulation mode can reach 80%, as shown in fig. 7. Figure 8 shows the confusion matrix for each category at-12 dB, 0dB, 6dB, 12 dB.
While the invention has been described in detail in the foregoing general description and specific examples, it will be apparent to those skilled in the art that modifications and improvements can be made thereto. Accordingly, such modifications or improvements may be made without departing from the spirit of the invention and are intended to be within the scope of the invention as claimed.
Claims (4)
1. The aliasing signal modulation identification method based on multi-feature fusion is characterized by comprising the following steps of:
A. Preprocessing data;
A1. for data tags: the modulation mode is represented by multi-heat coding;
A2. For data values: converting the original data into standard normal distribution by adopting Z-score standardization;
B. Extracting artificial features;
selecting instantaneous features, power spectrum features and high-order cumulant features of signals from time domain features, frequency domain features and statistical domain features as manual extraction features; wherein:
expressing the instantaneous characteristics of the signal in three aspects of instantaneous amplitude, instantaneous frequency and instantaneous phase;
Obtaining the spectrum characteristics of the signals through Welch spectrum analysis;
Extracting a fourth-order cumulant of the signal from the high-order cumulant feature;
C. Automatic feature extraction;
for the original data and amplitude phase, extracting shallow automatic features by adopting a convolution layer, and extracting advanced features by an attention mechanism, a capsule layer and BiLSTM;
D. Feature fusion;
D1. manually extracting feature combination;
B, converting the instantaneous characteristic, the power spectrum characteristic and the high-order cumulant characteristic obtained in the step B into three characteristic vectors, namely Feature1, feature2 and Feature3, splicing the three characteristic vectors, and extracting the three characteristic vectors from the three characteristic vectors by adopting a deep neural network to obtain F 2;
D2. Automatic feature combination of the original data and the amplitude phase;
b, expanding the shallow automatic features of the original data and the amplitude phase extracted by the convolution layer into a one-dimensional array, and then splicing to form F 0;
Splicing the high-level features extracted from the original data and the amplitude phase in the step C, inputting the high-level features into a capsule network and a BiLSTM network to extract higher-level features so as to form a high-dimensional automatic feature F 1;
D3. Multi-mode feature fusion;
Splicing and fusing the extracted artificial feature F 2, the shallow automatic feature F 0 and the high-dimensional automatic feature F 1 into F C;
D4. Learning a corresponding weight vector Q c by adopting an attention mechanism to fuse F C to form F a;
E. Classifying;
And obtaining the probability of each category through a sigmoid classifier, selecting and outputting the optimal threshold value according to F1-score under different signal to noise ratios, and finally outputting the final classification result.
2. The method for identifying modulation of aliasing signals based on multi-feature fusion according to claim 1, wherein in the step B:
Given a signal sequence a (i), i=1, 2., N, N s is the number of samples,
Average value of signal sequence:
Normalization of signal sequences: a n(i)=a(i)/ma;
Centering or shifting the already normalized signal sequence: a cn(i)=an (i) -1;
the expression formula of the transient characteristics is as follows:
normalized center instantaneous amplitude absolute standard deviation:
instantaneous phase center nonlinear component absolute value criterion:
Wherein, Is a nonlinear component of the instantaneous phase, C is/>A t is a determination threshold value of a weak signal, which is the number of samples of a n(i)>at;
nonlinear-centered absolute standard deviation of instantaneous frequency:
wherein f N(i)=(f(i)-E[f(i)])/Rs,Rs is the symbol rate;
The fourth-order cumulant formula of the high-order cumulant feature extraction signal is as follows:
Wherein,
3. The method for identifying the modulation of the aliasing signal based on the multi-feature fusion according to claim 1 or 2, wherein the following method is adopted in the step C:
The first layer of convolution layer takes the preprocessed original data and amplitude phase as input, and carries out convolution operation between the input and the weight set; the convolution layer is denoted as l, and assuming that the input three-dimensional data d l[x,y,c],Kl is a convolution kernel of l layers, the convolution kernel is denoted by four-dimensional data [ k x,ky,cl, c ], and 0 is equal to or less than k x≤kx-1,0≤ky≤ky-1; the output d i+1 of the convolutional layer versus the input data is expressed as:
channel attention and spatial attention are used in combination to focus on the channel dimension and spatial dimension of the data;
Channel attention attempts to find the most important channel; channel attention is separately calculated by global average pooling, GAP, and global maximum pooling, GMP, average a avg c and maximum a max c for each channel:
amax[c]=max(∑x∑ydl+1[x,y,c])
connecting the results of GAP and GMP and processing the results through a full connection layer and an activation function to obtain a channel attention weight;
Assuming that the weights and biases of the full connection layer are W 1,W2,b1,b2, and the activation functions are ReLU and sigmoid, the process is expressed as follows:
FCCon=concat(aavg,amax)
FC1=ReLU(W1*FCcon+b1)
FC2=σ(W2*FC1+b2)
wherein concat represents a join operation, x represents a matrix multiplication, + represents a matrix addition, and σ represents a sigmoid function;
The channel attention weight obtained by FC2 is applied to the input data d l+1, resulting in an attention weighted output f c[x,yc]:
fc[x,y,c]=FC2*dl+1[x,yc]
for the spatial attention mechanism, the average value AP [ x, y ] and the maximum value MP [ x, y ] of the input feature map f c[x,y,c] along the channel direction, i.e., the c-dimension, are first calculated:
Where C is the number of channels, Representing a maximum in the c-dimension;
The AP and MP are overlapped together to obtain M [ x, y ]:
M[x,y]=AP[x,y]+MP[x,y]
Processing M through a convolution layer and an activation function to obtain a spatial attention weight; the size of the convolution layer is 1x1, the convolution kernel is W s, and the bias term is b s:
as[x,y]=σ(Ws*M[x,y]+bs)
Wherein, represents convolution operation
The input feature map is weighted by using the spatial attention weight, and the output f s[x,y,c] after spatial attention processing is obtained:
fs[x,y,c]=as[x,y]×fc[x,y,c]
where x represents the multiplication operation at element level;
and f [x,y,c] is obtained after the output processed by the attention mechanism passes through a layer of convolution layer, and the capsule layer converts the features extracted by the convolution layer into vectors from scalar quantities:
First, a group of vector elements with the size of 1 XC 1 is obtained through the primary capsule layer, and the output vector of the capsule i of the first layer is obtained And weight c ij is expressed as:
Wherein, Is the input vector of the capsule i of the first layer, b ij is the linear combination of the capsule i and the capsule i of the first +1 layer, the formula is squash function, is a nonlinear activation function;
The capsule is recorded as:
Wherein W u represents a primary capsule layer convolution kernel weight parameter, N1 represents the number of capsules output by a primary capsule layer, C1 is 16, and the squaring function is a nonlinear activation function of the capsules;
Weighting and fusing the primary capsules to finally obtain N groups of capsules, wherein the length of each capsule represents the probability that the target belongs to different categories; the input is primary capsule ui, the output is N capsules:
vj=[v1,v2,…,vN],j=1,2,…,N,
each advanced capsule v i is a weighted sum of the output vectors u i of all primary capsules, where the weight c ij represents the contribution of the ith primary capsule to the jth advanced capsule:
S i is then converted to v i by a squash function to ensure that the length of v i does not exceed 1:
vi=squash(si)
Initially, all weights c ij are initialized to equal values, and then updated by a number of iterations, based on the consistency of the predictions for each primary capsule with the actual outputs of each advanced capsule; for each primary capsule i and each advanced capsule j, a "consistency" score is calculated:
aij=ui·vj
Wherein, represents the dot product of the vector;
the weights c ij are updated using a "softmax" function:
Iterating for a plurality of times until the weight c ij converges;
BiLSTM the calculation process is expressed as:
Forward propagation: for the previous output capsule v j=[v1,v2,…,vN, forward LSTM calculates the hidden state at each time step t F represents forward:
Wherein LSTM forward is a forward LSTM computation function whose inputs are input v t at the current time and the hidden state at the previous time
Back propagation: reverse LSTM calculates hidden states at each time step tB represents backsaward:
Wherein LSTM backward is a calculation function of inverse LSTM whose inputs are input v t at the current time and hidden state at the next time
Splicing the forward hidden state and the reverse hidden state to obtain a final output y t:
Wherein, [; and represents concatenation of vectors.
4. The method for identifying modulation of aliasing signals based on multi-feature fusion of claim 3, wherein in step D4:
The weight vector Q c is obtained by two fully connected layers, the number of neurons of the layer structure is the same as the number of features in F C, the activation function is Tanh, and the formula is as follows:
P1=Tanh(W1Fc+b1)
P2=Tanh(W2P1+b2)
Wherein, Representing the corresponding multiplication of two one-dimensional arrays.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410116526.XA CN118035885B (en) | 2024-01-26 | 2024-01-26 | Multi-feature fusion-based aliasing signal modulation identification method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410116526.XA CN118035885B (en) | 2024-01-26 | 2024-01-26 | Multi-feature fusion-based aliasing signal modulation identification method |
Publications (2)
Publication Number | Publication Date |
---|---|
CN118035885A true CN118035885A (en) | 2024-05-14 |
CN118035885B CN118035885B (en) | 2024-08-09 |
Family
ID=90999652
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202410116526.XA Active CN118035885B (en) | 2024-01-26 | 2024-01-26 | Multi-feature fusion-based aliasing signal modulation identification method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN118035885B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112702294A (en) * | 2021-03-24 | 2021-04-23 | 四川大学 | Modulation recognition method for multi-level feature extraction based on deep learning |
CN114254680A (en) * | 2022-02-28 | 2022-03-29 | 成都大公博创信息技术有限公司 | Deep learning network modulation identification method based on multi-feature information |
CN115238749A (en) * | 2022-08-04 | 2022-10-25 | 中国人民解放军军事科学院系统工程研究院 | Feature fusion modulation identification method based on Transformer |
CN116070136A (en) * | 2023-01-29 | 2023-05-05 | 北京林业大学 | Multi-mode fusion wireless signal automatic modulation recognition method based on deep learning |
CN116471154A (en) * | 2023-05-23 | 2023-07-21 | 西安电子科技大学 | Modulation signal identification method based on multi-domain mixed attention |
-
2024
- 2024-01-26 CN CN202410116526.XA patent/CN118035885B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112702294A (en) * | 2021-03-24 | 2021-04-23 | 四川大学 | Modulation recognition method for multi-level feature extraction based on deep learning |
CN114254680A (en) * | 2022-02-28 | 2022-03-29 | 成都大公博创信息技术有限公司 | Deep learning network modulation identification method based on multi-feature information |
CN115238749A (en) * | 2022-08-04 | 2022-10-25 | 中国人民解放军军事科学院系统工程研究院 | Feature fusion modulation identification method based on Transformer |
CN116070136A (en) * | 2023-01-29 | 2023-05-05 | 北京林业大学 | Multi-mode fusion wireless signal automatic modulation recognition method based on deep learning |
CN116471154A (en) * | 2023-05-23 | 2023-07-21 | 西安电子科技大学 | Modulation signal identification method based on multi-domain mixed attention |
Non-Patent Citations (2)
Title |
---|
BOYU DENG ETAL: ""Beam Scheduling with Various Mission Demands in Data Relay Satellite Systems"", 《JOURNAL OF COMMUNICATIONS AND INFORMATION NETWORKS》, 31 December 2021 (2021-12-31), pages 396 - 410 * |
曾泓然 等: ""一种采用预分类式神经网络的混叠信号识别方法"", 《工程与应用》, 31 December 2023 (2023-12-31), pages 1138 - 1144 * |
Also Published As
Publication number | Publication date |
---|---|
CN118035885B (en) | 2024-08-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109450834B (en) | Communication signal classification and identification method based on multi-feature association and Bayesian network | |
CN112702294B (en) | Modulation recognition method for multi-level feature extraction based on deep learning | |
CN112818891B (en) | Intelligent identification method for communication interference signal type | |
CN116866129A (en) | Wireless communication signal detection method | |
CN114881092A (en) | Signal modulation identification method based on feature fusion | |
CN114553648B (en) | Wireless communication modulation mode identification method based on space-time diagram convolutional neural network | |
CN110808932B (en) | Multi-layer sensor rapid modulation identification method based on multi-distribution test data fusion | |
CN115249332A (en) | Hyperspectral image classification method and device based on space spectrum double-branch convolution network | |
CN112039820A (en) | Communication signal modulation and identification method for quantum image group mechanism evolution BP neural network | |
CN110336594A (en) | A kind of deep learning signal detecting method based on conjugate gradient decent | |
CN113723556B (en) | Modulation mode identification method based on entropy weighting-multi-mode domain antagonistic neural network | |
CN114726692A (en) | Radiation source modulation mode identification method based on SEResNet-LSTM | |
CN111562541B (en) | Software platform for realizing electric energy meter detection data management by applying CART algorithm | |
CN116070136A (en) | Multi-mode fusion wireless signal automatic modulation recognition method based on deep learning | |
CN118035885B (en) | Multi-feature fusion-based aliasing signal modulation identification method | |
CN115982613A (en) | Signal modulation identification system and method based on improved convolutional neural network | |
CN113095162B (en) | Spectrum sensing method based on semi-supervised deep learning | |
CN112866156B (en) | Radio signal clustering method and system based on deep learning | |
CN113872904A (en) | Multi-classification communication signal automatic modulation identification method based on ensemble learning | |
CN114118151B (en) | Intelligent spectrum sensing method with environment adaptation capability | |
CN117493568A (en) | End-to-end software function point extraction and identification method | |
CN117978595A (en) | Automatic modulation classification method and device, equipment and computer readable storage medium | |
CN115238748A (en) | Modulation identification method based on Transformer and decision fusion | |
CN115291019A (en) | Electric energy quality disturbance classification and identification system based on multi-task learning model | |
CN117896020B (en) | Test method and test system of microwave receiving and transmitting communication system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |