CN115316955A - A lightweight and fast decoding method for motor imagery EEG signals - Google Patents
A lightweight and fast decoding method for motor imagery EEG signals Download PDFInfo
- Publication number
- CN115316955A CN115316955A CN202211087807.4A CN202211087807A CN115316955A CN 115316955 A CN115316955 A CN 115316955A CN 202211087807 A CN202211087807 A CN 202211087807A CN 115316955 A CN115316955 A CN 115316955A
- Authority
- CN
- China
- Prior art keywords
- eeg signal
- convolution
- motor imagery
- module
- decoding method
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/24—Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
- A61B5/316—Modalities, i.e. specific diagnostic methods
- A61B5/369—Electroencephalography [EEG]
- A61B5/372—Analysis of electroencephalograms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Mathematical Physics (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Public Health (AREA)
- General Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Veterinary Medicine (AREA)
- General Engineering & Computer Science (AREA)
- Animal Behavior & Ethology (AREA)
- Psychiatry (AREA)
- Surgery (AREA)
- Pathology (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Signal Processing (AREA)
- Physiology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Fuzzy Systems (AREA)
- Psychology (AREA)
- Feedback Control In General (AREA)
- Manipulator (AREA)
Abstract
本发明公开了一种轻量快速的运动想象脑电信号解码方法,按如下步骤进行:S1、构建深度学习模型,所述深度学习模型包括时空卷积模块、池化模块和全连接模块;所述时空卷积模块由用于减少可训练参数量的时间卷积层和用于减少通道连接的空间深度卷积层构成;所述池化模块是堆叠池化层以降低模型的维度并减少复杂度;所述全连接模块用于最终的分类;S2、对原始脑电信号进行预处理,然后利用深度学习模型对预处理后的脑电信号进行分类解码。本发明可以以较少的可训练参数量获得更好地解码性能,在分类精度与模型复杂度之间保持了相对的平衡。
The invention discloses a light-weight and fast method for decoding electroencephalographic signals of motor imagery. The steps are as follows: S1. Build a deep learning model, wherein the deep learning model includes a spatiotemporal convolution module, a pooling module and a fully connected module; The spatiotemporal convolution module consists of a temporal convolution layer for reducing the amount of trainable parameters and a spatial depth convolution layer for reducing channel connections; the pooling module stacks pooling layers to reduce the dimension of the model and reduce complexity degree; the fully connected module is used for final classification; S2, preprocess the original EEG signal, and then use a deep learning model to classify and decode the preprocessed EEG signal. The present invention can obtain better decoding performance with less trainable parameters, and maintains a relative balance between classification accuracy and model complexity.
Description
技术领域technical field
本发明涉及脑电信号分析技术领域,特别涉及一种轻量快速的运动想象脑电信号解码方法。The invention relates to the technical field of electroencephalogram signal analysis, in particular to a lightweight and fast motor imagery electroencephalogram signal decoding method.
背景技术Background technique
在运动想象脑电解码任务中,传统的手动提取特征的机器学习算法最典型的是共同空间模式、滤波器组共同空间模式、短时傅立叶变换以及主成分分析等。CSP算法的基本思想是找到一个空间滤波器,在运动想象脑电四分类任务中最大化此距离。相似地,FBCSP方法是CSP技术的拓展,也经常使用于脑电解码任务中。该算法通过一组带通滤波器提取最优的空间特征,从而选择及分类特征。同时,在过去的运动想象脑电研究中,经常会采用例如线性分类器、支持向量机、多层感知器以及随机森林这样的分类方法。尽管这些方法在运动想象脑电解码任务中已经获得了良好的结果,但是其将特征提取和分类分成了两个阶段。In motor imagery EEG decoding tasks, the most typical machine learning algorithms for manually extracting features are common spatial patterns, filter bank common spatial patterns, short-time Fourier transform, and principal component analysis. The basic idea of the CSP algorithm is to find a spatial filter that maximizes this distance in the motor imagery EEG four-category task. Similarly, the FBCSP method is an extension of the CSP technique and is often used in EEG decoding tasks. The algorithm selects and classifies features by extracting the optimal spatial features through a set of bandpass filters. At the same time, in past motor imagery EEG research, classification methods such as linear classifiers, support vector machines, multi-layer perceptrons, and random forests are often used. Although these methods have achieved good results in motor imagery EEG decoding tasks, they divide feature extraction and classification into two stages.
然而,传统机器学习算法因需要手动提取特征导致的偏差有其在分类精度方面的缺陷。随着深度学习方法在各个领域的广泛应用并且可以高效地提取更加有意义的特征而获得较好的效果,卷积神经网络在脑电分类方面也应运而生。与此相反地,深度学习将特征提取和分类合并成一个步骤。根据脑电信号的时间序列性的特点,长短时期记忆网络有着提取时间特征的能力,尤其是在处理时间序列方面占据独到的优势并且已经在语音识别以及自然语言处理领域有着广泛的理解。通过在单元结构中引进门函数,LSTM可以解决普通RNN由于相关的输入信息过大而不能学习到的数据的麻烦。 LSTM结构包含存储单元和非线性门控单元两个重要的分支。目前,已经有研究尝试采用BLSTM记忆两个通道在特定时间内的关系变化来提取时域特征并获得了良好的分类结果。惊奇地是,实验表明卷积神经网络可以更好地从运动想象脑电中提取时空域以及频域特征。值得注意地是,在已经出现的研究贡献中,大多数学者们使用CNNs或者融合模型来提取运动想象脑电信号,但是并未考虑到资源的消耗以及模型计算复杂度,尤其在复杂度较高的融合模型中占用了较多的可训练参数,并且他们的方案中未对提取到的特征以及卷积核输出进行可视化。However, traditional machine learning algorithms have shortcomings in classification accuracy due to the bias caused by the need to manually extract features. With the wide application of deep learning methods in various fields and the ability to efficiently extract more meaningful features and achieve better results, convolutional neural networks have also emerged in EEG classification. In contrast, deep learning combines feature extraction and classification into a single step. According to the characteristics of the time series of EEG signals, the long-short-term memory network has the ability to extract time features, especially in the processing of time series, which has a unique advantage and has been extensively understood in the fields of speech recognition and natural language processing. By introducing a gate function in the unit structure, LSTM can solve the problem of data that ordinary RNN cannot learn due to too much relevant input information. The LSTM structure contains two important branches of storage unit and nonlinear gating unit. At present, some researches have tried to use BLSTM to memorize the relationship changes of two channels in a specific time to extract time-domain features and obtain good classification results. Surprisingly, experiments show that convolutional neural networks can better extract spatiotemporal and frequency domain features from motor imagery EEG. It is worth noting that in the research contributions that have appeared, most scholars use CNNs or fusion models to extract motor imagery EEG signals, but do not consider resource consumption and model computational complexity, especially when the complexity is high The fusion model of . takes up more trainable parameters, and their scheme does not visualize the extracted features and convolution kernel output.
发明内容Contents of the invention
本发明的目的在于,提供一种轻量快速的运动想象脑电信号解码方法。本发明可以以较少的可训练参数量获得更好地解码性能,在分类精度与模型复杂度之间保持了相对的平衡。The object of the present invention is to provide a lightweight and fast motor imagery EEG signal decoding method. The present invention can obtain better decoding performance with fewer trainable parameters, and maintains a relative balance between classification accuracy and model complexity.
本发明的技术方案:一种轻量快速的运动想象脑电信号解码方法,按如下步骤进行:Technical solution of the present invention: a light-weight and fast motor imagery EEG signal decoding method is carried out as follows:
S1、构建深度学习模型,所述深度学习模型包括时空卷积模块、池化模块和全连接模块;所述时空卷积模块由用于减少可训练参数量的时间卷积层和用于减少通道连接的空间深度卷积层构成;所述池化模块是堆叠池化层以降低模型的维度并减少复杂度;所述全连接模块用于最终的分类;S1, build a deep learning model, the deep learning model includes a spatiotemporal convolution module, a pooling module and a fully connected module; the spatiotemporal convolution module is composed of a temporal convolution layer for reducing the amount of trainable parameters and a channel for reducing A connected spatial depth convolutional layer is formed; the pooling module is a stacked pooling layer to reduce the dimension of the model and reduce the complexity; the fully connected module is used for the final classification;
S2、对原始脑电信号进行预处理,然后利用深度学习模型对预处理后的脑电信号进行分类解码。S2. Preprocessing the original EEG signal, and then using the deep learning model to classify and decode the preprocessed EEG signal.
上述的轻量快速的运动想象脑电信号解码方法,所述时空卷积模块用于提取脑电 输入的空间和光谱特征;所述时间卷积层通过一个参数化函数定义时间卷积核的内核值, 使得时间卷积层的内核值描述属于时间滤波器的子集,以减少可训练参数量,降低资源消 耗,其中脑电信号与第个时间卷积核之间的一维卷积公式如下: In the light-weight and fast motor imagery EEG signal decoding method described above, the spatio-temporal convolution module is used to extract the spatial and spectral features of the EEG input; the temporal convolution layer defines the kernel of the temporal convolution kernel through a parameterized function value, so that the kernel value description of the temporal convolutional layer belongs to the subset of the temporal filter, so as to reduce the amount of trainable parameters and reduce resource consumption, where the EEG signal and the first The one-dimensional convolution formula between two temporal convolution kernels is as follows:
; ;
式中,是第个电极信号与第个时间卷积核之间的一维卷积;是总数;表示第个电极信号,代表一维卷积沿着时间维度的过滤器的长度,对应着时间卷积 核的数量;是中间量; In the formula, is the first The first electrode signal and the first One-dimensional convolution between temporal convolution kernels; is the total; Indicates the first electrode signal, Represents the length of the filter of the one-dimensional convolution along the time dimension, Corresponding to the number of temporal convolution kernels; is the intermediate quantity;
通过参数化函数定义时间卷积核的内核值,频域范围内带通滤波的振幅表达 为: by parameterizing the function Define the kernel value of the time convolution kernel, and the amplitude of the bandpass filter in the frequency domain is expressed as:
; ;
式中:是频率;是第个带通滤波的截止频率的劣值;是是第个带通滤波 的截止频率的优值; In the formula: is the frequency; is the first The inferior value of the cutoff frequency of a bandpass filter; is the first The optimal value of the cutoff frequency of a bandpass filter;
时域中带通滤波的振幅如下:The magnitude of bandpass filtering in the time domain is as follows:
; ;
式中:为正弦函数; In the formula: is a sine function;
在时间卷积层之后堆叠过滤器大小为(1,65)以及32个带通时间过滤器的深度空间卷积层,用以学习光谱和空间特征。A deep spatial convolutional layer with filter size (1, 65) and 32 bandpass temporal filters is stacked after the temporal convolutional layer to learn spectral and spatial features.
前述的轻量快速的运动想象脑电信号解码方法,所述池化模块以(1,109)的大小与(1,23)的步长的将采样信号作为输入并且调查特定的超参数,且利用扁平层来转换池化模块的输出并连接到全连接层。In the aforementioned lightweight and fast motor imagery EEG signal decoding method, the pooling module takes the sampled signal as input with a size of (1, 109) and a step size of (1, 23) and investigates specific hyperparameters, and The output of the pooling module is transformed using a flattened layer and connected to a fully connected layer.
前述的轻量快速的运动想象脑电信号解码方法,所述全连接模块使用softmax函 数进行分类,利用监督学习算法训练超参数,产生运动想象四分类所对应的真实输出值;通 过函数映射到真实的类以表示卷积的处理过程,函数中代表 电极数;为时间步长,为总的输入数据;由softmax函数将输出转换成给定输出的每个 类标签的特定受试者条件概率表达为: In the aforementioned light-weight and fast motor imagery EEG signal decoding method, the fully connected module uses the softmax function to classify, uses the supervised learning algorithm to train hyperparameters, and generates the real output values corresponding to the four classifications of motor imagery; through the function Mapped to a real class to represent the processing of convolution, in the function represents the number of electrodes; is the time step, is the total input data; the output transformed by the softmax function into the subject-specific conditional probability of each class label of the given output is expressed as:
式中:是标签的条件概率;为第个试验的输入数据,;表示 权重与偏置函数的参数;为映射函数; In the formula: is the conditional probability of the label; for the first The input data of a test, ; Indicates the parameters of the weight and bias functions; is the mapping function;
通过为正确输出的标签分配高概率来最小化每个样本损失之和,该计算过程公式展示为:The sum of per-sample losses is minimized by assigning a high probability to the correctly output label, which is shown as:
; ;
式中:为每次试验的输出类,的值由负对数可能性决定,其具体计算表达式 如下: In the formula: is the output class for each trial, The value of is determined by the negative logarithmic likelihood, and its specific calculation expression is as follows:
。 .
前述的轻量快速的运动想象脑电信号解码方法,步骤S2中,对原始脑电信号进行处理是将原始脑电信号转换为跨时间样本C*T矩阵的通道,然后在8Hz和30Hz之间应用带通滤波器滤波,然后对滤波后的脑电信号进行批量归一化处理。In the aforementioned light-weight and fast motor imagery EEG signal decoding method, in step S2, the processing of the original EEG signal is to convert the original EEG signal into a channel of the cross-time sample C*T matrix, and then between 8Hz and 30Hz Filter with a bandpass filter, and then perform batch normalization on the filtered EEG signals.
前述的轻量快速的运动想象脑电信号解码方法,所述原始脑电信号表示为:The aforementioned lightweight and fast motor imagery EEG signal decoding method, the original EEG signal is expressed as:
; ;
式中:代表第个记录的试验总数,运动想象信号的第次试验与输出类紧密相 连;对于一个给定的试验,第次输入试验的输出映射为输出类标签,以分别指代运动想 象四分类任务。 In the formula: On behalf of The total number of recorded trials, the first number of motor imagery signals Trials are closely linked to the output class; for a given trial , No. The output of the input trials is mapped to the output class label , to refer to the four categories of motor imagery tasks.
前述的轻量快速的运动想象脑电信号解码方法,利用BrainDecode框架,实现深度学习模型的训练;将原始脑电信号的划分成750个样本作为深度学习模型的输入,起初每次batch的值设置为58,当epochs执行标准超过预设最大值时,训练终止;同时结合使用早期停止策略与Adam优化器来提升分类精度。The aforementioned lightweight and fast motor imagery EEG signal decoding method uses the BrainDecode framework to realize the training of the deep learning model; the original EEG signal is divided into 750 samples as the input of the deep learning model, and the value of each batch is initially set When the epochs execution standard exceeds the preset maximum value, the training is terminated; at the same time, the early stopping strategy is combined with the Adam optimizer to improve the classification accuracy.
前述的轻量快速的运动想象脑电信号解码方法,利用批量归一化正则化方法防止过拟合现象发生。The aforementioned light-weight and fast motor imagery EEG signal decoding method uses batch normalization regularization method to prevent over-fitting phenomenon.
与现有技术相比,本发明提出了展示了一种高效的用于运动想象脑电解码的深度学习模块,该深度学习模块可以以较少的可训练参数量获得更好地解码性能,在分类精度与模型复杂度之间保持了相对的平衡。本发明将可变参数纳入考虑范围之内极大地提高了运动想象脑电解码精度。根据脑电特定受试者的实验结果论证,本发明获得了78.42%左右的平均分类精度,比其他很多SOA算法有更强的鲁棒性。此外,本发明可以解决在神经科学中基于脑电图的普遍问题,为建立实际临床应用铺平了道路。Compared with the prior art, the present invention presents an efficient deep learning module for motor imagery EEG decoding, which can obtain better decoding performance with fewer trainable parameters. A relative balance is maintained between classification accuracy and model complexity. The invention greatly improves the motor imagery EEG decoding accuracy by taking variable parameters into consideration. According to the experimental results of specific EEG subjects, the present invention obtains an average classification accuracy of about 78.42%, which is more robust than many other SOA algorithms. Furthermore, the present invention can solve common problems based on EEG in neuroscience, paving the way for establishing practical clinical applications.
附图说明Description of drawings
图1为本发明的深度学习模型示意图;Fig. 1 is a schematic diagram of a deep learning model of the present invention;
图2为本发明的流程示意图;Fig. 2 is a schematic flow sheet of the present invention;
图3为数据集BCI竞赛2a的训练集与测试集单次实验范式; Figure 3 is the data set BCI competition 2a training set and test set single experiment paradigm;
图4为FBCSP算法在BCI竞赛2a数据集上的所有受试者之间的不同输出类的平 均混淆矩阵; Figure 4 shows the FBCSP algorithm in the BCI competition Average confusion matrix for different output classes across all subjects on the 2a dataset;
图5为Mixed-ConvNet模型在BCI竞赛2a数据集上的所有受试者之间的不同输 出类的平均混淆矩阵; Figure 5 shows the Mixed-ConvNet model in the BCI competition Average confusion matrix for different output classes across all subjects on the 2a dataset;
图6是在BCI竞赛2a数据集上不同模型之间的解码精度与标准误差的性能比 较; Figure 6 is in the BCI competition Performance comparison of decoding accuracy and standard error between different models on the 2a dataset;
图7是本发明的深度模型与其他两种SOA算法在不同的性能衡量指标之间的对比结果展示。Fig. 7 is a comparison result between the depth model of the present invention and other two SOA algorithms in different performance measurement indexes.
具体实施方式Detailed ways
下面实施例对本发明作进一步的说明,但并不作为对本发明限制的依据。The following examples further illustrate the present invention, but are not as the basis for limiting the present invention.
实施例1:一种轻量快速的运动想象脑电信号解码方法,如图1所示,按如下步骤进行:Embodiment 1: A light-weight and fast motor imagery EEG signal decoding method, as shown in Figure 1, is carried out as follows:
S1、获取原始脑电信号,构建深度学习模型,所述深度学习模型包括时空卷积模块、池化模块和全连接模块;所述时空卷积模块由用于减少可训练参数量的时间卷积层和用于减少通道连接的空间深度卷积层构成;所述池化模块是堆叠池化层以降低模型的维度并减少复杂度;所述全连接模块用于最终的分类;S1. Obtain the original EEG signal and build a deep learning model. The deep learning model includes a spatiotemporal convolution module, a pooling module, and a fully connected module; the spatiotemporal convolution module is composed of temporal convolutions used to reduce the amount of trainable parameters layer and a spatial depth convolution layer for reducing channel connections; the pooling module is a stacked pooling layer to reduce the dimension of the model and reduce complexity; the fully connected module is used for the final classification;
S2、对原始脑电信号进行预处理,然后利用深度学习模型对预处理后的脑电信号进行分类解码。S2. Preprocessing the original EEG signal, and then using the deep learning model to classify and decode the preprocessed EEG signal.
具体的,所述原始脑电信号表示为:Specifically, the original EEG signal is expressed as:
; ;
式中:代表第个记录的试验总数,运动想象信号的第次试验与输出类紧密相 连;对于一个给定的试验,;其中代表电极数;为时间步长,为总 的输入数据,第次输入试验的输出映射为输出类标签,以分别指代运动想象四分类任 务。 In the formula: On behalf of The total number of recorded trials, the first number of motor imagery signals Trials are closely linked to the output class; for a given trial , ;in represents the number of electrodes; is the time step, For the total input data, the The output of the input trials is mapped to the output class label , to refer to the four categories of motor imagery tasks.
如图2所示,对原始脑电信号进行预处理是将原始脑电信号转换为跨时间样本C*T矩阵的通道,然后在8Hz和30Hz之间应用带通滤波器滤波,然后对滤波后的脑电信号进行批量归一化处理。在预处理过程中,为原始脑电信号过滤了噪声与伪影,剔除了三个EOG坏通道。在本发明中,从[0,38]或者[4,38]Hz的感兴趣的频带的均匀分布中随机初始化,这些频率在此范围内更新变化。As shown in Figure 2, the preprocessing of the original EEG signal is to convert the original EEG signal into a channel of a C*T matrix of samples across time, and then apply a bandpass filter between 8Hz and 30Hz, and then filter the filtered The EEG signals were batch normalized. In the preprocessing process, noise and artifacts are filtered for the original EEG signal, and three EOG bad channels are eliminated. In the present invention, random initialization is performed from a uniform distribution of frequency bands of interest in [0,38] or [4,38] Hz, and these frequencies are updated within this range.
本实施例中,所述深度学习模型每一层所使用的全局超参数,输出形状、参数的数量以及激活函数的详细信息描述在表格1-3中。In this embodiment, the global hyperparameters used by each layer of the deep learning model, the output shape, the number of parameters and the detailed information of the activation function are described in Table 1-3.
表1为深度学习模型的输入以及时空卷积模块Table 1 is the input of the deep learning model and the spatio-temporal convolution module
表2为池化模块Table 2 is the pooling module
表3为全连接模块Table 3 is the fully connected module
表1-3表示了本发明深度学习模型的详细超参数。表1-3重点展示了本发明巧妙应 用时间正弦卷积层模型的主要模块以及每个模块使用的卷积或者池化层,每一层的全局超 参数,输出形状,每一层的可训练参数以及采用的激活函数。其中C表示通道数量,T表示样 本数量,K和F指代卷积核的数量和大小,S和P则代表步长以及池化大小,D表示深度卷积层 中的multiplier,m代表使用BatchNorm正则化方法的参数值,以及表示混合卷积神经网 络的dropout rate. 而且,意味着EEG信号特征分类的数量。与此同时,采样频率为 250Hz,实验还使用了ELU指数线性单元激活函数。 Tables 1-3 have shown the detailed hyperparameters of the deep learning model of the present invention. Tables 1-3 focus on the main modules of the present invention's ingenious application of the time sinusoidal convolution layer model, the convolution or pooling layers used by each module, the global hyperparameters of each layer, the output shape, and the trainable parameters of each layer. parameters and the activation function used. Where C represents the number of channels, T represents the number of samples, K and F represent the number and size of convolution kernels, S and P represent the step size and pooling size, D represents the multiplier in the depth convolution layer, and m represents the use of BatchNorm parameter values for the regularization method, and Indicates the dropout rate of the hybrid convolutional neural network. And, means the number of EEG signal feature categories. At the same time, the sampling frequency is 250Hz, and the experiment also uses the ELU exponential linear unit activation function.
在脑电信号与处理领域,经常使用CNNs处理高纬度数据的问题,卷积窗口即内核大小的所有的神经元都会给定偏置值与权重。与之替代地是,CNNs巧妙地深化了权重共享理念,学习一组权重与应用于隐藏层神经元的单一偏置值,此过程的数学方法表示如下:In the field of EEG signals and processing, CNNs are often used to deal with high-latitude data problems, and all neurons of the convolution window, that is, the kernel size, are given bias values and weights. Instead, CNNs subtly deepen the idea of weight sharing, learning a set of weights and a single bias value to apply to hidden layer neurons, the mathematical method of this process is expressed as follows:
; ;
其中,表示隐藏层中第个过滤器的第个神经元的激活输出,对应着使用的 激活函数,指代过滤器的共享偏置值,是内核大小,是共享权重向量,意旨预知 神经元的输出向量,表示转置操作。 in, represents the hidden layer filter's The activation output of a neuron, Corresponding to the activation function used, Referral filter The shared bias value of , is the kernel size, is the shared weight vector, Intention to predict the output vector of the neuron, Represents a transpose operation.
为了提取脑电输入的空间和光谱特征,分别设计了时间与空间卷积层。所述时间 卷积层通过一个参数化函数定义时间卷积核的内核值,使得时间卷积层的内核值描述属于 时间滤波器的子集,以减少可训练参数量,降低资源消耗,其中脑电信号与第个时间卷积 核之间的一维卷积公式如下: In order to extract the spatial and spectral features of the EEG input, temporal and spatial convolutional layers are designed respectively. The temporal convolution layer defines the kernel value of the temporal convolution kernel through a parameterized function, so that the kernel value description of the temporal convolution layer belongs to a subset of temporal filters, so as to reduce the amount of trainable parameters and reduce resource consumption. electrical signal and the The one-dimensional convolution formula between two temporal convolution kernels is as follows:
; ;
式中,是第个电极信号与第个时间卷积核之间的一维卷积;是总数;表示第个电极信号,代表一维卷积沿着时间维度的过滤器的长度,对应着时间卷积 核的数量;是中间量; In the formula, is the first The first electrode signal and the first One-dimensional convolution between temporal convolution kernels; is the total; Indicates the first electrode signal, Represents the length of the filter of the one-dimensional convolution along the time dimension, Corresponding to the number of temporal convolution kernels; is the intermediate quantity;
通过参数化函数定义时间卷积核的内核值,为了描述频域范围内的带通滤波, 其振幅表达为: by parameterizing the function Define the kernel value of the time convolution kernel. In order to describe the bandpass filtering in the frequency domain, its amplitude is expressed as:
; ;
式中:是频率;是第个带通滤波的截止频率的劣值;是是第个带通滤波 的截止频率的优值; In the formula: is the frequency; is the first The inferior value of the cutoff frequency of a bandpass filter; is the first The optimal value of the cutoff frequency of a bandpass filter;
此计算过程减少了时间卷积层每一个内核的可训练参数量在时域中,同理,时域中带通滤波的振幅如下:This calculation process reduces the amount of trainable parameters for each kernel of the time convolutional layer in the time domain. Similarly, the amplitude of the bandpass filter in the time domain is as follows:
; ;
式中:为正弦函数; In the formula: is a sine function;
在时间卷积层之后堆叠过滤器大小为(1,65)以及32个带通时间过滤器的深度空间卷积层,用以学习光谱和空间特征。为了防止过拟合的发生,本发明灵活地将批量归一化正则化技术运用于特征图维度。A deep spatial convolutional layer with filter size (1, 65) and 32 bandpass temporal filters is stacked after the temporal convolutional layer to learn spectral and spatial features. In order to prevent the occurrence of overfitting, the present invention flexibly applies the batch normalization regularization technique to the feature map dimension.
时空卷积模块的输出作为池化模块的输入,池化层在保留原始有意义的信息的前提下减少了每个特征图的维度,空间池化可以划分为子采样与降采样几种类型,其中最著名的是最大池化与平均池化方法。本发明在每一个卷积层之后设置了批量归一化技术以及在池化层之后应用了Dropout方法。在平均池化模块中,以(1,109)的大小与(1,23)的步长的将采样信号作为输入并且调查特定的超参数。The output of the space-time convolution module is used as the input of the pooling module. The pooling layer reduces the dimension of each feature map while retaining the original meaningful information. The spatial pooling can be divided into several types of subsampling and downsampling. The most famous of these are the max pooling and average pooling methods. The present invention sets a batch normalization technique after each convolutional layer and applies a Dropout method after a pooling layer. In the average pooling module, we sample the signal with a size of (1, 109) and a stride of (1, 23) as input and investigate specific hyperparameters.
最后,深度学习模型还充分利用了扁平层来转换上述模块的输出并连接到全连接层,导致生成一个提取到的一维特征向量。所述全连接模块使用softmax函数进行分类,利用监督学习算法训练超参数,产生运动想象四分类所对应的真实输出值;通过函数映射到真实的类以表示卷积的处理过程,由softmax函数将输出转换成给定输出的每个类标签的特定受试者条件概率表达为:Finally, the deep learning model also makes full use of the flattening layer to transform the output of the above modules and connect to the fully connected layer, resulting in an extracted 1D feature vector. The fully connected module uses the softmax function to classify, utilizes a supervised learning algorithm to train hyperparameters, and produces the real output values corresponding to the four classifications of motor imagery; through the function mapping to the real class to represent the convolution process, the softmax function will The output transforms into the subject-specific conditional probability of each class label for a given output expressed as:
式中:是标签的条件概率;为第个试验的输入数据,;表示权重与偏置函数 的参数,是指数函数;为映射函数;代表电极数;为时间步长,为总的输入数据 In the formula: is the conditional probability of the label; for the first The input data of a test,; Indicates the parameters of the weight and bias functions, is an exponential function; is the mapping function; represents the number of electrodes; is the time step, For the total input data
通过为正确输出的标签分配高概率来最小化每个样本损失之和,该计算过程公式展示为:The sum of per-sample losses is minimized by assigning a high probability to the correctly output label, which is shown as:
; ;
式中:为每次试验的输出类,的值由负对数可能性决定,其具体计算表达式 如下: In the formula: is the output class for each trial, The value of is determined by the negative logarithmic likelihood, and its specific calculation expression is as follows:
。 .
本发明巧妙地借助于Schirrmeister提出的完全用于处理运动想象脑电分类任务 问题的BrainDecode框架,实现Mixed-ConvNet的训练。由于脑电受试者之间的差异性并且 随着时间的变化而变化,因而本发明设置了一系列可变的超参数,为每位受试者搜寻最佳 的参数设定值。将原始22个脑电通道信号的划分的750个样本作为模型的输入,起初每次 batch的值设置为58,当epochs执行标准超过预设最大值时,训练终止。同时结合使用早期 停止策略与Adam优化器来提升模型的分类精度,在感兴趣的分布中随机选取低频为0或者 4Hz的频带,为了使得模型训练更加稳定可靠,批量归一化正则化技术是防止过拟合现象发 生的关键所在并且将其值设定为0.99。以原始脑电信号与预处理过的脑电信号为例, 当与分别是矩阵和的第i行,那么归一化操作显而易见地由如下公式表示:The present invention skillfully utilizes the BrainDecode framework proposed by Schirrmeister, which is completely used for processing motor imagery EEG classification tasks, to realize the training of Mixed-ConvNet. Since the EEG varies among subjects and changes with time, the present invention sets a series of variable hyperparameters to search for the best parameter setting value for each subject. The 750 samples of the original 22 EEG channel signals are used as the input of the model. At first, the value of each batch is set to 58. When the epochs execution standard exceeds the preset maximum value, the training is terminated. At the same time, the early stopping strategy is combined with the Adam optimizer to improve the classification accuracy of the model, and the frequency band with a low frequency of 0 or 4 Hz is randomly selected in the distribution of interest. In order to make the model training more stable and reliable, the batch normalization regularization technology is to prevent The key to the overfitting phenomenon and set its value to 0.99. raw EEG signal with preprocessed EEG signals For example, when and are the matrices and The i-th line of , then the normalization operation is obviously expressed by the following formula:
; ;
其中,与分别代表平均值与标准差操作。当正则化完输出结果之后, 巧妙运用激活函数并且本研究使用的函数为指数线性单元与线性激活函数。其中,ELU的计 算公式可以表达为: in, and represent the mean and standard deviation operations, respectively. After regularizing the output results, the activation function is cleverly used and the functions used in this study are exponential linear units and linear activation functions. Among them, the calculation formula of ELU can be expressed as:
(9) (9)
Dropout技术在每次训练更新期间随机将前一层的输出置为0并且此模型在平均池化层之后设置的Dropout率为0.5。The dropout technique randomly sets the output of the previous layer to 0 during each training update and this model sets a dropout rate of 0.5 after the average pooling layer.
为了验证本发明的分类解码的准确性,本发明在数据集BCI竞赛 2a上与一些SOA算法的比较实验结果以及对卷积核本身进行可视化、可视化卷积后的结果以帮助理解卷积核的作用、通过热度图深刻鉴别图像分类问题中哪些部分的关键作用即类激活可视化以及隐藏层与特征可视化。In order to verify the accuracy of the classification decoding of the present invention, the present invention compares the experimental results with some SOA algorithms on the data set BCI competition 2a and visualizes the convolution kernel itself, and visualizes the results after convolution to help understand the convolution kernel. The key role of deeply identifying which parts of the image classification problem through the heat map is the visualization of class activation and visualization of hidden layers and features.
1、数据集描述1. Dataset description
作为运动想象脑电分类领域最为常见的数据集之一,BCI竞赛 2a是由格拉茨大学在采样频率为250Hz使用22个脑电电极(不包括3个EOG通道)从9个受试者实验收集而成的公开可获得数据集。特别提醒,在本发明实验研究中,由于此原始数据集的第4个受试者的训练数据缺失了部分试验样本,因此采用了除去第4个受试者以外的其他八个受试者进行实验。剔除三个额外的EOG通道之后,使用带通滤波器在0.5到100Hz过滤噪声与伪影。该数据集的受试者执行左手、右手、双脚和舌头运动四种不同的运动想象任务,训练集与测试集分别在不同的两天完成实验记录,每组实验包含288条试验次数,其中每类任务包含72次,运动想象的时间为4s,每次试验生成750个样本点,其实验范式如图3所示,包括“Beep”鸣声提示音、“Fixation cross”的2秒时间十字架、接下来的1.25秒左、右、上和下的“Cue”箭头提示(分别对应运动想象的左手、右手、双脚以及舌头类)、3到6秒为关键的“Motor imagery“运动想象时间以及最后6到7.5秒为十字架小时过后的“Break”短暂休息时间。由之前对运动想象相关数据的分类尝试研究表明,数据集的受试者个体性差异性比较大,衡量他们在模型上的解码效果也因人而异。而且,为了执行在优化过程中第一步的早期停止策略,训练集进一步划分成小部分训练集与验证集,其中验证集占总训练集的20%。As one of the most common data sets in the field of motor imagery EEG classification, the BCI competition 2a was collected from 9 subjects experimentally by the University of Graz using 22 EEG electrodes (excluding 3 EOG channels) at a sampling frequency of 250Hz A publicly available dataset. Special reminder, in the experimental research of the present invention, because the training data of the 4th subject of this original data set is missing part of the test samples, so the other eight subjects except the 4th subject are used to carry out experiment. After removing the three additional EOG channels, noise and artifacts were filtered using a bandpass filter at 0.5 to 100 Hz. The subjects in this data set performed four different motor imagery tasks of left hand, right hand, feet and tongue movement. The training set and the test set completed the experimental records on two different days. Each set of experiments contained 288 trials, of which Each type of task contains 72 times, the time of motor imagery is 4s, and each trial generates 750 sample points. The experimental paradigm is shown in Figure 3, including the "Beep" sound and the 2-second time cross of "Fixation cross". , the next 1.25 seconds left, right, up and down "Cue" arrow prompts (respectively corresponding to the left hand, right hand, feet and tongue of motor imagery), 3 to 6 seconds is the key "Motor imagery" motor imagery time And the last 6 to 7.5 seconds is the "Break" short rest period after the cross hour. Previous attempts to classify data related to motor imagery have shown that the individual subjects of the data set have relatively large individual differences, and the decoding effect of measuring them on the model also varies from person to person. Moreover, in order to implement the early stopping strategy in the first step of the optimization process, the training set is further divided into a small training set and a validation set, where the validation set accounts for 20% of the total training set.
2、性能评价指标2. Performance evaluation index
为了评价提出的方法的解码性能,保留了测试数据并且对带有标签的数据进行真实预测。实验使用如下几种性能指标评价提出的模型,其中分类精度是最为频繁采纳的衡量指标,还有预测值与查全率也是鉴别算法结构好与坏的重要方法,这三种常见的指标公式如下: To evaluate the decoding performance of the proposed method, the test data is retained and ground truth predictions are made on the labeled data. The experiment uses the following performance indicators to evaluate the proposed model, among which the classification accuracy is the most frequently adopted measurement indicator, and the prediction value and recall rate are also important methods to identify good or bad algorithm structures. The formulas of these three common indicators are as follows :
; ;
; ;
; ;
其中是真正例,是真负例,是假正例,是假负例。除此以外,科恩卡 kappa值也是较常使用的计算指标,其公式展示如下: in is a real example, is a true negative example, is a false positive, is a false negative. In addition, Konka's kappa value is also a commonly used calculation indicator, and its formula is shown as follows:
; ;
其中,代表观察到的一致性的比例精度,暗指随机猜测值的概率或者精度。分数值是最终经常采纳的指标,需要结合预测值与查全率两者共同计算得知,具体公式 表达如下: in, represents the scaled precision of the observed agreement, The probability or precision of a random guess value is implied. The score value is the index that is often adopted in the end. It needs to be calculated by combining the predicted value and the recall rate. The specific formula is expressed as follows:
。 .
3.比较结果3. Compare the results
本发明提出的深度学习模型与同类SOA算法在数据集BCI竞赛2a上相比较的实 验对比结果详细记录在表4中(忽略了部分训练数据缺失的受试者4)。 The deep learning model proposed by the present invention competes with similar SOA algorithms in the data set BCI competition The results of the experimental comparisons on 2a are detailed in Table 4 (subject 4 with missing part of the training data is ignored).
表4Table 4
为了提高说服力,实验选取了传统机器学习算法FBCSP以及深度学习ShallowConvNet、C2CM、DeepConvNet、WaSFConvNet、CM- -ConvNet、AMSI-ConvNet、Proposed-model模型,在此四分类运动想象脑电数据集上实施了一系列对比实验。滤波器组共同空间模式是最为经典的运动想象脑电传统手工特征提取算法,同时结合线性判别分析分类器完成四分类数据集任务,在此数据集上获得了68.59%的解码精度。图4和图5分别给定了FBCSP算法与本发明各自的所有受试者之间的不同输出类的平均混淆矩阵,图中都展示了“Left”和“Right”运动想象任务的平均分类精度比“Foot”和“Tongue”二者分类要高。值得注意地是,提出的Mixed-ConvNet模型在四个运动想象分类方面都比FBCSP+rLDA结构要高,尤其在“Foot”与“Tongue”二分类上都提取了为解码更明显可微的脑电特征。重点关注地是,此结构在“Tongue”分类方面比FBCSP传统算法足足高出了24%左右并且达到了80.4%的平均解码精度,以此证明提出的Mixed-ConvNet解码模型提取特征的效率之高。In order to improve persuasiveness, the experiment selected the traditional machine learning algorithm FBCSP and deep learning ShallowConvNet, C2CM, DeepConvNet, WaSFConvNet, CM-ConvNet, AMSI-ConvNet, Proposed-model models, and implemented them on this four-category motor imagery EEG dataset A series of comparative experiments were carried out. The filter bank common spatial mode is the most classic traditional manual feature extraction algorithm of motor imagery EEG, and combined with the linear discriminant analysis classifier to complete the task of the four-category data set, the decoding accuracy of 68.59% was obtained on this data set. Fig. 4 and Fig. 5 have respectively given the average confusion matrix of the different output classes between FBCSP algorithm and all subjects of the present invention respectively, all have demonstrated the average classification accuracy of " Left " and " Right " motor imagery task in the figure Classified higher than both "Foot" and "Tongue". It is worth noting that the proposed Mixed-ConvNet model is higher than the FBCSP+rLDA structure in the four categories of motor imagery, especially in the two categories of "Foot" and "Tongue". electrical characteristics. It is important to note that this structure is about 24% higher than the traditional FBCSP algorithm in terms of "Tongue" classification and achieves an average decoding accuracy of 80.4%, which proves that the proposed Mixed-ConvNet decoding model is highly efficient in feature extraction. high.
此外,从表格4中惊奇发现,运动想象脑电领域最广泛使用的浅层模型Shallow-ConvNet与大多数二维卷积结构相类似的二维卷积场景模型C2CM分别在此数据集上获取了74.05%与75.43%的平均分类准确率。同时,由Schirrmeister et al.搭建的Deep-ConvNet模型是在浅层结构的基础上又堆叠了几层卷积与池化层,相反地,不但增加了模型可训练参数量而且与Shallow-ConvNet相比较,平均分类精度降低了4%左右。值得一提地是,利用小波核直接考虑了脑电信号的频谱功率调制的模型WaSF-ConvNet,该网络包含两个特定的卷积层、紧随其后的是池化层,加上dropout层与全连接层五个完整的部分,在学习特征过程中显著减少了参数量并且在此四分类脑电数据集上仅仅得到了68.96%相对较低的精度。更进一步地,AMSI-ConvNet结构提出了两种空时卷积模块的变体形式以验证解码性能的可行性,进一步地基准测试得知在此脑电竞赛数据集上获得了76.27%的平均分类精度。与其他baseline算法相比较,本发明提出的模型的平均解码精度比传统FBCSP足足高了10%左右并且其标准差值也是相对较小的,观察得知该结构具有更高的鲁棒性与稳定性。In addition, it is surprisingly found from Table 4 that Shallow-ConvNet, the most widely used shallow model in the field of motor imagery EEG, and the two-dimensional convolutional scene model C2CM, which is similar to most two-dimensional convolutional structures, were obtained on this data set. The average classification accuracies of 74.05% and 75.43% were achieved. At the same time, the Deep-ConvNet model built by Schirrmeister et al. stacks several layers of convolution and pooling layers on the basis of the shallow structure. On the contrary, it not only increases the number of trainable parameters of the model but also has the same In comparison, the average classification accuracy is reduced by about 4%. It is worth mentioning that the model WaSF-ConvNet, which uses the wavelet kernel to directly consider the spectral power modulation of the EEG signal, contains two specific convolutional layers, followed by a pooling layer, plus a dropout layer With five complete parts of the fully connected layer, the number of parameters is significantly reduced in the process of learning features and only a relatively low accuracy of 68.96% is obtained on this four-category EEG dataset. Furthermore, the AMSI-ConvNet structure proposes two variants of the space-time convolution module to verify the feasibility of the decoding performance. Further benchmark tests show that the average classification of 76.27% has been obtained on this EEG competition dataset. precision. Compared with other baseline algorithms, the average decoding accuracy of the model proposed by the present invention is about 10% higher than the traditional FBCSP and its standard deviation is relatively small. It is observed that the structure has higher robustness and stability.
进一步地,本发明提出的深度学习模型与其他baseline算法在BCI竞赛2a数据
集上的解码精度实验对比如图6所示,其中图6中反映此数据集上八个受试者之间的性能的
误差线代表标准差(Std)。图中观察发现,Mixed-ConvNet模型的性能优于其他SOA方法,通
过充分使用正弦时间卷积模块,此算法比Shallow-ConvNet与Deep-ConvNet模型分别高于
达到4%以及8%精度,并且此模型在该数据集上的标准差也小于其他SOA算法,实验论证表明
时空卷积模块的确提升了脑电解码性能。从单独受试者方向考虑模型的性能,展示的模型
与其他SOA算法相比较而言,特定受试者1、5和9在此结构上性能表现最佳,分别达到了
88.54%、71.87%以及85.76%平均解码精度,这完全可以验证出提出的结构足以运用于振荡
运动想象脑电解码。表3充分展示了BCI竞赛2a数据集上采用特定受试者方法的几种同类
SOA运动想象算法与提出的模型的每位受试者之间的kappa值结果比较。Further, the deep learning model proposed by the present invention competes with other baseline algorithms in the BCI competition The experimental comparison of decoding accuracy on the 2a dataset is shown in Fig. 6, where the error bars in Fig. 6 reflecting the performance among eight subjects on this dataset represent the standard deviation (Std). It is observed in the figure that the performance of the Mixed-ConvNet model is better than other SOA methods. By fully using the sinusoidal time convolution module, this algorithm is 4% and 8% higher than the Shallow-ConvNet and Deep-ConvNet models, respectively, and this The standard deviation of the model on this data set is also smaller than that of other SOA algorithms. The experimental demonstration shows that the spatio-temporal convolution module does improve the performance of EEG decoding. Considering the performance of the model from the perspective of individual subjects, compared with other SOA algorithms, the performance of
表5table 5
从表5中不难比较发现,提出的模型获得了最优的性能评价指标kappa值并且达到 了0.72,与其他传统算法或者浅层模型相比,在特定受试者2、3、6和9中获取了最高的kappa 值。在图7中展示了本发明提出的模型Mixed-ConvNet(深度学习模型)与传统算法FBCSP以 及典型的浅层结构Shallow-ConvNet三种模型分别在分类精度、kappa值以及F1分数值不同 性能衡量指标之间的对比结果(每个评价指标的柱状图中,从上到下依次为Shallow- ConvNet、FBCSP和Mixed-ConvNet),从图中不难辨别,本发明提出的模型在三者评价指标均 比其他SOA算法(即FBCSP)表现更佳。而且,借助于Precision和Recall值其他的性能衡量指 标以充分验证提出的模型的竞争性,表格6简单阐明了提出的模型与其他几个baseline算 法BCI竞赛2a数据集上的Precision、Recall值以及F1分数的结果比较: From Table 5, it is not difficult to find that the proposed model has obtained the optimal performance evaluation index kappa value and reached 0.72. The highest kappa value was obtained in . In Figure 7, the model Mixed-ConvNet (deep learning model) proposed by the present invention and the traditional algorithm FBCSP and the typical shallow structure Shallow-ConvNet three models are shown in different performance metrics of classification accuracy, kappa value and F1 score value. The comparison results between (the histograms of each evaluation index, from top to bottom are Shallow-ConvNet, FBCSP and Mixed-ConvNet), it is not difficult to distinguish from the figure, the model proposed by the present invention is in the three evaluation indexes. Performs better than other SOA algorithms (ie FBCSP). Moreover, with the help of Precision and Recall values and other performance metrics to fully verify the competitiveness of the proposed model, Table 6 briefly illustrates the proposed model's competition with several other baseline algorithms BCI Comparison of Precision, Recall values and F1 scores on the 2a dataset:
表6Table 6
仔细观察表格得知,提出的轻量级结构在这三者性能指标上的总体平均值均优于其他相比较的SOA算法。Careful observation of the table shows that the overall average of the proposed lightweight structure is better than other comparative SOA algorithms in these three performance indicators.
从模型的复杂性来考虑,本发明提出的深度学习结构与其他几个baseline算法使用的可训练参数量、训练时间以及平均解码精度之间在BCI竞赛 2a数据集上的详细对比结果如表7所示:Considering the complexity of the model, the detailed comparison results between the deep learning structure proposed by the present invention and the amount of trainable parameters, training time and average decoding accuracy used by several other baseline algorithms on the BCI competition 2a data set are shown in Table 7 Shown:
表7Table 7
从表7中可以看出,在对比的几个浅层以及深度融合算法中,结合振幅扰动数据增强方法的通道映射混合尺寸卷积神经网络使用了大量的可训练超参数且达到了8.36×105,模型结构比较复杂的前提下,其消耗了大量的资源增加了计算成本然而在此数据集上的平均解码精度并未超越提出的模型。此外,深度卷积模型Deep-ConvNet、经典的浅层结构Shallow-ConvNet以及混合堆叠网络AMSI-ConvNet在可训练参数量、受试者平均训练时间(hh:mm:ss)以及平均解码精度三者之间的平衡性也并没有提出的算法更佳。我们提出的模型仅仅引进了最少的可训练参数量8324并且达到了最高的平均分类精度,在此数据集上的八个受试者之间的平均训练时间仅仅花费了10分钟不到的时间,这足以证明提出的轻量级结构完全可以适用于真实脑机接口应用场景中。It can be seen from Table 7 that in the comparison of several shallow and deep fusion algorithms, the channel mapping mixed-size convolutional neural network combined with the amplitude perturbation data enhancement method uses a large number of trainable hyperparameters and reaches 8.36×10 5. Under the premise that the model structure is relatively complex, it consumes a lot of resources and increases the calculation cost. However, the average decoding accuracy on this data set does not exceed the proposed model. In addition, the deep convolutional model Deep-ConvNet, the classic shallow structure Shallow-ConvNet, and the hybrid stacked network AMSI-ConvNet have the best performance in terms of the number of trainable parameters, the average training time of subjects (hh:mm:ss) and the average decoding accuracy. The balance between the proposed algorithm is not better. Our proposed model only introduces the minimum number of trainable parameters 8324 and achieves the highest average classification accuracy. The average training time between eight subjects on this dataset takes less than 10 minutes. This is enough to prove that the proposed lightweight structure can be applied to real brain-computer interface application scenarios.
综上所述,本发明提出了展示了一种高效的用于运动想象脑电解码的深度学习模块,该深度学习模块可以以较少的可训练参数量获得更好地解码性能,在分类精度与模型复杂度之间保持了相对的平衡。本发明将可变参数纳入考虑范围之内极大地提高了运动想象脑电解码精度。根据脑电特定受试者的实验结果论证,本发明获得了78.42%左右的平均分类精度,比其他很多SOA算法有更强的鲁棒性。此外,本发明可以解决在神经科学中基于脑电图的普遍问题,为建立实际临床应用铺平了道路。To sum up, the present invention proposes and demonstrates an efficient deep learning module for motor imagery EEG decoding. The deep learning module can obtain better decoding performance with fewer trainable parameters, and the classification accuracy A relative balance is maintained between model complexity. The invention greatly improves the motor imagery EEG decoding accuracy by taking variable parameters into consideration. According to the experimental results of specific EEG subjects, the present invention obtains an average classification accuracy of about 78.42%, which is more robust than many other SOA algorithms. Furthermore, the present invention can solve common problems based on EEG in neuroscience, paving the way for establishing practical clinical applications.
Claims (8)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211087807.4A CN115316955A (en) | 2022-09-07 | 2022-09-07 | A lightweight and fast decoding method for motor imagery EEG signals |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211087807.4A CN115316955A (en) | 2022-09-07 | 2022-09-07 | A lightweight and fast decoding method for motor imagery EEG signals |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115316955A true CN115316955A (en) | 2022-11-11 |
Family
ID=83929661
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211087807.4A Withdrawn CN115316955A (en) | 2022-09-07 | 2022-09-07 | A lightweight and fast decoding method for motor imagery EEG signals |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115316955A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117390542A (en) * | 2023-10-13 | 2024-01-12 | 上海韶脑传感技术有限公司 | A cross-subject motor imagery EEG modeling method based on domain generalization |
-
2022
- 2022-09-07 CN CN202211087807.4A patent/CN115316955A/en not_active Withdrawn
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117390542A (en) * | 2023-10-13 | 2024-01-12 | 上海韶脑传感技术有限公司 | A cross-subject motor imagery EEG modeling method based on domain generalization |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Dissanayake et al. | Deep learning for patient-independent epileptic seizure prediction using scalp EEG signals | |
Yang et al. | Emotion recognition from multi-channel EEG through parallel convolutional recurrent neural network | |
CN113011239A (en) | Optimal narrow-band feature fusion-based motor imagery classification method | |
Pandya et al. | Unveiling the Power of Collective Intelligence: A Voting-Based Approach for Dementia Classification | |
Jinliang et al. | EEG emotion recognition based on granger causality and capsnet neural network | |
Li et al. | SSTD: a novel spatio-temporal demographic network for EEG-based emotion recognition | |
Zhou et al. | Multimodal ECG heartbeat classification method based on a convolutional neural network embedded with FCA | |
Zhang et al. | Research on lung sound classification model based on dual-channel CNN-LSTM algorithm | |
CN113128384A (en) | Brain-computer interface software key technical method of stroke rehabilitation system based on deep learning | |
Tao et al. | A novel feature relearning method for automatic sleep staging based on single-channel EEG | |
Cao et al. | IFBCLNet: Spatio-temporal frequency feature extraction-based MI-EEG classification convolutional network | |
Stuchi et al. | A frequency-domain approach with learnable filters for image classification | |
Isik | Heart disease prediction with feature selection based on metaheuristic optimization algorithms and electronic filter model | |
CN115316955A (en) | A lightweight and fast decoding method for motor imagery EEG signals | |
CN115633938A (en) | Electroencephalogram analysis method for epileptic syndrome of multitask depth network | |
CN113974627A (en) | An emotion recognition method based on brain-computer generative confrontation | |
CN117338313B (en) | Multi-dimensional feature EEG signal recognition method based on stacking integration technology | |
Liu et al. | A learnable front-end based efficient channel attention network for heart sound classification | |
CN118766459A (en) | A method for emotion recognition of EEG signals integrating multi-scale residual attention network | |
CN118557150A (en) | Method and system for epileptic seizure detection and automatic labeling based on EEG data | |
Tengshe et al. | Sickle cell anemia detection using convolutional neural network | |
CN116662782A (en) | MSFF-SENET-based motor imagery electroencephalogram decoding method | |
He et al. | A parallel neural networks for emotion recognition based on EEG signals | |
Parveen et al. | Eeg-based emotion classification-a theoretical perusal of deep learning methods | |
Manasa et al. | EEG signal-based classification of mental tasks using a one-dimensional ConvResT model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WW01 | Invention patent application withdrawn after publication | ||
WW01 | Invention patent application withdrawn after publication |
Application publication date: 20221111 |