[go: up one dir, main page]

Next Article in Journal
Free Space Measurement Device Independent Quantum Key Distribution with Modulating Retro-Reflectors under Correlated Turbulent Channel
Next Article in Special Issue
Local Integral Regression Network for Cell Nuclei Detection
Previous Article in Journal
Image Encryption Scheme Based on Multiscale Block Compressed Sensing and Markov Model
Previous Article in Special Issue
Entropy-Based Graph Clustering of PPI Networks for Predicting Overlapping Functional Modules of Proteins
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fatigue Detection with Spatial-Temporal Fusion Method on Covariance Manifolds of Electroencephalography

1
Gansu Provincial Key Laboratory of Wearable Computing, School of Information Science and Engineering, Lanzhou University, Lanzhou 730000, China
2
School of Optics and Electronics, Beijing Institute of Technology, Beijing 100081, China
3
CAS Center for Excellence in Brain Science and Institutes for Biological Sciences, Shanghai Institutes for Biological Sciences, Chinese Academy of Sciences, Shanghai 200031, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2021, 23(10), 1298; https://doi.org/10.3390/e23101298
Submission received: 13 August 2021 / Revised: 22 September 2021 / Accepted: 28 September 2021 / Published: 30 September 2021
(This article belongs to the Special Issue Networks and Systems in Bioinformatics)
Figure 1
<p>The framework of our model: In the SR domain, we fused the output of the SPD network and the distance information based on the Stein divergence as SR-domain features. In the TR domain: we fed the covariance matrices into a two-layer RNN with seven LSTM cells per layer, and define the output as TR-domain features. Then, the concatenated SR-domain and TR-domain features were transformed into linear features by the fully connected layer, and the features were used to give the final prediction by the softmax layer.</p> ">
Figure 2
<p>Architecture of the SPD network: The bilinear map layer consists of a transformation matrix <span class="html-italic">W</span> with full rank and its transpose. <math display="inline"><semantics> <mi mathvariant="normal">ζ</mi> </semantics></math> of the second layer indicates a diagonal matrix, whose diagonal elements are the threshold value we set. In the logarithmic eigenvalue layer, we apply eigendecomposition to get the eigenvalues and eigenvectors of matrix <math display="inline"><semantics> <msub> <mi>X</mi> <mi>i</mi> </msub> </semantics></math>, which are <math display="inline"><semantics> <msub> <mi mathvariant="normal">Λ</mi> <mi>i</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>U</mi> <mi>i</mi> </msub> </semantics></math>.</p> ">
Figure 3
<p>The scheme of EEG segmentation: (<b>a</b>) For time series of each trial, a 3 s sliding time window and a 1 s step are used to divide the EEG signals. (<b>b</b>) From the segmented result, the relations between different channels are captured by the covariance matrices we calculated, and we reshape them into vectors with one dimension, separately. The covariance value between the <span class="html-italic">i</span>-th channel and the <span class="html-italic">j</span>-th channel is indicated by the element <math display="inline"><semantics> <mrow> <mi>C</mi> <mo>(</mo> <mi>i</mi> <mo>,</mo> <mi>j</mi> <mo>)</mo> </mrow> </semantics></math> of the covariance matrix.</p> ">
Figure 4
<p>The features of the TR domain are processed by a 2-layer RNN with 7 LSTM units per layer, where <math display="inline"><semantics> <msub> <mi>m</mi> <mi>i</mi> </msub> </semantics></math> means the flattened matrix of the <span class="html-italic">i</span>-th segment signals, and <math display="inline"><semantics> <msub> <mi>h</mi> <mi>i</mi> </msub> </semantics></math> indicates the <span class="html-italic">i</span>-th hidden state. The <math display="inline"><semantics> <msubsup> <mi>h</mi> <mrow> <mn>7</mn> </mrow> <msup> <mrow/> <mo>′</mo> </msup> </msubsup> </semantics></math> denotes the output of the RNN.</p> ">
Figure 5
<p>Simulated driving paradigm: Each lane-departure event is considered as the period from the first response offset to the second response offset, which contains the deviation onset and response onset of the current event. The time between the deviation onset and the response onset is defined as reaction time(RT), which is the main focus of our study. EEG signals were recorded during the whole task.</p> ">
Figure 6
<p>Comparison of the average RTs of 27 subjects in different driving states. The histogram with blue color (V_RT) represents the reaction time of the vigilant state, while the orange one (F_RT) represents the reaction time of the fatigue state. The horizontal dotted lines indicate the average RT of all subjects in different states. Besides, we marked the variance of reaction time, which are 0.775 and 0.043 of fatigue state and vigilant state, respectively.The average reaction time distribution of 27 subjects in the vigilant and drowsy driving state. The green histogram (V_Res) represents the response time of vigilant driving, while the purple histogram (F_Res) represents the response time of drowsy driving. The red and black horizontal dotted lines represent the average response time across all subjects in drowsy and vigilant driving, respectively. Besides, the variance of response time in drowsy and vigilant driving is 0.775 and 0.043.</p> ">
Figure 7
<p>Scatter diagram of RT values of all trials: The horizontal axis represents the value of local reaction time, and the vertical axis represents the value of global reaction time; 739 trials were labeled as vigilant states, which were marked as purple points in the lower left corner; 694 orange points in the upper right corner indicated trials with fatigue state.</p> ">
Figure 8
<p>Trajectories of variations in the inter-channel relations over time in the situation of 3, 5, 7 channel pairs. In each graph (<b>a</b>–<b>f</b>), the horizontal axis represents the 7 successive time windows, and the vertical axis indicates the relation values between different EEG channels, which are the covariance of two channels. As seen, the curve shapes within the same class are very similar to each other, whereas significant dissimilarities exist between the different classes. For example, graphs (<b>a</b>,<b>d</b>) show the dynamic variations of the same combination of 3 EEG channel pairs in <math display="inline"><semantics> <msub> <mi>M</mi> <mi>V</mi> </msub> </semantics></math> and <math display="inline"><semantics> <msub> <mi>M</mi> <mi>D</mi> </msub> </semantics></math>, respectively.</p> ">
Figure 9
<p>The box plots of classification accuracy of different numbers of channel pairs, which are 3, 5, 7 and all pairs. The <span class="html-italic">p</span>-values of the significance test are marked below the box plots.</p> ">
Figure 10
<p>Diagram of brain state analysis based on covariance: (<b>a</b>) We assumed that the electrical activity is the matrix with the dimensions of <math display="inline"><semantics> <mrow> <mn>364</mn> <mo>×</mo> <mn>2250</mn> </mrow> </semantics></math>. (<b>b</b>) We calculated the weight matrix set <math display="inline"><semantics> <mi mathvariant="bold">A</mi> </semantics></math> of fatigue and vigilant states, respectively. The histograms of the third column represent the weight assignment of all channels by the same neuron in two brain states.</p> ">
Figure 11
<p>Fitted t-distribution diagrams of the weight assignment of all channels by the brain neurons, in which the blue curve denotes the fatigue state and the yellow curve denotes the vigilant state. The normalized KL divergences between two distributions are on the top of each diagram.</p> ">
Review Reports Versions Notes

Abstract

:
With the increasing pressure of current life, fatigue caused by high-pressure work has deeply affected people and even threatened their lives. In particular, fatigue driving has become a leading cause of traffic accidents and deaths. This paper investigates electroencephalography (EEG)-based fatigue detection for driving by mining the latent information through the spatial-temporal changes in the relations between EEG channels. First, EEG data are partitioned into several segments to calculate the covariance matrices of each segment, and then we feed these matrices into a recurrent neural network to obtain high-level temporal information. Second, the covariance matrices of whole signals are leveraged to extract two kinds of spatial features, which will be fused with temporal characteristics to obtain comprehensive spatial-temporal information. Experiments on an open benchmark showed that our method achieved an excellent classification accuracy of 93.834% and performed better than several novel methods. These experimental results indicate that our method enables better reliability and feasibility in the detection of fatigued driving.

1. Introduction

The World Health Organization (WHO) stated that subjectively experienced dimensions, such as fatigue and chronic fatigue syndrome (CFS), have been included in the concept of health. Previous works have indicated that chronic fatigue may accompany with a series of physical symptoms [1] such as insomnia, memory loss, and inattention, which also bring out psychiatric diseases like depression and anxiety [2,3,4]. In more serious cases, these physical and mental illnesses may result in a vicious cycle. Besides these negative influences on human health, fatigue also impacts other aspects of human life. For instance, fatigue driving can greatly interfere with a driver’s judgment and response time, thus deeply threaten drivers’ safety. According to statistics mentioned by Li et al. [5], although 57% of drivers believed that fatigued driving is a serious problem, 20% of drivers have fallen asleep or napped while driving in the past year. Maclean et al. referred to some research and arrived at the conclusion that 20–30% of traffic accidents are related to fatigued driving [6]. Thus, effective and reliable fatigue detection, which is an urgent need, can keep drivers away from most traffic accidents. Thanks to the improvement of the computing performance of computers and servers, it is possible to process and analyze data in real-time. Feature engineering and machine learning (ML) algorithms have become a major focus of research on psychological disease detection [7,8], emotion recognition [9], and fatigue detection [10,11,12,13], and related researchers can effectively utilize physiological information such as eye movements, body movements, facial expressions, and signals from modalities such as EEG, electromyography (EMG), electrocardiography (ECG), and galvanic skin reaction (GSR) to monitor the brain activity [14,15].
Recently, the classification of EEG signals by using Riemannian geometry has been proposed by many researchers [16,17,18], and it has become a hot topic to utilize covariance matrices as features of signals. However, most researchers only calculated covariance matrices of the whole trial as the total features, which neglected the information of temporal relation in symmetric positive-definite (SPD) space throughout the duration of a trial. Alternatively, by utilizing the superiority in Riemannian geometry, except for the calculation of distance and mapping, which is the shallow use of Riemannian network, we can extract more effective higher-level features in SPD space [19]. Therefore, our study was centered on the changes in the spatial-temporal relations of signals recorded by different EEG channels, and we investigated whether it is discriminative in the fatigue driving detection. For ease of description of our model, we have defined some concepts. First, all the covariance matrices calculated from EEG signals we used were defined as relation domain, which means they can reflect the relationship between different EEG channels. Then, the relation domain was divided into two subdomains: temporal relation (TR) domain and spatial relation (SR) domain. In the spatial relation domain, the covariance matrix of each trial was calculated, then we took the characteristics of its distribution as the SR-domain features in SPD space. In the temporal relation domain, EEG signals are partitioned into several segments by leveraging the slide window technique, then covariance matrices corresponding to each segment are obtained and flattened. It has been proven that the long short-term memory (LSTM) network could solve the gradient vanishing problem of the basic RNN model [20], and the LSTM network performs better while processing long series problem such as language modeling [21]. Meanwhile, some modified model based on LSTM has been utilized in more scenarios such as real-time traffic flow prediction [22] and 3D action recognition [23,24]. Therefore, we use an LSTM network to obtain the TR-domain features of higher level by modeling the relationships between these flattened matrices. Then, SR-domain features and TR-domain features are fused by the fully connected (FC) layer, and the fused features were fed into the softmax layer to make the final prediction. The experimental results compared with other methods show that our method achieves an accuracy of 93.834% for detecting fatigue state and demonstrate that this method outperforms several advanced methods. The main contributions of our study are listed as follows:
  • It discusses the relationship between different EEG channels and explores the changes of temporal dynamics of this information, which further contributes to the extension of signal processing and modeling conception.
  • It combines two kinds of spatial features on the SPD space to improve the classification performance of different driving states.
  • It proposes a state-of-the-art fusion method, which utilizes the spatial-temporal information of covariance matrices on the Riemannian manifold, and achieves the mission of detecting fatigue driving state effectively.
The remainder of this paper is organized as follows. In Section 2, the construction of our method and extraction of TR and SR features are reported. In Section 3, we present the description of dataset and report the experimental results, we also analyze the relationship between brain state and the covariance matrix. Finally, in Section 4, we conclude our work and discuss the possibility of future expansion and application.

2. Methodology

As we defined before, in the relation domain constructed by the covariance matrices, we can learn the relationships between signals of different EEG channels by processing these matrices. The covariance matrices obtained from EEG signals have been shown to be widely used in many areas [25,26], which also prove that the relation domain we defined could greatly contribute to the feature extraction. As proven in [27], the driver’s mental state while driving can be revealed by the information of some specific regions of the brain under the cortices, and the design of our method to obtain some specific features in the relation domain can be supported by understanding these sources of information. The covariance matrices in SR domain reflect the characteristics of the whole trial, which means the EEG signals of different states are analyzed from a macro perspective. According to some advances in technology of brain–computer interface [26,28], the feasibility of classification method based on the spatial covariance matrix can be demonstrated. However, just as we mentioned before, most studies only considered the spatial features captured from the whole trial and neglected the temporal information over time. Moreover, in practical applications, the spatial distribution of covariance matrix is more concerned, while the temporal information of covariance characteristics of EEG signal has not been deeply studied. In time-series tasks such as action recognition, forecasting, and virtual driving task, the temporal changes over time contain more information that can be well used. We assumed that the covariance matrices from the partitioned signals can capture the relationship between brain regions over time, and based on this, the temporal domain features of covariance of EEG signal are extracted and investigated in our study. Thus, as shown in Figure 1, we propose the framework contains spatial features and temporal features and fuses them together, which will be introduced in detail in the remainder of this section.

2.1. Features of Spatial Relation Domain

Huang et al. [19] proposed a novel and special Riemannian network architecture in 2016. This network could project the SPD matrices to a more smoother SPD space because of the depth nonlinear characteristic mapping. Through the end-to-end method, the model can automatically learn the optimal mapping mode, so as to extract the higher-level information of the covariance matrix, which greatly improves the classification accuracy of the model. Figure 2 shows the pipeline of the SPD network; the first layer of this network is called Bilinear Map Layer, which is utilized to generate covariance matrices in the SPD space with more discrimination. However, it is possible that some large eigenvalues of the original matrices may be reduced to very small number after being processed by Bilinear Map Layer. These very small eigenvalues may affect the positive-definite properties of matrices, which violates the principle of SPD space. Thus, the Eigenvalue Regularization Layer is set to process this issue by replacing these special eigenvalues with the threshold. In order to apply the classical Euclidean computations, the Logarithmic Eigenvalue Layer utilizes the logarithm operation of matrices proposed by Arsigny [29] to reduce the matrices to a flat space. In this study, we applied two Bilinear Map Layer, one Eigenvalue Regularization Layer, and one Logarithmic Eigenvalue Layer to construct the SPD network, which uses the whole-trial covariance matrices as input.
Although the SPD Network extracts the higher-level information of the covariance matrix, when we reshape it into a 1-D vector, it may lose some information of the original matrix. It has been verified by Congedo [18] that the sample distribution of each class could be reflected by the geometric mean of that class with proper matrices based on SPD space. In addition, the classification algorithm could achieve reliable results by utilizing the distance information between the mean of classes and each sample, which also keeps the distance information of matrices in the original SPD space. Therefore, in our study, we believe that by using symmetric Stein divergence [30], the vigilant level of the driver can be distinguished to some extent. The Stein divergence is used to measure the distance between the centers of two classes (drowsy and vigilant) and each sample, which could supplement the information of the original matrix by utilizing the distance information.
The whole-trial matrices of SR domain are assumed to be zero-centered, which is defined as C:
C = 1 Q 1 X X T , X R N × Q
where X denotes the EEG signal of one whole trial, Q = 9 s × 250 Hz = 2250 denotes the sampling points of trial, and N = 30 indicates the number of EEG channels As proven by Sra [30], the center of these covariance matrices should be subject to the condition as follows:
C e n t e r ^ = arg min C e n t e r ( i = 1 M δ R 2 ( C e n t e r , C i ) )
where M denotes the number of covariance matrices, and in the training set, the covariance matrix of one single trial is calculated and defined as C i , C e n t e r ^ denotes the center of M trialwise covariance matrices. Moreover, the first order of (2) should be satisfied with
( C e n t e r ) = i = 1 M C e n t e r + C i 2 1 M C e n t e r 1 = 0
According to the work in [31], the (3) has no analytical solution, thus as illustrated in Algorithm 1, we utilize the iterative approximation to solve the problem.
Algorithm 1 Centre of Covariance Matrices
Input: 
C i , i 1 , M
Output: 
The Centre of C i , i 1 , M
1:
initial: Centre = 1 M i = 1 M C i , iterator = 0; t h r e = 1 × 10 9 ; M a x i t e r a t o r = 50;
2:
repeat
3:
     C n e w = 1 M i = 1 M C i + C e n t r e 2 1 1 ;
4:
     c r i = C n e w C e n t r e F ;
5:
     C e n t r e = C n e w ;
6:
     i t e r a t o r = i t e r a t o r + 1 ;
7:
untilcri <thre or iterator> Maxiterator
In Algorithm 1, the Frobenius norm c r i is defined to represent the difference between C n e w and C e n t e r . We define the M a x i t e r a t o r as the maximum number of iterations and t h r e as the threshold to satisfy the requirement of the convergence stability. Similar to the Eigenvalue Regularization Layer of SPD network, to ensure the positive-definite properties of the covariance matrices, each covariance matrix C is regularized as
C = U c m a x Λ c , ϵ I U c T
where e p s i l o n is the threshold we set ( 1 × 10 8 ), and I represents the identity matrix. The eigenvalues Λ c and eigenvectors U c are calculated by applying eigendecomposition:
C = U c Λ C U c T
The variate m a x ( Λ c , ϵ I ) of (4) is a diagonal matrix E whose diagonal elements are defined as
E ( i , i ) = Λ ( i , i ) , Λ ( i , i ) > ϵ ϵ , Λ ( i , i ) ϵ
In both training set and testing set, the distances between the covariance matrix C j of each trial and the centers of the two classes trials with vigilant or drowsy labels are calculated as
δ R ( C j , C V / D ) = log | ( C j + C V / D ) / 2 | | C j C V / D |
Finally, according to the work in [16], we took these distances information as another kind of feature in the SR domain. Then, we fuse the features obtained from SPD network and Stein divergence as the total features of SR domain.

2.2. Features of Temporal Relation Domain

In this study, we assumed that the variations in the relations between different regions of the brain over time can be captured by the partitioned EEG signals and their consecutive covariance matrix. The driving fatigue state can be reflected by learning these variations. Figure 3a shows the sample of signals from each trial. The 9 s signals were divided into seven segments with a time window of 3 s and a step of 1 s. Thus, we define the covariance matrix corresponding to the t-th segmented signal as follows:
C t = 1 d 1 X t X t T , t 1 , 2 , , 7
where N = 30 means the channel number, d = 3 s × 250 Hz = 750 indicates the sampling points. Thus, the time series of the t-th segment with N channels are represented by X t , which is assumed to have a mean of zero.
As shown in Figure 3b, according to the symmetry of the covariance matrix, we take the diagonal elements and the lower triangular elements to reconstruct the matrix into a one-dimension vector m t , which has 465 elements. Sabbagh [32] has proved that such a flattened covariance matrix can simulate the linear features of the brain source, thus avoiding the problem of solving the inverse matrix directly [32]. Let O denote the number of linearly independent brain sources, thus the transformation matrix A R N × O is required to connect signals of EEG channels X t and brain sources E t R O × d by using the equation X t = A E t . That means we can use E t and A to represent the covariance matrix C t as C t = 1 d 1 A E t E t T A T . Sabbagh et al. [33] suggested in their study that the signals of the brain sources are linearly independent with each other, which means the covariance of two different sources is zero. Thus, the matrix E t E t T is a diagonal matrix, whose diagonal elements are the variances. By utilizing this method, without calculating the transformation matrix explicitly, we can extract the power of brain sources linearly. Accordingly, we defined the features of the 9 s trial in TR domain as follows:
[ m 1 , m 2 , , m 7 ]
the power of brain sources can be represented by m t , and the dynamic changes over time could be learned from the series above.
By taking the advantage of RNN, which is always used to process the time series data by exploring the comprehensive underlying information over time. Thus, we choose LSTM networks to extract TR-domain features at a high level. As shown in Figure 4, in our model, each RNN layer is constructed with 7 LSTM units. Several previous works [34,35] have achieved excellent performance, and as proven in paper [36], such a well-designed RNN structure could obtain the underlying information of the input sequences such as long-term dependencies. In the TR domain, a 2-layer RNN model is designed to capture the intrinsic information of covariance matrices over time In Figure 4, the first RNN layer process the series of (9) and output the hidden variable sequence as [ h 1 , h 2 , , h 7 ] , where h t and h t + 1 represent the consecutive hidden state. The hidden variable sequence is also the input of the second RNN layer, which denoted by [ h 1 , h 2 , , h 7 ] . Thus, we got h 7 , which is the final output of the RNN, and we fused it with the features from the SR domain.

2.3. Fusion and Optimization

Considering that the features from both TR domain and SR domain contribute to the final fatigue detection, as illustrated in Figure 1, we fused the δ R ( C j , C V ) and δ R ( C j , C D ) , which calculated based on Stein divergence, features obtained from SPDNet and features of TR domain captured by the RNN together. Referring to [37], the spatial-temporal joint optimization method is illustrated in Algorithm 2. We optimize the parameters alternately in each iteration. Specifically, when optimizing the parameters of SPD network, we fix the parameters of RNN, and vice versa. Different from the other parameters, the bias and weight of fully connected layer are updated in the end of each iteration. As shown in the algorithm, X L t l , X S t s , X D denotes the vector of fused features from both domains, where X L is the feature of the TR domain, X S is the vector transformed from the trial-wise SPD covariance matrix after being mapped by the SPD network, and X D is the feature vector δ R ( C j , C V ) , δ R ( C j , C D ) introduced previously. The mapping operation by FC layer is represented by F X L t l , X S t s , X D , ω F t , where ω F t is the weight parameters of FC layer. Besides, we utilize the cross-entropy as our loss function £ Y , f t . In order to record the number of iterations of optimization, t s and t l are defined in the optimization of SPD network and RNN, separately, while t indicates the total number of iteration.
In summary, we obtain the SR-domain features, which are the fused vectors consist of the distances information between the covariance matrices of a single trial and the center of class based on Stein divergence as well as the reshaped vector obtained from the SPD network. Meanwhile, the TR-domain features are the final output of the 2-layer RNN by processing several time series segments of a single trial. Then, we concatenate the features and get the ultimate predictions by feeding the features into the fully connected layer and softmax layer.
Algorithm 2 Spatial-temporal joint optimization algorithm
Input: 
X L : input of LSTM; X S : input of SPDNet; X D : input of Stein divergence; Y: Training set label
Output: 
ω L : parameters of RNN; ω S : paremeters of SPDNet; ω F : parameters of FC layer
1:
initial M a x i t e = 50; t l = 1 ; t s = 1 ;
2:
for each t [ 1 , M a x i t e ] do
3:
    if t % 2 = 1 then
4:
         X L t l + 1 = L X L t l , ω L t l
5:
         f t = F X L t l + 1 , X S t s , X D , ω F t
6:
         L o s s = £ Y , f t
7:
         ω L t + 1 = L o s s f t · f t X L t l + 1 · X L t l + 1 ω L
8:
         ω ˜ L t l + 1 = ω L t η L ω L t l + 1
9:
         t l t l + 1
10:
    end if
11:
    if t % 2 = 0 then
12:
         X S t s + 1 = S X S t s , ω S t s
13:
         f t = F X L t l , X S t s + 1 , X D , ω F t
14:
         L o s s = £ Y , f t
15:
         ω S t + 1 = L o s s f t · f t X S t s + 1 · X S t s + 1 ω S
16:
         ω ˜ S t s + 1 = ω S t η S ω S t s + 1
17:
         t s t s + 1
18:
    end if
19:
     ω F t + 1 = L o s s f t · f t ω F
20:
     ω ˜ F t + 1 = ω F t η F ω F t
21:
end for

3. Experimental Results and Discussion

3.1. Dataset and Data Processing

3.1.1. Experimental Paradigm

In this study, we use the open benchmark dataset, which consists of EEG signals collected during the sustained-attention driving task [38] based on the experimental paradigm of immersive driving proposed by Huang et al. [39]. This paradigm and the dataset were also used in related works [40,41,42,43,44], and the validity and reliability have been verified by these works with effective result. As mentioned in [38], the study of the dataset we used was performed in strict accordance with the recommendations in the Guide for the Committee of Laboratory Care and Use of the National Chiao Tung University, Taiwan. Besides, the study mentioned was also approved by The Institutional Review Board of the Veterans General Hospital, Taipei, Taiwan. As shown in Figure 5, the participants were required to drive at a constant speed of 100 km/h in a virtual night-time scenario, and the car will randomly drift out of the lane at the speed of 5 km/h. Subjects were asked to steer the car back to the center of the lane as quickly as possible. Then, the time spent driving back to the original lane, which defined the reaction time, was used to represent the drowsy level of the subject. As shown in Figure 6, we can see a clear difference between the two states, and most reaction times of subjects in the vigilant state were around the average, which is also significantly smaller than that of the fatigue state. Thus, we believe that most people are more likely to behave differently between the vigilant and fatigue state.

3.1.2. Definition of Labels

In our study, in order to include the information which could reflect the driving state, we capture the data from 1 s before the response offset to 8 s after that as the data of one trial, which contains the deviation onset and response onset. We defined the reaction time mentioned before as local reaction time (RT). As defined in [45], we average several local RT values around the current trial and define the average as the global RT, by which the level of drowsy state while driving could be represented. Thus, the global RT for the l-th trial was defined as
g l o b a l R T ( l ) = i = l n l + n l o c a l R T ( i ) 2 n + 1
where n was set to 2 in our study. As shown in Figure 7, we defined the following vigilant and drowsy labels, where R T l denotes the local reaction time and R T g denotes the global reaction time:
T r i a l ( i ) R T l ( i ) 0.62 R T g ( i ) 0.62 , v i g i l a n t R T l ( i ) 1.5 R T g ( i ) 1.5 , d r o w s y
Accordingly, with a total number of 1433 trials, 739 trials were defined as the vigilant labels and 694 trials were defined as the drowsy labels. The length of each trial was 9 s, and the number of trials of each subject varies from 2 to 179.

3.1.3. Data Preprocessing

According to the work in [38], 27 volunteers were recruited to participate in a 90-min task of sustained virtual attention driving while 62 EEG recordings were collected. The original signals were collected using a system with 32 electrodes, which includes 2 reference electrodes, and digitised at 500 Hz with 16-bit resolution. For the purpose of eliminating the noise, we downsampled these EEG signals to 250 Hz and used bandpass filtering to retain the information between 1 and 50 Hz. In order to remove the artifacts that we are not interested in, such as muscle activity, eye movement, blinking, and ECG signals, we utilize Independent Component Analysis (ICA) to extract the pure EEG signals.

3.2. TR-Domain Experiments

3.2.1. Validity Analysis of TR-Domain Features

To visually investigate the temporal dynamics of the relationship between EEG channels, the reshaped flatten covariance matrices of one single trial were combined into a 7 × 465 matrix defined as M, whose column indicates the dynamic changes of one pair of EEG channels. By averaging the matrices M of all trials belonging to different classes, we get M V and M D to represent the dynamic changes of 739 trials with vigilant labels and 694 trials with drowsy labels. The Pearson’s correlation coefficient (PCC) was used to calculate the correlations between different columns of M, and the PCC could reflect whether we can learn useful identifying information from the variations. Considering the tradeoff of computational complexity, we chose only 2, 3, 5, and 7 channel pairs and all pairs and focused on comparisons of them, which can be used to prove the extensibility and practicability of our model. Given the chosen number of EEG channel pairs, we defined the measurement ζ for each possible combination of different columns of M as below to check the inter-class and intra-class correlations:
ζ = Γ + Θ Φ + Υ
where Γ , Θ , Φ , and Υ are defined as
Γ = i , j χ f V , i j exp ( ϕ ( p i v , p j v ) ω )
Θ = i , j χ f D , i j exp ( ϕ ( p i d , p j d ) ω )
Φ = i χ f V , j χ f D exp ϕ ( p i v , p j d )
Υ = i , j χ f V , k , l χ f D | | ϕ ( p i v , p j v ) ϕ ( p k d , p l d ) | | 2
Let χ V and χ D denote all possible combinations of a given number of column pairs in M V and M D , respectively. We define the combination function f to represent the set of combinations of columns, and the χ f V and χ f D indicate the same elements in χ V and χ D , respectively. The i-th column in M V and M D were presented by p i v and p i d , which belong to the column set in χ f V and χ f D . The PCC between p i v and p i d is denoted by ϕ ( p i v , p i d ) . For example, we want to choose 3 columns to calculate the measurement, so we can randomly choose 3 columns in M V and M D . Then, the χ V and χ D consist of 465 C 3 combinations, respectively. For a given combination fuction f, there are 3 corresponding columns in χ f V or χ f D , respectively. Here, ω is set to be 0.9 in our work, which is the threshold representing the strong correlation criterion between two columns. As shown in (13) and (14), Γ and Θ represent the intra-class correlations of the same combination function f in χ V and χ D . Φ denotes the inter-class correlation of the same combination f in χ V and χ D , and Υ indicates the similarity between the intra-class correlations for the same column pair in χ f V and χ f D , thus i and k are the same column, and j and l are the same column in (16).
The purpose of this experiment is to ascertain whether we can learn some useful discriminative information from the TR domain; ζ is defined to measure whether the combination we selected meet our requirement. Thus, we aim to find the combinations with relatively high value of ζ , which are achieved in cases of high values of Γ and Θ and low values of Φ and Υ , which means the combination with high value of ζ has relatively weak inter-class correlations and strong intra-class correlations. These combinations with weak inter-class correlations show the difference between vigilant and drowsy driving state. To intuitively illustrate the results based on the measurement ζ , Figure 8 clearly shows the dynamic variations in the relationship of different channels over time in the two states. The trajectories showed in the figure clearly illustrate the truth that the changes of the same class are similar to each other but the variations of different classes are completely different. Furthermore, Table 1 presents quantitative information on the intra-class and inter-class correlations. As shown in Table 1, the intra-class correlations are generally higher than the inter-class correlations.
The results of this section are consistent with the hypothesis of the previous section Features of Temporal Domain, which indicates that the variations of the relationship between different channels over time play an important role in the classification. Based on research on this hypothesis, the feasibility as well as reliability of developing such a method that utilizes TR-domain features can be further proven.

3.2.2. Comparison of Different Number of Channel Pairs

To evaluate the effectiveness of our method and avoid data leakage, we took the advantage of the leave-one-subject-out strategy to conduct the experiments on the open dataset mentioned in Dataset and Data Processing section, which also proves that our model performs well on subject independence. Considering the advantage of TensorFlow framework, we implemented our model in it, and the rectified linear unit (ReLU) and cross-entropy loss was set as the activation function and loss, respectively. To avoid overfitting, the early stop trick was applied in our experiments. Besides, we use the evaluation index as follows to evaluate our fatigue detection model: accuracy, sensitivity, specificity, and F1 score.
As shown in Figure 9, we selected different numbers of columns to conduct the classification experiment. The results show that the classification accuracy improved with the increase of the selected numbers, which indicates that model with more EEG channel pairs could obtain more comprehensive information. Therefore, classification using all EEG channel pairs outperforms that using the subset of channel pairs. By using the paired samples t-test, we found that there are significant differences between the different number of channel pairs because the p-value is smaller than 0.01. As shown in Figure 9, the model selected all channel pairs achieved the highest accuracy. With the decrease of channel number, the classification accuracy also decreased by about 4–10%. Considering the volume conduction effects in EEG [46], the signal of each EEG channel is not generated by a single neuron but by the discharge of many neurons. Therefore, with the increasing of the selected number of channel pairs, the linear power features of the neurons under test will increase, which can be reflected by the classification accuracy.

3.3. Comparative Experiments

To prove the effectiveness of our model by fusing features from the spatial and temporal domain, we compared several novel methods with ours. These methods also calculated the covariance matrices and utilized them as features to construct their models. Consistent with our method, the same leave-one-subject-out strategy and experimental settings were used in the other selected methods, in which accuracy, specificity, sensitivity, and F1 scores were utilized for evaluation.
By using the novel neural network, Mehdi Hajinoroozi et al. [47] proposed a method based on convolutional neural network (CNN) which utilized covariance matrices as the input features. In our experiments, we named that method as COV_CNN to compare with ours. Similarly, we abbreviated the method using CNN and RNN parallelly as CNN_RNN [48], which also used covariance matrices as input. Referring to the work in [45], we also choose a common method without using the covariance matrix. The FRE method cited in our manuscript utilized power spectral density (PSD) matrix and CNN to construct the model. Considering the information contained in the TR domain, we called the method with all channel pairs in the previous experiment as the TR domain classifier (TRDC) and designed an experiment based on that. Finally, in order to investigate the effectiveness of using only Stein divergence or SPDNet features to fused with TR-domain features, we designed a classifier using Stein divergence and TR-domain features (SDTR) and that using SPDNet and TR-domain features (SNTR), and then conducted classification experiments utilizing them, respectively. The constructions of these methods are as follows.
1.
COV_CNN: Consistent with the work in [47], a CNN was set to extract the information of whole-trial covariance matrices. The kernel size of CNN with 1 pooling layer were set to be 3 × 3 × 10 and 3 × 3 × 5 , respectively, while the stride was set to be 1.
2.
CNN_RNN: Referring to the work in [48] and being consistent with our work, a CNN with 2 layers and an RNN with 2 layers are combined parallelly to reconstruct the model. Then, the output features of CNN and RNN are fused together and sent into the softmax layer for final prediction. The kernel sizes of the first and second CNN layer were 3 × 3 × 32 and 3 × 3 × 64 , and the kernel stride was 1.
3.
FRE: We selected the signal of each channel by bandpass filtering with 1–30 Hz and divided it into 30 bands on average. Then, we utilized short-term fast Fourier transform (STFT) to calculate the PSD features and obtain the amplitude in each band. By processing the signals, each single trial could be denoted by a PSD matrix, whose size is 30 × 30 . Considering that CNN has excellent capabilities of image processing, the obtained PSD matrix was fed into a 3-layer CNN, whose parameters were set as follows: the kernel sizes of each layer were set to be 3 × 3 × 1 , 3 × 3 × 32 , and 3 × 3 × 64 , respectively, and the kernel stride was 1.
4.
TRDC: The output of RNN is thought to be the final feature of the TRDC model. Without being fused with other features, the output is fed into the fully connected layer. Besides, the parameter settings of this method are the same as the spatial-temporal fusion method we proposed.
5.
SDTR: Consistent with the method in [31], without fusing with features obtained from SPDNet, we only fused distance features calculated based on Stein divergence and features of the TR domain. Then, we fed this feature vector into the FC layer and softmax layer for prediction. The framework of TR domain is the same as our method.
6.
SNTR: Following the parameters setting of [37], the features obtained by mapping a single covariance matrix using SPDNet and features of the TR domain were fused together to be fed into the FC layer and softmax layer. Then, we used the prediction for the final classification.
As shown in Table 2, in comparison with our fusion method, the SDTR achieved an average accuracy of 89.280%, and the SNTR achieved an average accuracy of 91.042%, whose performances were reduced by approximately 4.554% and 2.792% compared to that of our method. Thus, it was proven that the SR domain features combined Stein divergence and SPDNet utilizing EEG signals could better reflect the difference between vigilant and drowsy states, which is consistent with the assumption of the previous section SR-Domain Features. TRDC achieved an average accuracy of 86.238% even without considering the features of the SR domain, which indicates that the features of the TR domain contain more discriminative information for the classification of fatigue. This result also shows us the importance of information over time in long-term time series tasks, which is one of the reasons why we select RNN to process this modeling task. However, as shown by the experimental results without asterisks of SDTR and SNTR in Table 2, through the significance test of the evaluation criteria, it was found that there was no statistically significant difference in these evaluation results between our methods and that of the two methods above. We analyzed the possible causes and checked the p-value between TRDC and SDTR, TRDC and SNTR, SDTR and SNTR, respectively. The results showed that there was no significant difference between SDTR and SNTR, while the other combinations had significant differences. According to the results of TRDC, we know that temporal information contributes more to the final results than spatial information; thus, we believe that the difference between different feature extraction methods in the spatial domain is not significant. However, when the spatial features were fused with temporal features, the model truly improved the classification accuracy, which can be proven by the experimental results. Furthermore, the spatial features based on the fusion of SPD network and Stein divergence also outperformed TRDC. However, compared with SDTR and SNTR, although it performed better than these two methods, because of the main contributions of TR-domain features, the differences between our method and SDTR as well as SNTR were not significant, which also can be proven by the significance test result between SDTR and SNTR. Therefore, how to improve the significance of the difference is worth to be explored in future work. The mean accuracy achieved by FRE is 75.155%, which is also inferior to that of our method, thus the truth that the relation-domain features contribute more to the classification model than features as PSD matrices can be proven. Compared with the methods that neglected the temporal information in the relation domain such as COV_CNN (the mean accuracy is 81.877%), our methods performed better, whose accuracy improved about 11.957%. Although CNN_RNN also uses an LSTM network to capture the features of covariance matrices, it only captures spatial features due to the use of the whole-trial matrix. As mentioned before, temporal features improves the performance of classification more, therefore the average accuracy of CNN_RNN is only 79.103%, which is reduced by about 14.731% compared to that of our method. In comparison, our method reduces the complexity of the model, avoids the introduction of noise, and improves classification performance by fusing both spatial and temporal features of the relation domain. In other words, learning from all experiments conducted previously, the spatial-temporal fusion method outperforms other methods that only consider spatial information alone.
As the experimental results showed in the section Comparison of Different Number of Channel Pairs, considering the trade-off of computational complexity, time cost, and classification accuracy, we compared the results with the different number of channel pairs and got the conclusion that the model can achieve relatively high accuracy although reducing the number of channel pairs. Thus, the possibility of future online use of our model can be supported by this result. GM Duma et al. [49] also discussed the possibility of whether such an EEG-based detection model could be used in practice to protect the safety of drivers. Some related works [50,51] also mentioned the online use of EEG-based models in the area of driving assistant systems, and the authors of [50] have already tested their system while driving the real car. With high classification accuracy and low time cost, we believed that our model could be used to detect the fatigue state in practice. In future studies, we will test the performance of our algorithm on other datasets.

3.4. Brain State Analysis Based on Covariance Matrix

Recent evidence [45] suggests that the locations of the EEG channels represent the distribution of different brain regions and that the covariance matrices capture the relations between different channels. As mentioned in section TR-Domain Features, the covariance matrix can simulate the linear features of the brain source. EEG signals captured from the scalp are superposed signals generated by the discharge of many neurons, then we let K denote the number of linearly independent brain sources. Therefore, the set of all neurons is represented as s ( n ) = [ s 1 ( n ) , s 2 ( n ) , , s K ( n ) ] R N × K , where n is the time-series length of signals produced by the neurons. Let M denote the number of EEG channels, x ( n ) = [ x 1 ( n ) , x 2 ( n ) , , x M ( n ) ] represents the signals of all channels. As illustrated in [52], in the study of brain sources, we often divide the problems into the forward problems and the inverse problems. The inverse problem is locating the position of the source of evoked potentials from measurements on the surface of the scalp. Thus, the matrix W is required for modeling the brain sources as s ( n ) = W T x ( n ) , which is based on that the EEG signals could be represented by the linear combination of discharge produced by the neurons. In contrast, the forward problem is solved by starting from a given electrical source and calculating the potentials at the electrodes, which means given the current signals s ( n ) , there is a matrix A that could denote the signals of channels as x ( n ) = A s ( n ) + ε ( n ) , where ε ( n ) denotes the noise during the conduction. In general, we assume that the noise is linearly independent with each other, thus the covariance Σ ε = 0 . Haufe et al. proposed that AW T = I , where I is the identity matrix [53]. However, the conduction matrix is not guaranteed to be a square matrix, so in the process of solving the forward problem, the following transformation should be made:
Σ x ( n ) = AE s ( n ) s T ( n ) A T + Σ ε
because Σ ε = 0 , so
Σ x ( n ) = AE s ( n ) s T ( n ) A T
we multiply W on the both sides of Equation (17):
Σ x ( n ) W = A Σ s ( n )
As shown in Equation (19), the covariance characteristics of EEG signals are closely related to the activity of the neurons, which is also the solution of the forward problem:
A = Σ x W Σ s 1
Sabbagh et al. [33] proposed that the electrical activity of the brain neurons is linearly independent, that is, Σ s ( n ) = 0 . Therefore, it is proven that the flattened covariance matrix can simulate the linear features of the brain source, which could avoid solving the inverse problem directly [32].
According to Equation (19), when the solution of the inverse problem is determined in the localization of the brain source, the covariance matrix corresponds to the weight matrix of the forward problem, which is closely related to the activity state of the brain. Based on this, we conducted research on the influence of the covariance matrix of EEG signals on the activity of the brain neurons under fatigue and vigilant conditions.
As shown in Figure 10a, we assumed that the number of neurons in brain is 364; Σ x R 30 × 30 denotes the covariance matrix obtained from the EEG signals. Under the condition that the head model is fixed, we utilized BrainStorm to obtain the solution of the inverse problem of brain source localization, which is represented by the transform matrix W R 30 × 364 . Thus, we obtained the electrical activity of the neurons S ^ R 364 × 2250 corresponded with each EEG segment, and the covariance matrix of neuron signals as Σ S ^ R 364 × 364 . Then, the solution A R 30 × 364 of the forward problem of each EEG segment could be obtained, which reflects the activity state of brain neurons. As shown in Figure 10b, we calculated the weight matrix set corresponded to the covariance matrices of all samples in two classes, respectively. Then, we calculated the average of weight matrix set of each class, and got the weight matrix A represents the fatigue and vigilant state, respectively. Thus, the i-th column of the matrix denotes the weight assignment of all channels by the i-th neuron during the conduction from the neurons to the scalp. In conclusion, in the different brain states, the weight assignment varies greatly, which can be seen in the third column of Figure 10b.
To explore the difference of the above weight assignment in fatigue and vigilant states intuitively, we utilized t-distribution [54] to fit the weight assignment and get a smoother distribution curve. The Kullback–Leibler Divergence (KL Divergence) defined as Equation (21) could quantify the difference between the two distributions P and Q, where X is the set of samples. In this study, the KL Divergence was used to quantify the difference of the weight assignment of all channels by the same neuron between different brain states; therefore, the influence of EEG covariance on different brain states can be observed intuitively. We calculated the KL Divergences of all neurons and the values fall in the range between 96.91 and 176.92. Then, these KL Divergence were normalized to the values between 0 and 1 for further comparison.
D ( P | | Q ) = i X P ( i ) · log ( P ( i ) Q ( i ) )
As shown in Figure 11, we can obviously obtain the difference of weight assignment between fatigue and vigilant brain states. Considering the large number of neurons, we only show the activity difference of nine neurons. The blue curve represents the t-distribution diagram fitted by the weight assignment in the fatigue state, and the yellow curve denotes that in the vigilant state. According to the concept of t-distribution, the shape of the curve represents the concentration degree of weight assignment of all channels by each neuron we assumed. Moreover, the curves with different colors obtained from the i-th neuron represent the difference between the concentration degree of assignment with different brain states. As shown in Source_136 of Figure 11, we can see that the assignment of all channels by the 136th neuron is more concentrated in the fatigue state than the vigilant state, which has a significant difference. In Source_88, the assignment is more concentrated in the vigilant state, which is opposite to Source_136. In conclusion, there is an obvious difference between the weight assignment of two states, which can be seen in the t-distribution diagrams and KL divergence on the top of each diagram. Therefore, it is proven that the covariance matrices obtained from the different level of fatigue state greatly reflect the information of brain activity in different states, which provide the effective features for classification.

4. Conclusions

In this study, we proposed a novel spatial-temporal fusion method on the Riemannian manifold by calculating the covariance matrices of EEG signals, which can well support the driver fatigue detection with high effectiveness. Compared with other traditional methods considering the spatial covariance matrices, we utilized the temporal features of the covariance matrix obtained from EEG signals. In addition, we took advantage of two aspects of spatial characteristics: numerical characteristics after mapping and distance information, which outperform other methods, and the combination of these two aspects are proven to be complementary, which is better than only use one of them. Furthermore, this paper offers a new perspective of research on EEG signals modeling and further contributes to the extension of signal processing and modeling conception. As we mentioned before, the proposed EEG-based model has the potential to be used in practice, and we believe the new perspective could help to construct more intelligent or emotion-aware applications, which will contribute to the improvement of the quality of human life.

Author Contributions

Conceptualization, N.Z., D.L. and X.Z.; methodology, N.Z. and D.L.; software, N.Z. and D.L.; validation, N.Z., D.L. and K.H.; formal analysis, N.Z. and D.L.; investigation, N.Z. and D.L.; resources, X.Z. and B.H.; data curation, N.Z. and D.L.; writing—original draft preparation, N.Z. and D.L.; writing—review and editing, X.Z.; visualization, N.Z., K.H., X.W. and M.C.; supervision, X.Z. and B.H.; project administration, X.Z. and B.H.; funding acquisition, X.Z. and B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Key R&D Program of China (Grant No. 2019YFA0706200), in part by the National Natural Science Foundation of China (Grant No. 62072219, 61632014), in part by the National Basic Research Program of China (973 Program, Grant No. 2014CB744600).

Data Availability Statement

Not applicable.

Acknowledgments

The authors are thankful to the reviewers for their comments and suggestions to improve the quality of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Krupp, L.B.; LaRocca, N.G.; Muir-Nash, J.; Steinberg, A.D. The fatigue severity scale: Application to patients with multiple sclerosis and systemic lupus erythematosus. Arch. Neurol. 1989, 46, 1121–1123. [Google Scholar] [CrossRef]
  2. Chen, M.K. The epidemiology of self-perceived fatigue among adults. Prev. Med. 1986, 15, 74–81. [Google Scholar] [CrossRef]
  3. Lewis, G.; Wessely, S. The epidemiology of fatigue: More questions than answers. J. Epidemiol. Community Health 1992, 46, 92. [Google Scholar] [CrossRef] [Green Version]
  4. Manu, P.; Matthews, D.A.; Lane, T.J. The mental health of patients with a chief complaint of chronic fatigue: A prospective evaluation and follow-up. Arch. Intern. Med. 1988, 148, 2213–2217. [Google Scholar] [CrossRef]
  5. Li, D.H.; Liu, Q.; Yuan, W.; Liu, H.X. Relationship between fatigue driving and traffic accident. J. Traffic Transp. Eng. (Xi’an Shaanxi) 2010, 10, 104–109. [Google Scholar]
  6. MacLean, A.W.; Davies, D.R.; Thiele, K. The hazards and prevention of driving while sleepy. Sleep Med. Rev. 2003, 7, 507–521. [Google Scholar] [CrossRef]
  7. Zhang, X.; Liu, J.; Shen, J.; Li, S.; Hou, K.; Hu, B.; Gao, J.; Zhang, T. Emotion Recognition From Multimodal Physiological Signals Using a Regularized Deep Fusion of Kernel Machine. IEEE Trans. Cybern. 2020, 51, 4386–4399. [Google Scholar] [CrossRef]
  8. Mammone, N.; De Salvo, S.; Bonanno, L.; Ieracitano, C.; Marino, S.; Marra, A.; Bramanti, A.; Morabito, F.C. Brain network analysis of compressive sensed high-density EEG signals in AD and MCI subjects. IEEE Trans. Ind. Inform. 2018, 15, 527–536. [Google Scholar] [CrossRef]
  9. Ahmed, F.; Bari, A.H.; Gavrilova, M.L. Emotion recognition from body movement. IEEE Access 2019, 8, 11761–11781. [Google Scholar] [CrossRef]
  10. Gu, H.; Ji, Q.; Zhu, Z. Active facial tracking for fatigue detection. In Proceedings of the Sixth IEEE Workshop on Applications of Computer Vision, Orlando, FL, USA, 4 December 2002; pp. 137–142. [Google Scholar]
  11. Zeng, H.; Li, X.; Borghini, G.; Zhao, Y.; Aricò, P.; Di Flumeri, G.; Sciaraffa, N.; Zakaria, W.; Kong, W.; Babiloni, F. An EEG-Based Transfer Learning Method for Cross-Subject Fatigue Mental State Prediction. Sensors 2021, 21, 2369. [Google Scholar] [CrossRef]
  12. Hu, J.; Wang, P. Noise robustness analysis of performance for EEG-based driver fatigue detection using different entropy feature sets. Entropy 2017, 19, 385. [Google Scholar] [CrossRef] [Green Version]
  13. Han, C.; Sun, X.; Yang, Y.; Che, Y.; Qin, Y. Brain complex network characteristic analysis of fatigue during simulated driving based on electroencephalogram signals. Entropy 2019, 21, 353. [Google Scholar] [CrossRef] [Green Version]
  14. Hu, B.; Dong, Q.; Hao, Y.; Zhao, Q.; Shen, J.; Zheng, F. Effective brain network analysis with resting-state EEG data: A comparison between heroin abstinent and non-addicted subjects. J. Neural Eng. 2017, 14, 046002. [Google Scholar] [CrossRef] [Green Version]
  15. Zhang, X.; Pan, J.; Shen, J.; Din, Z.U.; Li, J.; Lu, D.; Wu, M.; Hu, B. Fusing of Electroencephalogram and Eye Movement with Group Sparse Canonical Correlation Analysis for Anxiety Detection. IEEE Trans. Affect. Comput. 2020. [Google Scholar] [CrossRef]
  16. Mishra, P.K.; Jagadish, B.; Kiran, M.P.R.S.; Rajalakshmi, P.; Reddy, D.S. A novel classification for EEG based four class motor imagery using kullback-leibler regularized Riemannian manifold. In Proceedings of the 2018 IEEE 20th International Conference on e-Health Networking, Applications and Services (Healthcom), Ostrava, Czech Republic, 17–20 September 2018; pp. 1–5. [Google Scholar]
  17. Barachant, A.; Bonnet, S.; Congedo, M.; Jutten, C. Classification of covariance matrices using a Riemannian-based kernel for BCI applications. Neurocomputing 2013, 112, 172–178. [Google Scholar] [CrossRef] [Green Version]
  18. Congedo, M.; Barachant, A.; Bhatia, R. Riemannian geometry for EEG-based brain-computer interfaces; a primer and a review. Brain-Comput. Interfaces 2017, 4, 155–174. [Google Scholar] [CrossRef]
  19. Huang, Z.; Van Gool, L. A riemannian network for spd matrix learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  20. Jozefowicz, R.; Zaremba, W.; Sutskever, I. An empirical exploration of recurrent network architectures. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 2342–2350. [Google Scholar]
  21. Sundermeyer, M.; Schlüter, R.; Ney, H. LSTM neural networks for language modeling. In Proceedings of the Thirteenth Annual Conference of the International Speech Communication Association, Portland, OR, USA, 9–13 September 2012. [Google Scholar]
  22. Tian, Y.; Zhang, K.; Li, J.; Lin, X.; Yang, B. LSTM-based traffic flow prediction with missing data. Neurocomputing 2018, 318, 297–305. [Google Scholar] [CrossRef]
  23. Rao, H.; Xu, S.; Hu, X.; Cheng, J.; Hu, B. Augmented skeleton based contrastive action learning with momentum lstm for unsupervised action recognition. Inf. Sci. 2021, 569, 90–109. [Google Scholar] [CrossRef]
  24. Xu, S.; Rao, H.; Peng, H.; Jiang, X.; Guo, Y.; Hu, X.; Hu, B. Attention based multi-level co-occurrence graph convolutional lstm for 3d action recognition. IEEE Internet Things J. 2020, 1. [Google Scholar] [CrossRef]
  25. Miah, A.S.M.; Rahim, M.A.; Shin, J. Motor-imagery classification using riemannian geometry with median absolute deviation. Electronics 2020, 9, 1584. [Google Scholar] [CrossRef]
  26. Park, Y.; Chung, W. Optimal channel selection using covariance matrix and cross-combining region in EEG-based BCI. In Proceedings of the 2019 7th International Winter Conference on Brain-Computer Interface (BCI), Gangwon, Korea, 18–20 February 2019; pp. 1–4. [Google Scholar]
  27. Zhang, Y.; Prasad, S.; Kilicarslan, A.; Contreras-Vidal, J.L. Multiple kernel based region importance learning for neural classification of gait states from EEG signals. Front. Neurosci. 2017, 11, 170. [Google Scholar] [CrossRef] [Green Version]
  28. Olias, J.; Martín-Clemente, R.; Sarmiento-Vega, M.A.; Cruces, S. EEG signal processing in MI-BCI applications with improved covariance matrix estimators. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 895–904. [Google Scholar] [CrossRef] [PubMed]
  29. Arsigny, V.; Fillard, P.; Pennec, X.; Ayache, N. Geometric means in a novel vector space structure on symmetric positive-definite matrices. SIAM J. Matrix Anal. Appl. 2007, 29, 328–347. [Google Scholar] [CrossRef] [Green Version]
  30. Sra, S. A new metric on the manifold of kernel matrices with application to matrix geometric means. In Proceedings of the 25th International Conference on Neural Information Processing Systems—Volume 1; Curran Associates Inc.: Red Hook, NY, USA, 2012; pp. 144–152. [Google Scholar]
  31. Zhang, X.; Lu, D.; Pan, J.; Shen, J.; Wu, M.; Hu, X.; Hu, B. Fatigue Detection With Covariance Manifolds of Electroencephalography in Transportation Industry. IEEE Trans. Ind. Inform. 2020, 17, 3497–3507. [Google Scholar] [CrossRef]
  32. Baillet, S. Magnetoencephalography for brain electrophysiology and imaging. Nat. Neurosci. 2017, 20, 327–339. [Google Scholar] [CrossRef]
  33. Sabbagh, D.; Ablin, P.; Varoquaux, G.; Gramfort, A.; Engemann, D.A. Predictive regression modeling with MEG/EEG: From source power to signals and cognitive states. NeuroImage 2020, 222, 116893. [Google Scholar] [CrossRef] [PubMed]
  34. Du, Y.; Wang, W.; Wang, L. Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1110–1118. [Google Scholar]
  35. Zhu, W.; Lan, C.; Xing, J.; Zeng, W.; Li, Y.; Shen, L.; Xie, X. Co-occurrence feature learning for skeleton based action recognition using regularized deep LSTM networks. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  36. Zhang, S.; Yang, Y.; Xiao, J.; Liu, X.; Yang, Y.; Xie, D.; Zhuang, Y. Fusing geometric features for skeleton-based action recognition using multilayer LSTM networks. IEEE Trans. Multimed. 2018, 20, 2330–2343. [Google Scholar] [CrossRef]
  37. Zhang, X.; Lu, D.; Shen, J.; Gao, J.; Huang, X.; Wu, M. Spatial-temporal joint optimization network on covariance manifolds of electroencephalography for fatigue detection. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Korea, 16–19 December 2020; pp. 893–900. [Google Scholar]
  38. Cao, Z.; Chuang, C.H.; King, J.K.; Lin, C.T. Multi-channel EEG recordings during a sustained-attention driving task. Sci. Data 2019, 6, 1–8. [Google Scholar] [CrossRef] [Green Version]
  39. Huang, R.S.; Jung, T.P.; Makeig, S. Tonic changes in EEG power spectra during simulated driving. In Proceedings of the International Conference on Foundations of Augmented Cognition, San Diego, CA, USA, 19–24 July 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 394–403. [Google Scholar]
  40. Lin, F.C.; Ko, L.W.; Chuang, C.H.; Su, T.P.; Lin, C.T. Generalized EEG-based drowsiness prediction system by using a self-organizing neural fuzzy system. IEEE Trans. Circuits Syst. I Regul. Pap. 2012, 59, 2044–2055. [Google Scholar] [CrossRef]
  41. Chang, C.Y.; Hsu, S.H.; Pion-Tonachini, L.; Jung, T.P. Evaluation of artifact subspace reconstruction for automatic artifact components removal in multi-channel EEG recordings. IEEE Trans. Biomed. Eng. 2019, 67, 1114–1121. [Google Scholar] [CrossRef]
  42. Wang, Y.K.; Jung, T.P.; Lin, C.T. EEG-based attention tracking during distracted driving. IEEE Trans. Neural Syst. Rehabil. Eng. 2015, 23, 1085–1094. [Google Scholar] [CrossRef]
  43. Yang, Y.; Gao, Z.; Li, Y.; Wang, H. A CNN identified by reinforcement learning-based optimization framework for EEG-based state evaluation. J. Neural Eng. 2021, 18, 046059. [Google Scholar] [CrossRef]
  44. Cui, Y.; Xu, Y.; Wu, D. EEG-based driver drowsiness estimation using feature weighted episodic training. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 2263–2273. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Chuang, C.H.; Huang, C.S.; Ko, L.W.; Lin, C.T. An EEG-based perceptual function integration network for application to drowsy driving. Knowl.-Based Syst. 2015, 80, 143–152. [Google Scholar] [CrossRef]
  46. van den Broek, S.P.; Reinders, F.; Donderwinkel, M.; Peters, M. Volume conduction effects in EEG and MEG. Electroencephalogr. Clin. Neurophysiol. 1998, 106, 522–534. [Google Scholar] [CrossRef]
  47. Hajinoroozi, M.; Zhang, J.M.; Huang, Y. Driver’s fatigue prediction by deep covariance learning from EEG. In Proceedings of the 2017 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Banff, AB, Canada, 5–8 October 2017; pp. 240–245. [Google Scholar]
  48. Saha, P.; Fels, S.; Abdul-Mageed, M. Deep learning the eeg manifold for phonological categorization from active thoughts. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 2762–2766. [Google Scholar]
  49. Duma, G.M.; Mento, G.; Manari, T.; Martinelli, M.; Tressoldi, P. Driving with intuition: A preregistered study about the EEG anticipation of simulated random car accidents. PLoS ONE 2017, 12, e0170370. [Google Scholar] [CrossRef]
  50. Zhang, H.; Chavarriaga, R.; Khaliliardali, Z.; Gheorghe, L.; Iturrate, I.; Millán, J.d.R. EEG-based decoding of error-related brain activity in a real-world driving task. J. Neural Eng. 2015, 12, 066028. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Kim, J.W.; Kim, I.H.; Lee, S.W. Detection of braking intention during simulated driving based on EEG analysis: Online study. In Proceedings of the 2015 IEEE International Conference on Systems, Man, and Cybernetics, Hong Kong, China, 9–12 October 2015; pp. 887–891. [Google Scholar]
  52. Ary, J.P.; Klein, S.A.; Fender, D.H. Location of sources of evoked scalp potentials: Corrections for skull and scalp thicknesses. IEEE Trans. Biomed. Eng. 1981, 447–452. [Google Scholar] [CrossRef]
  53. Haufe, S.; Meinecke, F.; Görgen, K.; Dähne, S.; Haynes, J.D.; Blankertz, B.; Bießmann, F. On the interpretation of weight vectors of linear models in multivariate neuroimaging. Neuroimage 2014, 87, 96–110. [Google Scholar] [CrossRef] [Green Version]
  54. Statistics, X. A forerunner of the t-distribution. Biometrika 1996, 83, 891–898. [Google Scholar]
Figure 1. The framework of our model: In the SR domain, we fused the output of the SPD network and the distance information based on the Stein divergence as SR-domain features. In the TR domain: we fed the covariance matrices into a two-layer RNN with seven LSTM cells per layer, and define the output as TR-domain features. Then, the concatenated SR-domain and TR-domain features were transformed into linear features by the fully connected layer, and the features were used to give the final prediction by the softmax layer.
Figure 1. The framework of our model: In the SR domain, we fused the output of the SPD network and the distance information based on the Stein divergence as SR-domain features. In the TR domain: we fed the covariance matrices into a two-layer RNN with seven LSTM cells per layer, and define the output as TR-domain features. Then, the concatenated SR-domain and TR-domain features were transformed into linear features by the fully connected layer, and the features were used to give the final prediction by the softmax layer.
Entropy 23 01298 g001
Figure 2. Architecture of the SPD network: The bilinear map layer consists of a transformation matrix W with full rank and its transpose. ζ of the second layer indicates a diagonal matrix, whose diagonal elements are the threshold value we set. In the logarithmic eigenvalue layer, we apply eigendecomposition to get the eigenvalues and eigenvectors of matrix X i , which are Λ i and U i .
Figure 2. Architecture of the SPD network: The bilinear map layer consists of a transformation matrix W with full rank and its transpose. ζ of the second layer indicates a diagonal matrix, whose diagonal elements are the threshold value we set. In the logarithmic eigenvalue layer, we apply eigendecomposition to get the eigenvalues and eigenvectors of matrix X i , which are Λ i and U i .
Entropy 23 01298 g002
Figure 3. The scheme of EEG segmentation: (a) For time series of each trial, a 3 s sliding time window and a 1 s step are used to divide the EEG signals. (b) From the segmented result, the relations between different channels are captured by the covariance matrices we calculated, and we reshape them into vectors with one dimension, separately. The covariance value between the i-th channel and the j-th channel is indicated by the element C ( i , j ) of the covariance matrix.
Figure 3. The scheme of EEG segmentation: (a) For time series of each trial, a 3 s sliding time window and a 1 s step are used to divide the EEG signals. (b) From the segmented result, the relations between different channels are captured by the covariance matrices we calculated, and we reshape them into vectors with one dimension, separately. The covariance value between the i-th channel and the j-th channel is indicated by the element C ( i , j ) of the covariance matrix.
Entropy 23 01298 g003
Figure 4. The features of the TR domain are processed by a 2-layer RNN with 7 LSTM units per layer, where m i means the flattened matrix of the i-th segment signals, and h i indicates the i-th hidden state. The h 7 denotes the output of the RNN.
Figure 4. The features of the TR domain are processed by a 2-layer RNN with 7 LSTM units per layer, where m i means the flattened matrix of the i-th segment signals, and h i indicates the i-th hidden state. The h 7 denotes the output of the RNN.
Entropy 23 01298 g004
Figure 5. Simulated driving paradigm: Each lane-departure event is considered as the period from the first response offset to the second response offset, which contains the deviation onset and response onset of the current event. The time between the deviation onset and the response onset is defined as reaction time(RT), which is the main focus of our study. EEG signals were recorded during the whole task.
Figure 5. Simulated driving paradigm: Each lane-departure event is considered as the period from the first response offset to the second response offset, which contains the deviation onset and response onset of the current event. The time between the deviation onset and the response onset is defined as reaction time(RT), which is the main focus of our study. EEG signals were recorded during the whole task.
Entropy 23 01298 g005
Figure 6. Comparison of the average RTs of 27 subjects in different driving states. The histogram with blue color (V_RT) represents the reaction time of the vigilant state, while the orange one (F_RT) represents the reaction time of the fatigue state. The horizontal dotted lines indicate the average RT of all subjects in different states. Besides, we marked the variance of reaction time, which are 0.775 and 0.043 of fatigue state and vigilant state, respectively.The average reaction time distribution of 27 subjects in the vigilant and drowsy driving state. The green histogram (V_Res) represents the response time of vigilant driving, while the purple histogram (F_Res) represents the response time of drowsy driving. The red and black horizontal dotted lines represent the average response time across all subjects in drowsy and vigilant driving, respectively. Besides, the variance of response time in drowsy and vigilant driving is 0.775 and 0.043.
Figure 6. Comparison of the average RTs of 27 subjects in different driving states. The histogram with blue color (V_RT) represents the reaction time of the vigilant state, while the orange one (F_RT) represents the reaction time of the fatigue state. The horizontal dotted lines indicate the average RT of all subjects in different states. Besides, we marked the variance of reaction time, which are 0.775 and 0.043 of fatigue state and vigilant state, respectively.The average reaction time distribution of 27 subjects in the vigilant and drowsy driving state. The green histogram (V_Res) represents the response time of vigilant driving, while the purple histogram (F_Res) represents the response time of drowsy driving. The red and black horizontal dotted lines represent the average response time across all subjects in drowsy and vigilant driving, respectively. Besides, the variance of response time in drowsy and vigilant driving is 0.775 and 0.043.
Entropy 23 01298 g006
Figure 7. Scatter diagram of RT values of all trials: The horizontal axis represents the value of local reaction time, and the vertical axis represents the value of global reaction time; 739 trials were labeled as vigilant states, which were marked as purple points in the lower left corner; 694 orange points in the upper right corner indicated trials with fatigue state.
Figure 7. Scatter diagram of RT values of all trials: The horizontal axis represents the value of local reaction time, and the vertical axis represents the value of global reaction time; 739 trials were labeled as vigilant states, which were marked as purple points in the lower left corner; 694 orange points in the upper right corner indicated trials with fatigue state.
Entropy 23 01298 g007
Figure 8. Trajectories of variations in the inter-channel relations over time in the situation of 3, 5, 7 channel pairs. In each graph (af), the horizontal axis represents the 7 successive time windows, and the vertical axis indicates the relation values between different EEG channels, which are the covariance of two channels. As seen, the curve shapes within the same class are very similar to each other, whereas significant dissimilarities exist between the different classes. For example, graphs (a,d) show the dynamic variations of the same combination of 3 EEG channel pairs in M V and M D , respectively.
Figure 8. Trajectories of variations in the inter-channel relations over time in the situation of 3, 5, 7 channel pairs. In each graph (af), the horizontal axis represents the 7 successive time windows, and the vertical axis indicates the relation values between different EEG channels, which are the covariance of two channels. As seen, the curve shapes within the same class are very similar to each other, whereas significant dissimilarities exist between the different classes. For example, graphs (a,d) show the dynamic variations of the same combination of 3 EEG channel pairs in M V and M D , respectively.
Entropy 23 01298 g008
Figure 9. The box plots of classification accuracy of different numbers of channel pairs, which are 3, 5, 7 and all pairs. The p-values of the significance test are marked below the box plots.
Figure 9. The box plots of classification accuracy of different numbers of channel pairs, which are 3, 5, 7 and all pairs. The p-values of the significance test are marked below the box plots.
Entropy 23 01298 g009
Figure 10. Diagram of brain state analysis based on covariance: (a) We assumed that the electrical activity is the matrix with the dimensions of 364 × 2250 . (b) We calculated the weight matrix set A of fatigue and vigilant states, respectively. The histograms of the third column represent the weight assignment of all channels by the same neuron in two brain states.
Figure 10. Diagram of brain state analysis based on covariance: (a) We assumed that the electrical activity is the matrix with the dimensions of 364 × 2250 . (b) We calculated the weight matrix set A of fatigue and vigilant states, respectively. The histograms of the third column represent the weight assignment of all channels by the same neuron in two brain states.
Entropy 23 01298 g010
Figure 11. Fitted t-distribution diagrams of the weight assignment of all channels by the brain neurons, in which the blue curve denotes the fatigue state and the yellow curve denotes the vigilant state. The normalized KL divergences between two distributions are on the top of each diagram.
Figure 11. Fitted t-distribution diagrams of the weight assignment of all channels by the brain neurons, in which the blue curve denotes the fatigue state and the yellow curve denotes the vigilant state. The normalized KL divergences between two distributions are on the top of each diagram.
Entropy 23 01298 g011
Table 1. Defined measurement of selected combinations of EEG channel pairs.
Table 1. Defined measurement of selected combinations of EEG channel pairs.
Γ Θ Φ ζ
2_pairs_combination_11.02981.05210.030268.9370
2_pairs_combination_21.06031.08440.077329.7875
2_pairs_combination_31.05141.01000.159512.9241
3_pairs_combination_11.09591.09600.023892.0631
3_pairs_combination_21.06651.10070.052332.9318
3_pairs_combination_31.06201.08590.102320.8332
5_pairs_combination_11.05001.07620.059134.7570
5_pairs_combination_21.05451.08470.065131.8253
5_pairs_combination_31.08751.06860.090723.5354
7_pairs_combination_11.06511.07040.067436.1222
7_pairs_combination_21.07221.08060.059631.3334
7_pairs_combination_31.06511.08340.093122.7714
Table 2. Comparison of our method with other methods in terms of various evaluation criteria (accuracy, specificity, sensitivity, and F1 score).
Table 2. Comparison of our method with other methods in terms of various evaluation criteria (accuracy, specificity, sensitivity, and F1 score).
Accuracy (%)Specificity (%)Sensitivity (%)F1 (%)
C O V _ C N N Mean81.877 **74.149 **72.435 **77.204 **
Variance3.18912.41810.7756.001
C N N _ R N N Mean79.103 **73.036 **72.487 **74.749 **
Variance5.02813.05910.6687.476
F R E Mean75.155 **70.708 **74.218 **73.047 **
Variance3.89412.87810.6406.882
T R D C Mean86.238 **74.577 **80.719 **80.966 *
Variance2.37711.98710.3757.741
S D T R Mean89.280 *78.17782.327 * 85.368
Variance1.71110.8918.5685.077
S N T R Mean 91.042 72.571 90.800 81.321
Variance1.36914.0134.6297.118
O u r M e t h o d Mean93.83483.16885.86385.686
Variance0.90212.3988.9487.847
(p:obtained with paired samples t-test; *: p 0.05 , **: p 0.01 ).
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, N.; Lu, D.; Hou, K.; Chen, M.; Wei, X.; Zhang, X.; Hu, B. Fatigue Detection with Spatial-Temporal Fusion Method on Covariance Manifolds of Electroencephalography. Entropy 2021, 23, 1298. https://doi.org/10.3390/e23101298

AMA Style

Zhao N, Lu D, Hou K, Chen M, Wei X, Zhang X, Hu B. Fatigue Detection with Spatial-Temporal Fusion Method on Covariance Manifolds of Electroencephalography. Entropy. 2021; 23(10):1298. https://doi.org/10.3390/e23101298

Chicago/Turabian Style

Zhao, Nan, Dawei Lu, Kechen Hou, Meifei Chen, Xiangyu Wei, Xiaowei Zhang, and Bin Hu. 2021. "Fatigue Detection with Spatial-Temporal Fusion Method on Covariance Manifolds of Electroencephalography" Entropy 23, no. 10: 1298. https://doi.org/10.3390/e23101298

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop