An Adaptive Multi-Graph Neural Network With Multimodal Feature Fusion Learning For MDD Detection

Download as pdf or txt
Download as pdf or txt
You are on page 1of 14

www.nature.

com/scientificreports

OPEN An adaptive multi-graph neural


network with multimodal feature
fusion learning for MDD detection
Tao Xing1,2,4, Yutao Dou2,4, Xianliang Chen3, Jiansong Zhou3, Xiaolan Xie1 &
Shaoliang Peng2
Major Depressive Disorder (MDD) is an affective disorder that can lead to persistent sadness and
a decline in the quality of life, increasing the risk of suicide. Utilizing multimodal data such as
electroencephalograms and patient interview audios can facilitate the timely detection of MDD.
However, existing depression detection methods either consider only a single modality or do not fully
account for the differences and similarities between modalities in multimodal approaches, potentially
overlooking the latent information inherent in various modal data. To address these challenges,
we propose EMO-GCN, a multimodal depression detection method based on an adaptive multi-
graph neural network. By employing graph-based methods to model data from various modalities
and extracting features from them, the potential correlations between modalities are uncovered.
The model’s performance on the MODMA dataset is outstanding, achieving an accuracy (ACC) of
96.30%. Ablation studies further confirm the effectiveness of the model’s individual components.The
experimental results of EMO-GCN demonstrate the application prospects of graph-based multimodal
analysis in the field of mental health, offering new perspectives for future research.

Keywords MDD detection, GCN, Multimodal

MDD is a severe mental illness1. Patients with MDD often experience a lack of energy, loss of interest in life,
and low mood. In severe cases, delusions or symptoms of hallucinations and auditory hallucinations may
occur2. Severe depression negatively impacts daily life, work, sleep, diet, and physical health. Among adults
with depression, 2-8% die by suicide3. Moreover, studies4 indicate that in high-income countries, 70-80% of
individuals who die by suicide have a mental illness, with depression being the most common cause. In low-
and middle-income countries, this proportion is about half. In recent years, the global prevalence of MDD
has increased by 28%, with an actual count of 246 million cases5. Many studies6 show that timely therapeutic
intervention can help alleviate the worsening of MDD.Therefore, there is an urgent need for an efficient and
reliable method for detecting MDD.Currently, the clinical diagnosis of MDD primarily relies on psychiatric
interviews and various depression rating scales to measure the severity of depression5, such as Patient Health
Questionnaire (PHQ-9).However, the diagnosis of MDD primarily relies on patient self-reported symptoms and
the interpretations of clinicians, and variations in MDD across individuals7 may lead to misjudgments about the
patient’s condition.
Existing studies89 indicates that MDD patients often speak slowly, with numerous pauses, and their content
tends to be negative and lacking in energy89. It has also been found that the electroencephalogram (EEG)
waveforms of MDD patients show significant differences from those of healthy individuals10, highlighting the
potential of EEG and voice analysis in the field of depression detection. To improve the effectiveness of MDD
detection, many studies now utilize machine learning or deep learning to identify depression in patients. The
inherent complexity of physiological signals poses a significant challenge for traditional machine learning
methods1112, which often rely on manually extracted features. The crucial question is whether these handcrafted
features possess sufficient discriminative power to enable traditional machine learning algorithms, such as
Support Vector Machines (SVM), to effectively differentiate between different categories of individuals.

1College of Computer Science and Engineering, Guilin University of Technology, Guilin 541006, China. 2College of
Computer Science and Electronic Engineering, Hunan University, Changsha 410082, China. 3Hunan Key Laboratory
of Psychiatry and Mental Health, Department of Psychiatry, National Clinical Research Center for Mental Disorders,
National Center for Mental Disorders, National Technology Institute on Mental Disorders, The Second Xiangya
Hospital of Central South University, Changsha 410011, China. 4These authors contributed equally: Tao Xing and
Yutao Dou. email: xie_xiao_lan@foxmail.com; slpeng@hnu.edu.cn

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 1


www.nature.com/scientificreports/

Current deep learning methods have shown certain limitations in extracting latent information from data.
For instance, in studies dealing with EEG data, a common practice is to select only a few specific channels
for analysis13, which may lead to overlooking important information contained in other channels, thereby
affecting the comprehensiveness and accuracy of the final analysis. On the other hand, when using audio data
as input, some studies tend to adopt Long Short-Term Memory networks (LSTM)1415 or Convolutional Neural
Networks (CNN)16 as the primary architecture. Although these models perform well in handling time-series
data and extracting local features, they may not fully mine and utilize global information and deeper features in
audio data. This localized approach to information processing can limit the overall performance of the model.
Therefore, to enhance the capabilities of deep learning models in the field of MDD detection, it is necessary to
develop more advanced techniques and methods to more comprehensively and deeply analyze and utilize the
information within these complex physiological data.
To tackle the challenges mentioned above, this article introduces a new multimodal graph neural network
method for depression detection, named EMO-GCN. This method designs a set of graph neural networks for
each modality to model the relationships within the data. By fusing features extracted from each modality into a
multimodal embedding feature and introducing an attention mechanism on this embedding, the model focuses
on the most important features in each modality. Finally, the attention-featured embedding is input into a
classification network for subsequent MDD detection tasks. The main contributions of this article are as follows:

1. We propose a multimodal depression detection framework called EMO-GCN, which uses multiple graph
convolutional networks to extract structural features from EEG signals and acoustic features from speech,
achieving effective multimodal feature fusion. These fused multimodal representations provide more accu-
rate indicators for depression detection, thereby significantly enhancing performance.
2. We propose a Multi-GCN module that combines stacked graph convolution and graph pooling layers, in-
troducing a structural learning mechanism that reconstructs the graph structure through sparse attention
after pooling. This approach accurately captures complex graph features while preserving node relationships,
enabling efficient representation of data characteristics.
3. We evaluate our proposed method on the MODMA public dataset through extensive experiments. The re-
sults demonstrate that our approach outperforms existing baseline algorithms in the depression detection
task, achieving an accuracy of 96.30%, with its effectiveness further validated through comprehensive abla-
tion studies.The rest of this paper is structured as follows. “Related work” briefly discusses related work and
technologies in the task of MDD detection. In “Methodology”, we provide detailed information about the
method we propose. After that, “Experiments setup” introduces our experimental design and settings and
describes the datasets. In “Results”, findings from comparison experiments with other models and the abla-
tion study are presented. Then, in “Discussion”, we discuss the current limitations of our work and propose
some possible future directions. Finally, “Conclusion” concludes the paper.

Related work
Traditional depression detection
In the field of clinical depression diagnosis, there are already many detection methods. The most common methods
involve psychological tests or questionnaires. For example, the PHQ-9 is a commonly used screening tool for
depression17. The PHQ-9 contains 9 questions, mainly asking about the frequency of depressive symptoms over
the past two weeks and can be completed in about 5 minutes. Another commonly used clinical assessment scale
is the HAMD18. It is a clinician-administeblack tool consisting of 17 items to assess the severity of depressive
symptoms, with scores ranging from 0 to 52. The higher the score, the more severe the depression. The drawback
of traditional questionnaire methods is their subjectivity and variability19.Additionally, these scales have
limitations20. When using self-assessment scales like PHQ-9, there can be differences in understanding among
participants, which may lead to varying scores. With other-assessment scales like HAMD, different doctors may
give inconsistent scores for the same patient. Therefore, it’s challenging to accurately detect whether a subject has
depression solely through the use of scales.
To objectively diagnose depression, machine learning is increasingly being used to identify depression.
For example, Deshpande et al.21 analyzed tweets collected using Naive Bayes and SVM methods, identifying
potential depressive moods among a large set of user data. However, significant noise in the collected tweets
before preprocessing, such as third-person references and news quotes, led to the elimination of about one-
third of the data. Islam et al.22 utilized SVM, Decision Tree, KNN, and other methods to build predictive
models for identifying and processing emotion data related to depression in Facebook posts. Despite the use of
various machine learning techniques, the accuracy rates were only between 60% to 80%, indicating room for
improvement.

Graph neural network detection method


Graph Neural Network(GNN) was first introduced by Scarselli et al.23, defining them as a type of recursive
neural network capable of directly operating on graph-structured data. GNN are gradually gaining traction in
the medical field as well. Zhao et al.24 proposed ECGNN, composed of a feature extractor backbone and a GNN
module, which extracts electrocardiogram (ECG) features for the diagnosis of cardiovascular diseases. Wang et
al.25 introduced the MGREL, integrating knowledge extraction and graph learning channels. This model uses
graph representation learning to acquire network topology representations for predicting associations between
genes and diseases. The study by Fritz et al.26 combines GNN with epidemiological models to enhance the
predictive accuracy of weekly COVID-19 cases across various regions in Germany.
GNN have also been applied in depression detection. Yu et al.27 proposed a method based on GNN that
combines temporal and spatial features of functional near-infrared spectroscopy (fNIRS) data for automatic

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 2


www.nature.com/scientificreports/

depression recognition. Sun et al.28 constructed two GNN modules sequentially to explore latent connections
within and between audio signals, providing relevant cues for detecting depression in the model. Luo et al.29
designed a Graph Convolutional Gated Recurrent Unit (GCGRU) module to capture the temporal dynamic
changes in brain networks within EEG data, further extracting differential features between depressed individuals
and healthy controls.

Multimodal depression detection


With the advancement of computer technology, depression detection methods have evolved from unimodal to
multimodal approaches, particularly combining speech and EEG data to capture a more comprehensive set of
multidimensional features in individuals with depression. Qayyum et al.13 integrated different levels of speech
and EEG features and applied visual transformers and various pre-trained networks, significantly improving
the diagnostic effectiveness for patients with depression. Zheng et al.30 proposed a novel time-convolutional
transformer with knowledge embedding to link audio and EEG, capturing effective features and enhancing
the performance of depression detection models. Addressing both heterogeneity and homogeneity between
the speech and EEG modalities, Chen et al.31 proposed a multimodal fusion strategy based on graph neural
networks to explore potential relationships between samples. Current research indicates that multimodal
methods demonstrate clear advantages in depression detection, as integrating data from different modalities,
such as audio and EEG, enables a more comprehensive reflection of the emotional and behavioral characteristics
of individuals with depression. Although some progress has been made with multimodal depression detection
methods, there remains a limited number of methods in this area, particularly those based on EEG and speech.
Existing approaches also have certain limitations. For example, Chen et al.’s31 method introduces a reconstruction
network, which increases model complexity and computational burden and poses a higher risk of overfitting
when data samples are limited. In Qayyum et al.’s13 method, EEG and speech data undergo numerous processing
and feature extraction steps, including spectrogram generation and temporal correlation extraction; while these
steps help improve classification accuracy, they also increase computational complexity.

Methodology
In this section, we will introduce EMO-GCN, which consists of a graph feature vector extraction module for
each modality and a graph neural network module.Figure1 shows the framework of EMO-GCN. The graph
feature vector extraction module takes EEG and audio data from patients as inputs, with the data undergoing
preprocessing and feature extraction to produce initial graph feature matrices for each modality. The graph neural
network module learns from these initial graph feature matrices, generating new subgraphs through iterative
execution of graph convolution, graph pooling, and structural learning mechanisms. Node features within the
subgraphs are aggregated to produce fixed-size graph embedding vectors. Finally, the graph embedding vectors
from both modalities are fused and an attention mechanism is introduced before feeding them into a fully
connected layer to obtain the final detection results.

Problem definition
Given a multimodal dataset {Di}N i=1 , comprising N multimodal patient samples, {Di}N i=1 includes audio data
N ×Rowa ×Cola N ×Rowe ×Cole
a
M ∈R and EEG data M ∈ R e
, along with one-hot labels for each sample Y ∈ RN ×C
. Here, Rowa and Rowe respectively represent the number of sample points in the audio data and the number
of electrodes in the EEG data, where the sample points equal the audio duration multiplied by the sampling
rate; Cola and Cole respectively represent the feature dimensions of M a and M e; and C represents the number
of label categories. In this article, our objective is to input the multimodal dataset {Di}N i=1 into the model, and
subsequently output the detection results for each sample Ŷi ∈ R1×C (i = 1, 2, 3, · · · ).

Figure 1. Framework of EMO-GCN.

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 3


www.nature.com/scientificreports/

For each sample in M e, obtained using an EEG device with ne electrodes, we construct an arbitrary graph
Ge = (ν e, εe, X e), where ν e, εe, and X e ∈ Rne×me respectively represent the nodes, edges, and feature matrix
of Ge, with me denoting the dimension of X e. Also, let Ae ∈ Rne×ne be the adjacency matrix representing the
graph connectivity information, where the element Ae(j, k) indicates the connection status between the j-th
and k-th electrodes. If these electrodes are connected, then Ae(j, k) = 1; otherwise, Ae(j, k) = 0. Similarly, for
each sample in M a, consisting of na audio segments, we construct a graph Ga = (ν a, εa, X a), where ν a, εa,
X a ∈ Rna×ma respectively represent the nodes, edges, and feature matrix in Ga. The connectivity information of
the nodes in Ga is stored in the adjacency matrix Aa ∈ Rna×na .

Graph construction
In this subsection, we describe the process of constructing graph structures for M e and M a.
The spatial distribution of electrodes in the EEG collection device can help us preliminarily construct the
graph structure of the EEG. This allows for accurate modeling of local activities in various brain regions. By
considering each electrode as a node in Ge and connecting corresponding nodes based on the spatial distribution
of electrodes, we can obtain the initial adjacency matrix Aeinit. Additionally, a study focusing on the EEG
symmetry in patients with depression found that analyzing the EEG symmetry of homologous regions in the
left and right brain is significant for the diagnosis of depression32. Inspired by this, we empirically selected some
symmetric electrodes from the left and right hemispheres and connected them, resulting in a symmetrically
distributed electrode adjacency matrix Aesym. The final adjacency matrix Ae for M e is derived using the formula
as Ae = Aeinit + Aesym.Figure 2 shows the graph structure of EEG data. The resulting adjacency matrix Ae can
express the connectivity between brain localities and between the left and right hemispheres. Such an adjacency
matrix not only helps in exploring the associations between local regions of the brain but also assists the graph
neural network module in learning and recognizing cross-hemispheric brain activity patterns. It can provide a
more comprehensive representation of brain activity data, which is beneficial for training a more accurate model.
The temporal nature of M a is key to constructing its graph structure.The audio samples are cut into slices
of equal time length, and each slice is regarded as a node. The adjacency matrix Aa ∈ Rna×na for M a can be
obtained using the following formula:

1 if |j − k| = 1
Aa(j, k) = (1)
0 otherwise

When |j − k| = 1, it indicates that the j-th and k-th nodes are adjacent. In this case, we set the element in Aa
corresponding to the connection between these two nodes to 1, signifying that the adjacent nodes are connected.
For other cases, the element in Aa representing the connection between the j-th and k-th nodes is set to 0. This
graph structure naturally reflects the temporal sequence of M a and is capable of capturing changes in the audio
signal over different time periods. A sequentially connected graph structure can also assist in exploring the
emotional changes of patients over a period of time. This is helpful for analyzing the temporal information in
M a and learning complex audio features.

Graph feature extraction module


The feature extraction module converts raw EEG signals and audio data into graph-level feature matrices. These
matrices serve as the initial graph features for training and learning within the graph neural network. Since the

Figure 2. Local connection and symmetrical connection.The red dotted line represents symmetrical
connections, while the black solid line represents local connections

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 4


www.nature.com/scientificreports/

two modalities have their own characteristics, we have designed suitable feature extraction modules for each
modality separately.
Due to the high-dimensionality characteristic of M e, reflected in the data containing a large number of
electrode channels, each electrode collects a substantial amount of data, involving multi-dimensional time-series
signal acquisition. Excessively high dimensions are not conducive to model fitting and also require more time for
training. To preserve the complete electrode channels while addressing this challenge, we use the GraphSAGE
algorithm to perform dimensionality reduction on the original M e. This involves learning node embeddings
by subsampling neighboring node features on the graph dimension. This approach effectively reduces the data
dimensionality, decreasing the computational training load. Given an EEG signal data Mie = {e1, e2, · · · , ene }
with ne channels, and its graph structure Ge = (ν e, εe), we first calculate the average feature of the neighboring
nodes for each node. For each node v and its set of neighboring nodes N(v), the neighbor average feature hN (v) is
computed as hN (v) = mean{eu, ∀u ∈ N (v)}. After obtaining hN (v), we concatenate each node’s feature Xve with
the average feature of its neighbors to get hv , which is given by hv = ev ∥ hN (v). Subsequently, hv is fed into a
linear layer and an activation function is applied. Let W and b represent the weight and bias of the linear layer,
respectively. The new feature representation h′v is obtained using the following formula:

h′v = ReLU(W · hv + b)(2)

Finally, the h′v calculated



for all nodes are combined to form a matrix, resulting in the initial graph feature
matrix X0e ∈ Rne×me for the subsequent graph neural network. Here, m′e is the size of the feature vector after the
feature extraction process. The initial graph feature matrix obtained through dimensionality reduction retain all
channels, enabling the graph neural network to explore the associations between channels.

Although both M a and M e are waveform data, when processing M a, it is crucial to focus on the patient’s
emotional changes from a speech perspective. Features obtained from a perspective different from EEG can
complement other aspects of MDD characteristics in a single modality. Assuming that each sample in M a
has na audio segments, we compute various sound features in each segment that are effective for emotion
recognition, such as Mel Frequency Cepstral Coefficients (MFCC), pitch, Root Mean Square (RMS) energy,
and Mel-spectrogram(Mel). Let the j -th feature of the i-th audio segment be denoted as fij , leading to an audio
feature row ri = [fi1, fi2, . . . , fim′a ], where m′a is the size of the feature vector after extraction. Subsequently, we
concatenate all the obtained′ audio feature rows along the Y-axis to form the feature matrix. The audio emotion
feature matrix X0a ∈ Rna×ma , obtained through the calculation of audio emotional features, facilitates a deeper
understanding of the relationship between speech and emotion in subsequent analyses.

Multi-GCN component
In the Multi-GCN component, a graph convolution layer and a graph pooling layer together form a basic unit.
By stacking such units in three layers, we have constructed the complete Multi-GCN. Such a structure is capable
of capturing more complex and abstract features of the graph structure. Figure 3 shows the workflow of Multi-
GCN component.

Graph convolutional network


Graph Convolutional Network (GCN)33 is a powerful method for learning and capturing graph structures,
which can effectively understand both local and global relationships within the graph to enhance prediction and
classification accuracy. The initial graph feature matrix of M e and M a are obtained through the feature extraction
module. If traditional LSTM or CNN methods are used to process the feature matrix, they fail to capture and
learn the complex relationships and dependencies between nodes of M e and M a at the graph level. GCNs, on
the other hand, have a unique advantage in scenarios that require capturing and utilizing complex relationships
between nodes. The key idea of GCN is to update the feature representation of each node by aggregating the
feature information of adjacent nodes. The basic operation of GCN can be represented by the following formula:
 1 1

H (l+1) = σ D− 2 AD− 2 H (l)W (l) (3)

Figure 3. Workflow of multi-GCN component.

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 5


www.nature.com/scientificreports/

Where H (0) = X 0, H (l) is the node feature representation of the l-th layer, H (l+1) is the node feature representation
of the (l + 1)-th layer, A is the adjacency matrix of the graph, D is the 1diagonal1
matrix, W (l) is the weight matrix
of the l-th layer, and σ is the activation function. In this formula, D− 2 AD− 2 is responsible for normalizing the
adjacency matrix to maintain numerical stability when aggregating the features of neighboring nodes. H (l)W (l)
represents feature transformation, i.e., mapping the current features to a new feature space through the weight
matrix. Finally, the activation function σ provides a non-linear mapping, allowing the network to learn complex
patterns.

Graph pooling
Graph structure data may contain noise and information irrelevant to MDD detection. We employ graph pooling
to remove redundant information from the data. The purpose of graph pooling is to reduce the number of nodes
in a graph neural network while retaining the most important information as much as possible. First, a score is
assigned to each node, which is calculated using the following formula:
  
 
score =  I − D−1A H (i)(4)

Where I is the identity matrix, D is the diagonal matrix, D−1 is the inverse of D, and H (i) is the feature matrix of
the nodes. This formula evaluates the information content of a node by calculating the difference between each
node and its neighbors. This difference is computed through the expression (I − D−1A)H (i), which essentially
looks at the difference between the features of a node and those of its neighbors. After calculating the information
score of each node, the nodes with the highest scores are selected according to the calculated scores, formula
described as idx = top(score). Subsequently, the node feature matrix and adjacency matrix are updated based
on the high-scoring nodes selected:

H̃ (i+1) = H (i)(idx, :)
(5)
Ã(i+1) = A(i)(idx, idx)

H (i)(idx, :) represents selecting the features corresponding to the highest scoring nodes from the original node
feature matrix H (i). A(i)(idx, idx) indicates updating the adjacency matrix according to these nodes, meaning
retaining the connections between these high-scoring nodes. Graph pooling is used to reduce the number of
nodes and edges in Ge and Ga. Identifying and preserving information that is effective for the MDD detection
task can reduce the negative impact of irrelevant or misleading information on MDD detection performance.

Structure learning
In our framework, we constructed graph structures for M e and M a separately. However, graph pooling may lead
to the disconnection of originally closely related nodes in the subgraph, thereby losing the integrity of the graph
structure information and hindering the message passing process. To learn a refined graph structure after graph
pooling, encoding the underlying pairwise relationships between nodes, we introduced a structure learning
mechanism following the graph pooling operation. This mechanism learns a sparse graph structure through
a sparse attention mechanism. For the subgraph Gki obtained after the k -th layer pooling of graph Gi, we take
nki ×nki k
its adjacency matrix information Ai ∈ R
k
and node feature matrix Hik ∈ Rni ×d as inputs. The structure
learning mechanism is implemented through a single-layer neural network, parameterized by the weight vector
a ∈ R1×2d. The similarity score between nodes vp and vq can be represented as follows:

Eik (p, q) = σ(a[Hik (p, :) || Hik (q, :)]T ) + λ · Aki (p, q)(6)

In the above formula, σ(·) denotes the activation function. Hik (p, :) and Hik (q, :) represent the p-th and q -th
rows of the matrix Hik , respectively, corresponding to nodes vp and vq . When Aki (p, q) > 0, it indicates a direct
connection between the two nodes, and the attention mechanism tends to assign higher similarity scores to
directly connected nodes. At the same time, it attempts to learn the underlying pairwise relationships between
directly connected nodes and nodes that are not directly connected, with λ being the balancing parameter
between the two. To make the similarity scores between different nodes more easily comparable, the structure
learning mechanism uses the softmax function for normalization:

exp(Eik (p, q))


Sik (p, q) =  k (7)
ni k
m=1 exp(Ei (p, m))

The non-zero values resulting from the softmax transformation can lead to a densely connected graph,
introducing a significant amount of noise. To address this issue, we use the sparsemax function, which converts
the original softmax function into a piecewise function. This approach avoids introducing noise in the structure
learning process:

Sik (p, q) = sparsemax(Eik (p, q))


  (8)
sparsemax(Eik (p, q)) = Eik (p, q) − τ (Eik (p, :)) +

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 6


www.nature.com/scientificreports/

In the above formula, [x]+ = max{0, x}. The function τ (·) determines the threshold for the sparsemax
function transformation, thereby producing a sparse distribution. Through the aforementioned operations on
the subgraphs, the model can learn a more refined subgraph structure. This ensures the integrity of the graph
structure information after graph pooling operations, while not introducing additional noise. To generate a fixed-
size graph-level embedding, we concatenate the results of mean pooling and max pooling for each subgraph,
aggregating the representations of all nodes within the subgraph:
 k 
ni
1  d
rik =  H k (p, :) || max Hik (:, q)(9)
σ p=1 i q=1

Then, we sum up the node representations from different subgraphs to obtain a graph neural network embedding
fg = ri1 + ri2 + . . . + riK .

In Multi-GCN, we alternated between graph convolutional networks and graph pooling. The graph convolutional
network effectively extracts relational features between nodes by propagating information between the nodes in
Ge and Ga. Graph pooling reduces the complexity of the graph, thereby lowering the risk of overfitting and also
minimizing noise in the data. Utilizing a structural learning mechanism after graph pooling can address the
issue of lost node associations in subgraphs resulting from pooling.

Modal fusion
For the initial graph feature vectors X0e and X0a outputted from the graph feature vector extraction module,
we applied attention mechanisms at both the EEG channel-level and the audio slice-level to assign greater
importance to electrode channels and time segments that exhibit characteristics of depression. The attention
scores at the EEG channel-level and the audio slice-level are calculated using the following formulas:

α = softmax(V tanh(W X0 + b))


n
 (10)
fb = αiXi, (n = ne f or EEG, n = na f or audio)
i=1

Where V , W , and b are parameters at the EEG channel-level and audio slice-level learned during training, α
represents the attention scores, and fb ∈ Rm is a context vector that encodes the EEG signals of ne electrode
channels or the audio signals of na audio slices for different modalities. The context vectors fb from each
modality are used for the final results.

We aggregate the representations of all nodes in the subgraphs generated by Multi-GCN from each modality and
sum them up to obtain a fixed-size graph-level representation.
 
l ni
 1  d
fg = σ H (i)(p, :) ∥ max H (i)(:, q)(11)
i=1
n i p=1
q=1

fg ∈ R2d is the final aggregated feature vector, where l represents the number of subgraphs, and the ∥ operation
denotes concatenating the results of mean pooling and max pooling for each subgraph. Subsequently, the two
features fb and fg are concatenated to obtain the final graph embedding feature for each modality individually
fembedding = fb ∥ fg . For the fembedding of each modality, we first concatenate them and then apply a feature-level
attention mechanism to assign greater importance to specific features indicative of depression.

output = AttEEG(fembedding
e
) ∥ AttAudio(fembedding
a
)(12)

Where fembedding
e
and fembedding
a
are the graph embedding features derived from the neural network for the EEG
and audio modalities, respectively. The attention weighting function Attmodality(·) is defined as:

Attmodality(fmodality) = fmodality ⊙ σ(wmodality)(13)

In this context, ⊙ denotes the Hadamard product, and wmodality is learned through training. This attention
mechanism allows the model to dynamically adjust the contributions of the two modalities when processing the
fused features, thereby enhancing the flexibility and effectiveness of handling multimodal data.

Experiments setup
Dataset
The Multimodal Open Dataset for Mental Disorder Analysis (MODMA) offers clinically accurate data
for analyzing mental disorders like depression. It includes data from patients professionally diagnosed with
depression and matched controls. The dataset contains resting-state and stimulus-state EEG data from 51 subjects
(22 with depression, 29 controls), recorded at 250 Hz using a 128-channel HydroCel Geodesic Sensor Net, with
electrode impedance below 50 kΩ. Audio data from the subjects were recorded using a Neumann TLM102

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 7


www.nature.com/scientificreports/

microphone and an RME FIREFACE UCX interface at 44.1 kHz and 24-bit depth, under environmental noise
below 60 dB. Each patient has 29 audio segments. To match the number of samples between audio and EEG data,
we divided the EEG data into 29 segments, resulting in 638 MDD samples and 841 healthy control (HC) samples.
Table 1 presents the demographics of the dataset segmented across different dimensions.

Performance metrics
In the experiments of this paper, we use four key metrics to evaluate the performance of the model: accuracy
(ACC), precision (PRE), recall (REC), and F1 score. These indicators measure model performance in terms
of overall correctness, accuracy of positive class detection, completeness of positive class identification, and
the balance between the two. Combining these four indicators, we can fairly evaluate and optimize model
performance in different application scenarios.

Implementation details
Data splitting: In our experiments, we used a k-fold cross-validation method (k=10) and averaged the 10
validation results to obtain the overall performance metric of the model, thereby validating the model across
different partitions of the dataset to ensure robust generalization ability.
Parameter settings: In the experiments of this article, we used the MODMA dataset as the experimental data.
We denote M e and M a as the 128-channel resting-state EEG data and audio data from the MODMA dataset,
respectively. For the graph structures Ge and Ga corresponding to M e and M a, we set their number of nodes
to ne = 128 and na = 32, respectively. The dimensions m′e and m′a of X0e and X0a, obtained through the graph
feature vector extraction module for the two modalities, are set to 600 and 580, respectively.

Baselines
We present the results of 15 baseline methods for the depression detection task on the MODMA dataset,
covering multimodal, EEG, and audio approaches. Additionally, we include two EMO-GCN variants: EMO-
GCN-α (EEG only) and EMO-GCN-β (audio only).
Multimodal models:

1. MS2-GNN31: offers an effective multimodal fusion approach based on graph neural networks that signifi-
cantly enhances the precision in identifying depression.
2. Ahmed et al.34: proposed a multimodal classifier based on attention mechanisms that combines selective
dropout and normalization techniques to handle missing modalities in different multimodal datasets.
3. EfficientNet13: is a variant of CNN models that scales model depth, width, and resolution in a balanced way,
thereby optimizing performance and computational cost.
4. MobileNet13: is a lightweight CNN model specifically designed for mobile and embedded devices, optimized
for speed and memory usage without significantly affecting accuracy.
5. Hu et al.35: used large language models (LLMs) to perform mental health assessments on multimodal data
through zero-shot and few-shot prompting.EEG models:

1. Tasci et al.36: introduced a novel and computationally light manual feature engineering technique called the
Twin Pascal Triangle Layer Pattern (TPTLP).
2. SGP-SL37: progressively optimizes EEG-based graph structures by utilizing multiple self-attention graph
pooling modules and introduces a soft label strategy to construct the loss function, enhancing the discrimi-
nability of features.
3. Soni et al.38: used the Node2vec algorithm to generate node embeddings of EEG data as features to distin-
guish between patients with depression and healthy subjects.
4. Shen et al.39: proposed an adaptive channel fusion method based on EEG signals, enhancing the separability
of difficult samples by assigning higher weights to their losses through an improved focal loss (FL) function.
5. Sun et al.40: extracted various types of EEG features to comprehensively represent the EEG signals of MDD
patients, using machine learning algorithms and statistical analysis to evaluate these EEG features.Audio
models:

Characteristic Category MDD HC Total


Male 16 20 36
Gender
Female 6 9 15
< 45 19 27 46
Age (years)
≥ 45 3 2 5
≤ 12 12 3 15
Education (years)
> 12 10 26 36
≤ 19 13 29 42
PHQ-9 (score)
> 19 9 0 9

Table 1. Patient demographic information (values are number of people)

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 8


www.nature.com/scientificreports/

1. GNN-SDA41: includes a GNN-based domain alignment module and an uncertainty-guided optimization


module, which respectively achieve multi-domain alignment through an information propagation mecha-
nism and analyze the uncertainty of pseudo-labels to mitigate the adverse effects of noisy predictions.
2. Gheorghe et al.42: performed audio preprocessing, multidimensional feature extraction, and classified sam-
ples using a multilayer perceptron (MLP) and a 1D-CNN.
3. Sun et al.28: constructed a GNN model that integrates the temporal sequence information within audio sig-
nals, the potential associations between different audio pieces, and the extraction of emotional features.
4. Chen et al.43: constructed a large database with 1,479 speech feature samples for modeling. Through 10-fold
cross-validation and algorithm selection, they established a decision tree model for MDD screening.
5. Das et al.44: combined extracted MFCC and spectrogram features into multimodal data based on audio data,
using a CNN model with optimized residual blocks and a “glorot uniform” kernel initializer to identify MDD
patients.

Results
In this section, we evaluate the effectiveness of the proposed EMO-GCN on the MODMA45 dataset, compare
it with existing methods, and conduct a series of ablation experiments. Additionally, we analyze the model’s
attention to EEG electrode channels and audio feature attention.

Comparison with baseline


Table 2 presents the comparative experimental results. From these results, it can be observed that EMO-
GCN-α, which uses EEG data, and EMO-GCN-β , which uses audio data, achieve accuracy rates of 90.06% and
90.48%, respectively. These rates are approximately 6% lower than the multimodal performance of EMO-GCN,
highlighting the value of combining modalities as it enhances the overall accuracy of the model in identifying
MDD. Compared to other multimodal methods in the table, EMO-GCN excels in all evaluation metrics, with an
accuracy of 96.76%, precision of 96.26%, recall of 95.37%, and F1 score of 95.81%. Although the model proposed
by Ahmed et al. achieves a similar accuracy of 95.78%, it falls slightly short in precision and F1 score. Similarly,
pretrained models such as Effnetv2s and Mobile-Net reach accuracies of 93.07% and 83.89%, respectively, while
the MS2-GNN model and the method by Hu et al. show moderate performance, with accuracies of 86.49% and
80.59%. These results clearly emphasize EMO-GCN’s significant advantage in multimodal data analysis.
In the comparison of EEG unimodal methods, the EMO-GCN-α model demonstrates a strong advantage,
achieving the highest scores across all evaluation metrics, with an accuracy of 90.06%, precision of 90.20%, recall
of 88.46%, and an F1 score of 89.32%. While other models perform relatively well-such as the model by Soni et
al., which achieves an accuracy of 88.80%-their overall performance remains slightly lower than that of EMO-
GCN-α. Similarly, the SGP-SL model performs well with an accuracy of 84.91%, but its precision and F1 score
are lower than those of EMO-GCN-α. Other models, such as those proposed by Tasci et al., Shen et al., and Sun
et al., achieve relatively lower scores, further underscoring the advantage of EMO-GCN-α in modeling EEG data
through graph structures and employing multiple GCNs to learn EEG features.
In the audio modality, EMO-GCN-β also outperforms other audio-based models across various metrics,
with accuracy, precision, recall, and F1 scores of 90.48%, 92.36%, 90.48%, and 91.41%, respectively. Although the
model by Das et al. performs similarly in terms of accuracy and precision, it lags slightly in recall and F1 score.

Modality Method ACC(%) PRE(%) REC(%) F1 Score(%)


Multimodal MS2-GNN31 86.49 82.35 87.50 84.85
Ahmed et al.34 95.78 93.45 95.64 94.53
Effnetv2s13 93.07 92.92 91.76 93.92
Mobile-Net13 83.89 78.81 77.94 78.07
Hu et al.35 80.59 - - -
EMO-GCN 96.76 96.26 95.37 95.81
EEG Tasci et al.36 83.96 86.76 76.14 81.10
SGP-SL37 84.91 80.77 87.50 84.00
Soni et al.38 88.80 86.60 87.20 87.10
Shen et al.39 72.25 - 81.88 -
Sun et al.40 84.18 - 78.29 -
EMO-GCN-α 90.06 90.20 88.46 89.32
Audio GNN-SDA41 82.70 82.60 79.20 80.90
Gheorghe et al.42 84.16 85.30 83.80 84.00
Sun et al.28 90.35 88.25 90.33 89.15
Chen et al.43 83.40 83.50 76.80 80.00
Das et al.44 90.47 89.53 89.43 89.47
EMO-GCN-β 90.48 92.36 90.48 91.41

Table 2. Comparison of the performance of EEG and audio data fusion models at different depths of GCN
layers

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 9


www.nature.com/scientificreports/

Likewise, the model by Sun et al. achieves an accuracy of 90.35%, but its precision and F1 score are lower than
those of EMO-GCN-β . Other models, including GNN-SDA and the method by Chen et al., exhibit relatively
average performance, with all metrics significantly lower than those of EMO-GCN-β . These results demonstrate
the effectiveness of applying GCNs to audio data.
Overall, the performance of EMO-GCN in both multimodal and unimodal experiments confirms its superior
capability, making it a powerful model for MDD detection tasks.

Ablation study
We conducted a detailed ablation study on the network layers of the model to assess the contribution of each
part of the model. Given the high dimensionality of EEG signals, we adopted the GraphSAGE technique to
achieve dimensionality reduction. After removing the GraphSAGE layer, the model’s ACC dropped to 77.44%
, indicating that this technique can effectively reduce the signal dimensions while preserving features critical to
differentiating between patients with depression and healthy individuals.
Furthermore, we conducted ablation experiments on the Multi-GCN, as detailed in Table 3. By fixing the
GCN layers of one modality and varying the GCN layers of the other, we explored the optimal number of layers
for the model. The results show that increasing the GCN layers to 4 in the EEG modality, compared to 3 GCN
layers, led to a 4-5% decrease in all metrics. When the GCN layers were increased to 4 in the audio modality,
there was an approximate 20% decline in all performance metrics, suggesting that overly deep GCN layers might
lead to model overfitting. The accuracy of the model also decreased with the reduction of GCN layers in both
modalities. It can also be seen from Table 3 that reducing the GCN layers to 2 for each modality resulted in a
6-7% decrease in all metrics. When the number of GCN layers was reduced to 0, the accuracy metric decreased
by about 17% compared to 3 layers, indicating that fewer GCN layers cannot sufficiently learn signal features.
The experimental results confirm the effectiveness of the proposed three-layer Multi-GCN structure.

EEG electrode channel attention analysis


To investigate the attention of the model proposed in this paper on different EEG electrode channels, we
extracted the attention scores for each electrode channel after training the model and mapped these scores
onto a brain topography map, resulting in a brain topography Fig. 4a that displays the attention scores for the
electrode channels. The areas trending towards red on the chart indicate that the model pays special attention
to the electrode channels in the frontal, parietal, and temporal lobes. Moreover, we conducted a categorical
analysis based on demographic characteristics and health indicators to explore whether the model’s attention to
the electrode channels changes in different contexts. By comparing Fig. 4, we observed significant differences
in the brain topography when the model processes data from individuals of different genders, age groups (with
45 years as the threshold), and education levels (with 12 years of education as the threshold). This indicates
that the model’s focus shifts when dealing with data from individuals with varying demographic backgrounds,
suggesting that there may be differences in the characteristics of depression among different demographic
groups. Additionally, we divided the patients with depression into moderately severe and severe groups based
on their PHQ-9 scores and conducted studies, finding that the model’s focus on electrode channels also differs
between patients with moderate and severe depression.
In our study, the areas of high attention in the Fig. 4a largely coincide with regions that were found to differ
in degree distribution brain topography between MDD patients and healthy subjects in the research by Li et al.46
This finding suggests that the neural network model we proposed can automatically identify differences between
MDD patients and healthy individuals in EEG data, and significantly focus on these regions of pronounced
differences.

Audio feature attention analysis


In this study, the developed model identifies signs of depression by analyzing audio features. To assess the
importance of different audio features, the model assigns attention scores to seven features-MFCC, Pitch, Energy,
Mel-Spectrum, Chroma, Contrast, and Tonnetz. These attention scores offer a quantified means to evaluate the
contribution of each feature to the model’s ability to recognize depression.

Ablation condition GCN LayerCount ACC(%) PRE(%) REC(%) F1 Score(%)


EEG 4 86.11 85.85 84.26 85.05
3 90.06 90.20 88.46 89.32
2 84.72 87.50 84.26 85.85
1 81.94 83.18 82.41 82.79
0 73.15 76.04 67.59 71.57
Audio 4 62.50 68.32 63.89 66.03
3 90.48 92.36 90.48 91.41
2 83.80 88.00 84.15 84.62
1 81.94 83.18 82.41 82.79
0 73.61 72.64 71.30 71.96

Table 3. Comparison of the performance of EEG and audio data fusion models at different depths of GCN
layers

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 10


www.nature.com/scientificreports/

Figure 4. The brain topographic of the attention score.“Ed”, “MS” and “S” denote the education, moderately
severe, and severe, respectively.

As shown in Fig. 5, the Mel feature obtained the highest attention scores, indicating that the model considers
the Mel to be the most important among all the considered sound features for recognizing depression. By
simulating the auditory perception characteristics of the human ear, the Mel can effectively capture the
perceptual content of sounds, making it particularly useful in identifying emotional and mental states. This
result proves that the Mel plays a key role in analyzing emotion-related sound changes. Pitch and energy features
also received relatively high attention scores, suggesting an association between these fundamental acoustic
features and emotional expression. In contrast, MFCC, Chroma, Contrast, and Tonnetz features received lower
attention scores, indicating that in the specific context and dataset of this study, these features contribute le ss
to distinguishing depressive states. However, this does not mean that these features are irrelevant in emotional
analysis; rather, their direct impact on identifying depression is less significant within the framework of this
study.

Discussion
In this study, we developed an innovative method for depression recognition, which relies on a graph neural
network architecture that combines graph convolutional layers, graph pooling operations, and structural
learning to extract key features from multimodal data and effectively integrate these features through an
attention mechanism. We named this method EMO-GNN and conducted a detailed comparative analysis with
existing depression detection methods. The results show that our method consistently demonstrates superior
performance across various scenarios, particularly in multimodal contexts, where EMO-GCN emerged as the
most outstanding model. Furthermore, through a series of ablation experiments, we validated the importance of
each network layer in the Multi-GCN component of our model. The results clearly showed a significant decline

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 11


www.nature.com/scientificreports/

Figure 5. Speech feature attention rank.Here we apply a linear transformation to the attention score, taking the
smallest attention score as the baseline value and subtracting the baseline value

in model performance with the addition or removal of GCN layers, thereby confirming the effectiveness of
our proposed three-layer GCN structure. Lastly, we conducted an in-depth analysis of the model’s attention to
different EEG electrode channels and various acoustic features. The results are consistent with existing research,
further validating that EMO-GCN can reveal complex relationships within the data.
Our study still faces several limitations. On the data front, like many others, we resorted to segmenting data
to expand the dataset, primarily due to the current scarcity of depression-related data. Additionally, we aim to
integrate more types of modal data, allowing for a more comprehensive perspective in diagnosing depression.
In terms of clinical validation and acceptance, while our initial results show potential effectiveness, our research
has yet to undergo extensive clinical trials. Furthermore, the interpretability of the model is a key factor in its
adoption by medical professionals. Deep learning models are often criticized as “black box” models, which poses
a significant challenge in the medical field that demands high transparency and interpretability. In the future, we
hope to access more EEG and audio data from MDD patients, as well as data from other modalities, to explore
the diversity and variability across different datasets more deeply. Additionally, we aim to enhance the model’s
interpretability further, enabling medical professionals and patients to better understand the decision-making
process of the model.

Conclusion
In this paper, we introduce EMO-GCN, a multimodal depression detection framework based on graph neural
networks. We consider a graph convolutional layer and a graph pooling layer together as a fundamental unit. By
stacking such units threefold, we construct the core component of EMO-GCN, Multi-GCN. This configuration
enables the model to extract key features from different modalities and integrate these features through an
attention mechanism, achieving high-accuracy diagnosis of depression and exploring the potential connections
between EEG data and audio data.
Through a series of rigorous experimental validations, our method surpassed the performance of all currently
known methods in multimodal depression detection. The experimental analysis indicated that the GraphSAGE
component within the model, as well as the graph convolutional layers and graph pooling operations in the
graph neural network, played significant roles in enhancing model performance. Moreover, the model also
demonstrated high accuracy when processing unimodal data, and the integration of multimodal features further
increased detection precision. These results not only confirm the effectiveness of EMO-GCN in the field of
depression detection but also highlight the potential of multimodal data fusion in improving diagnostic accuracy.

Data availability
This study used the publicly available MODMA dataset, and these data can be obtained directly from the corre-
sponding official website https://modma.lzu.edu.cn/data/index/.

Received: 11 April 2024; Accepted: 13 November 2024

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 12


www.nature.com/scientificreports/

References
1. Kraus, C., Kadriu, B., Lanzenberger, R., Zarate, C. A. Jr. & Kasper, S. Prognosis and improved outcomes in major depression: A
review. Transl. Psychiatry 9, 127 (2019).
2. Otte, C. et al. Major depressive disorder. Nat. Rev. Dis. Primers 2, 1–20 (2016).
3. Strakowski, S. & Nelson, E. Major Depressive Disorder. (Oxford American Psychiatry Library/Oxford University Press, 2015).
4. Herrman, H. et al. Time for united action on depression: A lancet-world psychiatric association commission. Lancet 399, 957–
1022 (2022).
5. Santomauro, D. F. et al. Global prevalence and burden of depressive and anxiety disorders in 204 countries and territories in 2020
due to the covid-19 pandemic. Lancet 398, 1700–1712 (2021).
6. Marwaha, S. et al. Novel and emerging treatments for major depression. The Lancet 401, 141–153 (2023).
7. Jermy, B. S., Glanville, K. P., Coleman, J. R., Lewis, C. M. & Vassos, E. Exploring the genetic heterogeneity in major depression
across diagnostic criteria. Mol. Psychiatry 26, 7337–7345 (2021).
8. Berardi, M. et al. Relative importance of speech and voice features in the classification of schizophrenia and depression. Transl.
Psychiatry 13, 298 (2023).
9. Wang, J. et al. Acoustic differences between healthy and depressed people: A cross-situation study. BMC Psychiatry 19, 1–12 (2019).
10. Simmatis, L., Russo, E. E., Geraci, J., Harmsen, I. E. & Samuel, N. Technical and clinical considerations for electroencephalography-
based biomarkers for major depressive disorder. npj Mental Health Res. 2, 18 ( 2023).
11. Wu, C.-T. et al. Depression detection using relative eeg power induced by emotionally positive images and a conformal kernel
support vector machine. Appl. Sci. 8, 1244 (2018).
12. Hosseinifard, B., Moradi, M. H. & Rostami, R. Classifying depression patients and normal subjects using machine learning
techniques and nonlinear features from eeg signal. Comput. Methods Prog. Biomed. 109, 339–345 (2013).
13. Qayyum, A., Razzak, I., Tanveer, M., Mazher, M. & Alhaqbani, B. High-density electroencephalography and speech signal based
deep framework for clinical depression diagnosis. In IEEE/ACM Transactions on Computational Biology and Bioinformatics
(2023).
14. Zhao, Y. et al. Multi-head attention-based long short-term memory for depression detection from speech. Front. Neurorobot. 15,
684037 (2021).
15. Xie, Y. et al. Speech emotion classification using attention-based lstm. IEEE/ACM Trans. Audio Speech Lang. Process. 27, 1675–
1685 (2019).
16. Zhou, M. et al. Epileptic seizure detection based on eeg signals and cnn. Front. Neuroinform. 12, 95 (2018).
17. Molebatsi, K., Motlhatlhedi, K. & Wambua, G. N. The validity and reliability of the patient health questionnaire-9 for screening
depression in primary health care patients in Botswana. BMC Psychiatry 20, 1–10 (2020).
18. Seemüller, F. et al. A factor analytic comparison of three commonly used depression scales (hamd, madrs, bdi) in a large sample of
depressed inpatients. BMC Psychiatry 23, 548 (2023).
19. van Dijk, S. E. et al. Measurement properties of depression questionnaires in patients with diabetes: A systematic review. Qual. Life
Res. 27, 1415–1430 (2018).
20. Hobbs, C. et al. Comparison between self-administered depression questionnaires and patients’ own views of changes in their
mood: A prospective cohort study in primary care. Psychol. Med. 51, 853–860 (2021).
21. Deshpande, M. & Rao, V. Depression detection using emotion artificial intelligence. In 2017 International Conference on Intelligent
Sustainable Systems (ICISS). 858–862 ( IEEE, 2017).
22. Islam, M. R. et al. Depression detection from social network data using machine learning techniques. Health Inf. Sci. Syst. 6, 1–12
(2018).
23. Scarselli, F., Gori, M., Tsoi, A. C., Hagenbuchner, M. & Monfardini, G. The graph neural network model. IEEE Trans. Neural Netw.
20, 61–80 (2008).
24. Zhao, X., Liu, Z., Han, L. & Peng, S. Ecgnn: Enhancing abnormal recognition in 12-lead ecg with graph neural network. In 2022
IEEE International Conference on Bioinformatics and Biomedicine (BIBM). 1411–1416 ( IEEE, 2022).
25. Wang, Z., Gu, Y., Zheng, S., Yang, L. & Li, J. Mgrel: A multi-graph representation learning-based ensemble learning method for
gene-disease association prediction. Comput. Biol. Med. 155, 106642 (2023).
26. Fritz, C., Dorigatti, E. & Rügamer, D. Combining graph neural networks and spatio-temporal disease models to improve the
prediction of weekly covid-19 cases in Germany. Sci. Rep. 12, 3930 (2022).
27. Yu, Q. et al. Gnn-based depression recognition using spatio-temporal information: A FNIRS study. IEEE J. Biomed. Health Inform.
26, 4925–4935 (2022).
28. Sun, C., Jiang, M., Gao, L., Xin, Y. & Dong, Y. A novel study for depression detecting using audio signals based on graph neural
network. Biomed. Signal Process. Control 88, 105675 (2024).
29. Luo, G. et al. Exploring adaptive graph topologies and temporal graph networks for eeg-based depression detection. IEEE Trans.
Neural Syst. Rehabil. Eng. 31, 3947–3957 (2023).
30. Zheng, W., Yan, L. & Wang, F.-Y. Two birds with one stone: Knowledge-embedded temporal convolutional transformer for
depression detection and emotion recognition. IEEE Trans. Affect. Comput. 14, 2595–2613 (2023).
31. Chen, T., Hong, R., Guo, Y., Hao, S. & Hu, B. Ms 2 -gnn: Exploring gnn-based multimodal fusion network for depression detection.
IEEE Trans. Cybern. ( 2022).
32. Lin, Z. et al. Electroencephalography symmetry in power, waveform and power spectrum in major depression. In 2020 42nd
Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC). 5280–5283 ( IEEE, 2020).
33. Kipf, T. N. & Welling, M. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (
2016).
34. Ahmed, S., Yousuf, M. A., Monowar, M. M. & Hamid, M. A. & Alassafi, M. A Multimodal Depression Classification with Uncertainty
Approximation. Taking All the Factors We Need (IEEE Access, 2023).
35. Hu, Y. et al. Exploring large-scale language models to evaluate eeg-based multimodal data for mental health. In Companion of the
2024 on ACM International Joint Conference on Pervasive and Ubiquitous Computing. 412–417 ( 2024).
36. Tasci, G. et al. Automated accurate detection of depression using twin Pascal’s triangles lattice pattern with eeg signals. Knowl.-
Based Syst. 260, 110190 (2023).
37. Chen, T., Guo, Y., Hao, S. & Hong, R. Exploring self-attention graph pooling with eeg-based topological structure and soft label for
depression detection. IEEE Trans. Affect. Comput. 13, 2106–2118 (2022).
38. Soni, S., Seal, A., Yazidi, A. & Krejcar, O. Graphical representation learning-based approach for automatic classification of
electroencephalogram signals in depression. Comput. Biol. Med. 145, 105420 (2022).
39. Shen, J. et al. Depression recognition from eeg signals using an adaptive channel fusion method via improved focal loss. IEEE J.
Biomed. Health Inform. 27, 3234–3245 (2023).
40. Sun, S. et al. Eeg based depression recognition by combining functional brain network and traditional biomarkers. In 2020 IEEE
International Conference on Bioinformatics and Biomedicine (BIBM). 2074–2081 ( IEEE, 2020).
41. Chen, T., Guo, Y., Hao, S. & Hong, R. Semi-supervised domain adaptation for major depressive disorder detection. IEEE Trans.
Multimed. ( 2023).
42. Gheorghe, M., Mihalache, S. & Burileanu, D. Using deep neural networks for detecting depression from speech. In 2023 31st
European Signal Processing Conference (EUSIPCO). 411–415 ( IEEE, 2023).

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 13


www.nature.com/scientificreports/

43. Chen, X. & Pan, Z. A convenient and low-cost model of depression screening and early warning based on voice data using for
public mental health. Int. J. Environ. Res. Public Health 18, 6441 (2021).
44. Das, A. K. & Naskar, R. A deep learning model for depression detection based on mfcc and cnn generated spectrogram features.
Biomed. Signal Process. Control 90, 105898 (2024).
45. Cai, H. et al. Modma dataset: A multi-modal open dataset for mental-disorder analysis. arXiv preprint arXiv:2002.09283 ( 2020).
46. Li, X. et al. A resting-state brain functional network study in mdd based on minimum spanning tree analysis and the hierarchical
clustering. Complexity 2017 ( 2017).

Acknowledgements
This work was supported by NSFC-FDCT Grants 623611 66662; National Key R &D Program of Chi-
na 2023YFC3503 400, 2022YFC3400400; The Innovative Research Group Project of Hunan Province
2024JJ1002; Key R &D Program of Hunan Province 2023GK2004, 2023SK2059, 2023SK2060; Top 10 Tech-
nical Key Project in Hunan Province 2023GK1 010; Key Technologies R &D Program of Guangdong Prov-
ince (2023B1111030004 to FFH); Graduate Research Innovation Project of Hunan Province CX20240417; The
National Natural Science Foundation of China (No.62262011); The Guangxi key research and development
program(No.GuikeA A23062001); National Natural Science Foundation of China (82071543); STI2030-Major
Projects(2021ZD0200700); The Funds of State Key Laboratory of Chemo/Biose nsing and Chemometrics, the
National Supercomputing Center in Changsha (http://nscc.hnu.edu.cn/), and Peng Cheng Lab.

Declarations

Competing interests
The authors declare no competing interests.

Additional information
Correspondence and requests for materials should be addressed to X.X. or S.P.
Reprints and permissions information is available at www.nature.com/reprints.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and
institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which
permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give
appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and
indicate if changes were made. The images or other third party material in this article are included in the article’s
Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included
in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or
exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy
of this licence, visit http://creativecommons.org/licenses/by/4.0/.

© The Author(s) 2024

Scientific Reports | (2024) 14:28400 | https://doi.org/10.1038/s41598-024-79981-0 14

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy