Cross-subject affective analysis based on dynamic brain functional networks

Emotion recognition is crucial in facilitating human-computer emotional interaction. To enhance the credibility and realism of emotion recognition, researchers have turned to physiological signals, particularly EEG signals, as they directly reflect cerebral cortex activity. However, due to inter-sub...

Full description

Saved in:
Bibliographic Details
Published inFrontiers in human neuroscience Vol. 19; p. 1445763
Main Authors You, Lifeng, Zhong, Tianyu, He, Erheng, Liu, Xuejie, Zhong, Qinghua
Format Journal Article
LanguageEnglish
Published Switzerland Frontiers Media S.A 14.04.2025
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Emotion recognition is crucial in facilitating human-computer emotional interaction. To enhance the credibility and realism of emotion recognition, researchers have turned to physiological signals, particularly EEG signals, as they directly reflect cerebral cortex activity. However, due to inter-subject variability and non-smoothness of EEG signals, the generalization performance of models across subjects remains a challenge. In this study, we proposed a novel approach that combines time-frequency analysis and brain functional networks to construct dynamic brain functional networks using sliding time windows. This integration of time, frequency, and spatial domains helps to effectively capture features, reducing inter-individual differences, and improving model generalization performance. To construct brain functional networks, we employed mutual information to quantify the correlation between EEG channels and set appropriate thresholds. We then extracted three network attribute features-global efficiency, local efficiency, and local clustering coefficients-to achieve emotion classification based on dynamic brain network features. The proposed method is evaluated on the DEAP dataset through subject-dependent (trial-independent), subject-independent, and subject- and trial-independent experiments along both valence and arousal dimensions. The results demonstrate that our dynamic brain functional network outperforms the static brain functional network in all three experimental cases. High classification accuracies of 90.89% and 91.17% in the valence and arousal dimensions, respectively, were achieved on the subject-independent experiments based on the dynamic brain function, leading to significant advancements in EEG-based emotion recognition. In addition, experiments with each brain region yielded that the left and right temporal lobes focused on processing individual private emotional information, whereas the remaining brain regions paid attention to processing basic emotional information.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
Reviewed by: Aurora Saibene, University of Milano-Bicocca, Italy
Edited by: Noman Naseer, Air University, Pakistan
Sevgi Şengül Ayan, Antalya Bilim University, Türkiye
Minmin Miao, Huzhou University, China
ISSN:1662-5161
1662-5161
DOI:10.3389/fnhum.2025.1445763