Identifying relevant EEG channels for subject-independent emotion recognition using attention network layers

Electrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built using two approaches: subject-dependent and subject-independent. Although subject-independent models offer greater practical utility compared...

Full description

Saved in:
Bibliographic Details
Published inFrontiers in psychiatry Vol. 16; p. 1494369
Main Authors Valderrama, Camilo E, Sheoran, Anshul
Format Journal Article
LanguageEnglish
Published Switzerland Frontiers Media S.A 10.02.2025
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Electrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built using two approaches: subject-dependent and subject-independent. Although subject-independent models offer greater practical utility compared to subject-dependent models, they face challenges due to the significant variability of EEG signals between individuals. One potential solution to enhance subject-independent approaches is to identify EEG channels that are consistently relevant across different individuals for predicting emotion. With the growing use of deep learning in emotion recognition, incorporating attention mechanisms can help uncover these shared predictive patterns. This study explores this method by applying attention mechanism layers to identify EEG channels that are relevant for predicting emotions in three independent datasets (SEED, SEED-IV, and SEED-V). The model achieved average accuracies of 79.3% (CI: 76.0-82.5%), 69.5% (95% CI: 64.2-74.8%) and 60.7% (95% CI: 52.3-69.2%) on these datasets, revealing that EEG channels located along the head circumference, including , , , , , , , , , and , are the most crucial for emotion prediction. These results emphasize the importance of capturing relevant electrical activity from these EEG channels, thereby facilitating the prediction of emotions evoked by audiovisual stimuli in subject-independent approaches.
AbstractList Electrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built using two approaches: subject-dependent and subject-independent. Although subject-independent models offer greater practical utility compared to subject-dependent models, they face challenges due to the significant variability of EEG signals between individuals. One potential solution to enhance subject-independent approaches is to identify EEG channels that are consistently relevant across different individuals for predicting emotion. With the growing use of deep learning in emotion recognition, incorporating attention mechanisms can help uncover these shared predictive patterns. This study explores this method by applying attention mechanism layers to identify EEG channels that are relevant for predicting emotions in three independent datasets (SEED, SEED-IV, and SEED-V). The model achieved average accuracies of 79.3% (CI: 76.0-82.5%), 69.5% (95% CI: 64.2-74.8%) and 60.7% (95% CI: 52.3-69.2%) on these datasets, revealing that EEG channels located along the head circumference, including , , , , , , , , , and , are the most crucial for emotion prediction. These results emphasize the importance of capturing relevant electrical activity from these EEG channels, thereby facilitating the prediction of emotions evoked by audiovisual stimuli in subject-independent approaches.
BackgroundElectrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built using two approaches: subject-dependent and subject-independent. Although subject-independent models offer greater practical utility compared to subject-dependent models, they face challenges due to the significant variability of EEG signals between individuals. ObjectiveOne potential solution to enhance subject-independent approaches is to identify EEG channels that are consistently relevant across different individuals for predicting emotion. With the growing use of deep learning in emotion recognition, incorporating attention mechanisms can help uncover these shared predictive patterns.MethodsThis study explores this method by applying attention mechanism layers to identify EEG channels that are relevant for predicting emotions in three independent datasets (SEED, SEED-IV, and SEED-V). ResultsThe model achieved average accuracies of 79.3% (CI: 76.0-82.5%), 69.5% (95% CI: 64.2-74.8%) and 60.7% (95% CI: 52.3-69.2%) on these datasets, revealing that EEG channels located along the head circumference, including Fp1, Fp2, F7, F8, T7, T8, P7, P8, O1, and O2, are the most crucial for emotion prediction. ConclusionThese results emphasize the importance of capturing relevant electrical activity from these EEG channels, thereby facilitating the prediction of emotions evoked by audiovisual stimuli in subject-independent approaches.
Electrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built using two approaches: subject-dependent and subject-independent. Although subject-independent models offer greater practical utility compared to subject-dependent models, they face challenges due to the significant variability of EEG signals between individuals.BackgroundElectrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built using two approaches: subject-dependent and subject-independent. Although subject-independent models offer greater practical utility compared to subject-dependent models, they face challenges due to the significant variability of EEG signals between individuals.One potential solution to enhance subject-independent approaches is to identify EEG channels that are consistently relevant across different individuals for predicting emotion. With the growing use of deep learning in emotion recognition, incorporating attention mechanisms can help uncover these shared predictive patterns.ObjectiveOne potential solution to enhance subject-independent approaches is to identify EEG channels that are consistently relevant across different individuals for predicting emotion. With the growing use of deep learning in emotion recognition, incorporating attention mechanisms can help uncover these shared predictive patterns.This study explores this method by applying attention mechanism layers to identify EEG channels that are relevant for predicting emotions in three independent datasets (SEED, SEED-IV, and SEED-V).MethodsThis study explores this method by applying attention mechanism layers to identify EEG channels that are relevant for predicting emotions in three independent datasets (SEED, SEED-IV, and SEED-V).The model achieved average accuracies of 79.3% (CI: 76.0-82.5%), 69.5% (95% CI: 64.2-74.8%) and 60.7% (95% CI: 52.3-69.2%) on these datasets, revealing that EEG channels located along the head circumference, including Fp 1, Fp 2, F 7, F 8, T 7, T 8, P 7, P 8, O 1, and O 2, are the most crucial for emotion prediction.ResultsThe model achieved average accuracies of 79.3% (CI: 76.0-82.5%), 69.5% (95% CI: 64.2-74.8%) and 60.7% (95% CI: 52.3-69.2%) on these datasets, revealing that EEG channels located along the head circumference, including Fp 1, Fp 2, F 7, F 8, T 7, T 8, P 7, P 8, O 1, and O 2, are the most crucial for emotion prediction.These results emphasize the importance of capturing relevant electrical activity from these EEG channels, thereby facilitating the prediction of emotions evoked by audiovisual stimuli in subject-independent approaches.ConclusionThese results emphasize the importance of capturing relevant electrical activity from these EEG channels, thereby facilitating the prediction of emotions evoked by audiovisual stimuli in subject-independent approaches.
Author Valderrama, Camilo E
Sheoran, Anshul
AuthorAffiliation 1 Department of Applied Computer Science, University of Winnipeg , Winnipeg, MB , Canada
2 Department of Community Health Sciences, Cumming School of Medicine, University of Calgary , Calgary, AB , Canada
AuthorAffiliation_xml – name: 2 Department of Community Health Sciences, Cumming School of Medicine, University of Calgary , Calgary, AB , Canada
– name: 1 Department of Applied Computer Science, University of Winnipeg , Winnipeg, MB , Canada
Author_xml – sequence: 1
  givenname: Camilo E
  surname: Valderrama
  fullname: Valderrama, Camilo E
  organization: Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
– sequence: 2
  givenname: Anshul
  surname: Sheoran
  fullname: Sheoran, Anshul
  organization: Department of Applied Computer Science, University of Winnipeg, Winnipeg, MB, Canada
BackLink https://www.ncbi.nlm.nih.gov/pubmed/39995952$$D View this record in MEDLINE/PubMed
BookMark eNpVkU9vFCEYxompsbX2C3gwc_QyKwwMM5yMada6SRMveib8edmysrACU7PfXratpuUADy-8v4cnvEVnMUVA6D3BK0pn8ckdyrGuBjyMK8IEo1y8QheEc9ZjzvDZM32OrkrZ4TaoEJSPb9B5E2IU43CBwsZCrN4dfdx2GQLcq1i79fqmM3cqRgilcyl3ZdE7MLX30cIB4qmng32qPsXWZdI2-ge9lBNH1XqCtn2E-iflX11QR8jlHXrtVChw9bReop9f1z-uv_W33282119ue0vxVHtHLdcAVivRJq4dn53AbtZaDGAIITC3Ips0p8TO9hSUtWxUOKOYgZFeos0j1ya1k4fs9yofZVJePhRS3kqVqzcBpMKMWWWxnsaR8XFqFlab5sPMZDERjfX5kXVY9B6sacGyCi-gL0-iv5PbdC8Jmdk0D7QRPj4Rcvq9QKly74uBEFSEtBRJyYQFZe397eqH52b_Xf79F_0LTE6fxg
ContentType Journal Article
Copyright Copyright © 2025 Valderrama and Sheoran.
Copyright © 2025 Valderrama and Sheoran 2025 Valderrama and Sheoran
Copyright_xml – notice: Copyright © 2025 Valderrama and Sheoran.
– notice: Copyright © 2025 Valderrama and Sheoran 2025 Valderrama and Sheoran
DBID NPM
7X8
5PM
DOA
DOI 10.3389/fpsyt.2025.1494369
DatabaseName PubMed
MEDLINE - Academic
PubMed Central (Full Participant titles)
DOAJ Directory of Open Access Journals
DatabaseTitle PubMed
MEDLINE - Academic
DatabaseTitleList PubMed

MEDLINE - Academic
Database_xml – sequence: 1
  dbid: DOA
  name: DOAJ Directory of Open Access Journals
  url: https://www.doaj.org/
  sourceTypes: Open Website
– sequence: 2
  dbid: NPM
  name: PubMed
  url: https://proxy.k.utb.cz/login?url=http://www.ncbi.nlm.nih.gov/entrez/query.fcgi?db=PubMed
  sourceTypes: Index Database
DeliveryMethod fulltext_linktorsrc
EISSN 1664-0640
ExternalDocumentID oai_doaj_org_article_a044dad0b7554657b92dbc92e4c7d019
PMC11847823
39995952
Genre Journal Article
GroupedDBID 53G
5VS
9T4
AAFWJ
AAKDD
ABIVO
ACGFO
ACGFS
ACXDI
ADBBV
ADRAZ
AFPKN
ALMA_UNASSIGNED_HOLDINGS
AOIJS
BAWUL
BCNDV
DIK
EMOBN
GROUPED_DOAJ
GX1
HYE
IPNFZ
KQ8
M48
M~E
NPM
O5R
O5S
OK1
PGMZT
RIG
RNS
RPM
7X8
5PM
ID FETCH-LOGICAL-d307t-f3d6beedba9edb6bf68f90f8bb92ec111e86bf47b631d8d1664403939fca4ce53
IEDL.DBID DOA
ISSN 1664-0640
IngestDate Wed Aug 27 01:31:51 EDT 2025
Thu Aug 21 18:27:50 EDT 2025
Fri Jul 11 03:18:32 EDT 2025
Thu Apr 03 07:06:16 EDT 2025
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed true
IsScholarly true
Keywords deep learning
affective computing
attention mechanism
emotion recognition
electroencephalogram
EEG signal processing
Language English
License Copyright © 2025 Valderrama and Sheoran.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-d307t-f3d6beedba9edb6bf68f90f8bb92ec111e86bf47b631d8d1664403939fca4ce53
Notes ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
Reviewed by: Jiahui Pan, South China Normal University, China
Edited by: Panagiotis Tzirakis, Hume AI, United States
Konstantinos Barmpas, Imperial College London, United Kingdom
OpenAccessLink https://doaj.org/article/a044dad0b7554657b92dbc92e4c7d019
PMID 39995952
PQID 3170934939
PQPubID 23479
ParticipantIDs doaj_primary_oai_doaj_org_article_a044dad0b7554657b92dbc92e4c7d019
pubmedcentral_primary_oai_pubmedcentral_nih_gov_11847823
proquest_miscellaneous_3170934939
pubmed_primary_39995952
PublicationCentury 2000
PublicationDate 2025-02-10
PublicationDateYYYYMMDD 2025-02-10
PublicationDate_xml – month: 02
  year: 2025
  text: 2025-02-10
  day: 10
PublicationDecade 2020
PublicationPlace Switzerland
PublicationPlace_xml – name: Switzerland
PublicationTitle Frontiers in psychiatry
PublicationTitleAlternate Front Psychiatry
PublicationYear 2025
Publisher Frontiers Media S.A
Publisher_xml – name: Frontiers Media S.A
SSID ssj0000399365
Score 2.360814
Snippet Electrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can be built...
BackgroundElectrical activity recorded with electroencephalography (EEG) enables the development of predictive models for emotion recognition. These models can...
SourceID doaj
pubmedcentral
proquest
pubmed
SourceType Open Website
Open Access Repository
Aggregation Database
Index Database
StartPage 1494369
SubjectTerms affective computing
attention mechanism
deep learning
EEG signal processing
electroencephalogram
emotion recognition
Psychiatry
Title Identifying relevant EEG channels for subject-independent emotion recognition using attention network layers
URI https://www.ncbi.nlm.nih.gov/pubmed/39995952
https://www.proquest.com/docview/3170934939
https://pubmed.ncbi.nlm.nih.gov/PMC11847823
https://doaj.org/article/a044dad0b7554657b92dbc92e4c7d019
Volume 16
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwrV1LS8NAEF7EkxdRfMUXK3hdmmQfyR5VWougJwu9hd3sRgslLTY9-O-d2a2lFcGLlxCSww4zuzPfJN_MEHKb514KYzkD95gyIXnBrCo0c1JyiHeyNKGK__lFDUfiaSzHG6O-kBMW2wNHxfVMKoQzLrUF8qlkYXXubK1zL-rCpaHhZw4xbyOZCj4Y466SsUoGsjDda-aLT-RO5hKcgxY8MJxxtd-g5U-G5EbIGRyQ_RVWpHdRxkOy49sjMo2ltaE8ieLEE4DCHe33HynW8LYQ6ijgULpYWvzCwibrMbcd9XFkD12ThuAeee9vFJtsBtojbSMtnE4NYvFjMhr0Xx-GbDUygTk4rB1ruFMWwp41Gi7KNqpsdNqUFtTma_BrvoSHorCKZ650mQI4hNW5uqmNqL3kJ2S3nbX-jFCjVZF5l1k40cJlzmCmJyB70xyMa8uE3KP6qnnsilFhn-rwAKxXraxX_WW9hNx8K7-CfY0_K0zrZ8tFBbgmhZVAtIScRmOslwLjaqllnpByy0xbsmy_aSfvoXc25FMCQBE__w_pL8gebigkcWfpJdntPpb-CjBKZ6_DdvwCmzTpGA
linkProvider Directory of Open Access Journals
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Identifying+relevant+EEG+channels+for+subject-independent+emotion+recognition+using+attention+network+layers&rft.jtitle=Frontiers+in+psychiatry&rft.au=Valderrama%2C+Camilo+E&rft.au=Sheoran%2C+Anshul&rft.date=2025-02-10&rft.issn=1664-0640&rft.eissn=1664-0640&rft.volume=16&rft.spage=1494369&rft_id=info:doi/10.3389%2Ffpsyt.2025.1494369&rft_id=info%3Apmid%2F39995952&rft.externalDocID=39995952
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/lc.gif&issn=1664-0640&client=summon
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/mc.gif&issn=1664-0640&client=summon
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=/sc.gif&issn=1664-0640&client=summon