ALFEE: Adaptive Large Foundation Model for EEG Representation
While foundation models excel in text, image, and video domains, the critical biological signals, particularly electroencephalography(EEG), remain underexplored. EEG benefits neurological research with its high temporal resolution, operational practicality, and safety profile. However, low signal-to...
Saved in:
Main Authors | , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
07.05.2025
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | While foundation models excel in text, image, and video domains, the critical
biological signals, particularly electroencephalography(EEG), remain
underexplored. EEG benefits neurological research with its high temporal
resolution, operational practicality, and safety profile. However, low
signal-to-noise ratio, inter-subject variability, and cross-paradigm
differences hinder the generalization of current models. Existing methods often
employ simplified strategies, such as a single loss function or a
channel-temporal joint representation module, and suffer from a domain gap
between pretraining and evaluation tasks that compromises efficiency and
adaptability. To address these limitations, we propose the Adaptive Large
Foundation model for EEG signal representation(ALFEE) framework, a novel hybrid
transformer architecture with two learning stages for robust EEG representation
learning. ALFEE employs a hybrid attention that separates channel-wise feature
aggregation from temporal dynamics modeling, enabling robust EEG representation
with variable channel configurations. A channel encoder adaptively compresses
variable channel information, a temporal encoder captures task-guided
evolution, and a hybrid decoder reconstructs signals in both temporal and
frequency domains. During pretraining, ALFEE optimizes task prediction, channel
and temporal mask reconstruction, and temporal forecasting to enhance
multi-scale and multi-channel representation. During fine-tuning, a full-model
adaptation with a task-specific token dictionary and a cross-attention layer
boosts performance across multiple tasks. After 25,000 hours of pretraining,
extensive experimental results on six downstream EEG tasks demonstrate the
superior performance of ALFEE over existing models. Our ALFEE framework
establishes a scalable foundation for biological signal analysis with
implementation at https://github.com/xw1216/ALFEE. |
---|---|
DOI: | 10.48550/arxiv.2505.06291 |