TY - JOUR
T1 - Guiding LLMs to decode text via aligning semantics in EEG signals and language
AU - Zheng, Huanran
AU - Wu, Yuanbin
AU - Qian, Tianwen
AU - Yue, Wenjing
AU - Wang, Xiaoling
N1 - Publisher Copyright:
© 2025 Elsevier Ltd. All rights are reserved, including those for text and data mining, AI training, and similar technologies.
PY - 2026/3/1
Y1 - 2026/3/1
N2 - With the rapid development of brain-computer interfaces (BCI) in recent years, the electroencephalography (EEG) to text task has drawn increasing attention. This task aims to generate natural text based on EEG signals to assist individuals who have lost their communication ability. Previous methods have defined it as a sequence-to-sequence translation task. However, their model was trained using a teacher-forcing strategy, which introduced language bias and could not effectively utilize EEG signals. To address this issue, we propose a novel framework in this paper, which innovatively treats the EEG-to-text task as a fine-grained controllable text generation task. Specifically, since large language models (LLMs) have strong text generation capabilities, we guide LLMs in generating the desired sentences step by step by re-ranking the predicted candidate words based on their semantic similarity with the EEG segment representations. Therefore, our approach focuses on training a word-level EEG representation model to effectively extract information from EEG signals and align EEG representations with word semantics without using teacher-forcing strategies. Extensive experiments on the ZuCo benchmark demonstrate the effectiveness of our approach, which achieves state-of-the-art performance in both multi-subject and single-subject settings. Furthermore, experimental results in cross-subject scenarios further verify that our method has a strong generalization ability and can be applied to unseen subjects.
AB - With the rapid development of brain-computer interfaces (BCI) in recent years, the electroencephalography (EEG) to text task has drawn increasing attention. This task aims to generate natural text based on EEG signals to assist individuals who have lost their communication ability. Previous methods have defined it as a sequence-to-sequence translation task. However, their model was trained using a teacher-forcing strategy, which introduced language bias and could not effectively utilize EEG signals. To address this issue, we propose a novel framework in this paper, which innovatively treats the EEG-to-text task as a fine-grained controllable text generation task. Specifically, since large language models (LLMs) have strong text generation capabilities, we guide LLMs in generating the desired sentences step by step by re-ranking the predicted candidate words based on their semantic similarity with the EEG segment representations. Therefore, our approach focuses on training a word-level EEG representation model to effectively extract information from EEG signals and align EEG representations with word semantics without using teacher-forcing strategies. Extensive experiments on the ZuCo benchmark demonstrate the effectiveness of our approach, which achieves state-of-the-art performance in both multi-subject and single-subject settings. Furthermore, experimental results in cross-subject scenarios further verify that our method has a strong generalization ability and can be applied to unseen subjects.
KW - Brain-computer interface
KW - Brain-to-Text
KW - Contrastive learning
KW - Electroencephalography
KW - Large language models
UR - https://www.scopus.com/pages/publications/105023823572
U2 - 10.1016/j.eswa.2025.130300
DO - 10.1016/j.eswa.2025.130300
M3 - 文章
AN - SCOPUS:105023823572
SN - 0957-4174
VL - 299
JO - Expert Systems with Applications
JF - Expert Systems with Applications
ER -