好色先生

好色先生

Explore the latest content from across our publications

Log In

Forgot Password?
Create New Account

Loading... please wait

Abstract Details

EEG-VL: Integrating Visual Features with Large Language Models for Automated Seizure Detection
Epilepsy/Clinical Neurophysiology (EEG)
P2 - Poster Session 2 (11:45 AM-12:45 PM)
11-011

To address the increasing demand for accurate EEG-based epileptic seizure detection, we propose EEG-VL, a novel vision–language framework that treats EEG signals as visual patterns and integrates them with large language models to improve seizure detection.

Traditional approaches often struggle to capture the complex spatiotemporal patterns inherent in EEG signals and typically lack high-level contextual understanding, limiting their applicability in real-world clinical settings. Recent advances in large language models (LLMs) have enabled new opportunities for semantic reasoning in biomedical tasks, offering a promising strategy for more sophisticated and semantically informed methodologies. 

A pretrained EfficientNet encoder is used to extract abstract visual features from EEG representations, which are embedded into structured prompts and processed by the Qwen language model. This design synergistically combines the spatial modeling capabilities of convolutional networks with the semantic reasoning strengths of large language models. We conduct experiments on two widely used benchmark datasets for seizure detection, the TUH EEG Seizure Corpus (TUSZ) and the CHB-MIT Scalp EEG Database. To address the class imbalance commonly present in seizure datasets, we adopt a logit adjustment strategy based on label distribution priors.

We employed several classic EEG networks for performance comparisons, including EEGNet, EEG-TCNet, ShallowCNN-Tran, EEGNet-Tran and EEG2ViT. Extensive experiments on the TUSZ and CHB-MIT datasets demonstrate that EEG-VL achieves state-of-the-art performance. On TUSZ, our model attains an AUROC of 0.9466 and an AUPRC of 0.7599, surpassing previous best results by 0.80% and 8.19%, respectively. On CHB-MIT, EEG-VL achieves average AUROC of 0.91 and AUPRC of 0.47, superior previous best performance by 1% and 14.6%.

Our study underscores the potential of the proposed vision–language paradigm for robust, scalable, and clinically applicable EEG-based seizure detection. By integrating a pretrained visual encoder with a large language model, EEG-VL exhibits excellent performance on multiple public datasets, demonstrating promising potential for clinical adoption.

Authors/Disclosures
Nan Lin, MD
PRESENTER
Dr. Lin has nothing to disclose.
Zi Liang, Jr., MD, no Mr. Liang has nothing to disclose.
Ming Lu Ming Lu has received personal compensation for serving as an employee of Eli Lilly and Company. Ming Lu has stock in Eli Lilly and Company.
Weifang Gao Dr. Gao has nothing to disclose.
Heyagn Sun Dr. Sun has nothing to disclose.
Peng Hu Mr. Hu has nothing to disclose.
Thomas Li, PhD Dr. Li has nothing to disclose.