Ocular artifacts remain a significant challenge in electroencephalography (EEG), potentially compromising data integrity in both basic neuroscience and clinical drug development.
Ocular artifacts remain a significant challenge in electroencephalography (EEG), potentially compromising data integrity in both basic neuroscience and clinical drug development. This article provides a comprehensive overview of EEG pre-processing techniques specifically for ocular artifact removal, tailored for researchers and biomedical professionals. We explore the foundational principles of ocular artifacts and their impact on neural signals, systematically compare the efficacy of established and emerging methodologies—including regression, Independent Component Analysis (ICA), and wavelet-based techniques—and provide evidence-based strategies for pipeline optimization and troubleshooting. A critical validation framework is presented to guide the selection of methods based on specific research goals, emphasizing the balance between signal preservation and artifact removal to ensure the reliability of neural data for clinical applications.
Ocular artifacts in electroencephalographic (EEG) recordings originate from two primary physiological phenomena: the corneo-retinal dipole (CRD) and the mechanical action of the eyelids. A comprehensive understanding of this bioelectrical source is fundamental to developing effective artifact removal strategies.
The eye functions as a steady electrical dipole with a positive pole at the cornea and a negative pole at the retina, generating a standing potential known as the corneo-fundal or corneo-retinal potential [1] [2]. This potential difference is substantial, with the cornea exhibiting a steady DC potential of approximately +13 mV relative to the forehead [1]. This electrical field is not static; it changes orientation with eyeball rotation and is altered by the movement of the eyelids, producing measurable potential changes at the scalp surface that contaminate the EEG signal [3].
Two distinct types of artifacts are generated by this system. Blink artifacts are caused by the eyelid sliding over the positively charged cornea during a blink, which inverts the polarity and creates a positive current that spreads across the scalp [2]. Eye movement artifacts, such as those from saccades, result from changes in the orientation of the entire corneo-retinal dipole, producing box-shaped deflections with opposite polarity on opposite sides of the head [1] [2]. The scalp distribution of these artifacts is not uniform and can be described using propagation factors—the fraction of the electrooculographic (EOG) signal recorded at periocular electrodes that appears at a particular scalp location [1]. These factors vary significantly with electrode location and differ between blinks and vertical eye movements [1].
The table below summarizes the key characteristics of the primary ocular artifacts, essential for their identification and removal.
Table 1: Quantitative and Topographical Characteristics of Primary Ocular Artifacts
| Artifact Type | Physiological Origin | Typical Amplitude | Spectral Profile | Maximum Topographic Expression |
|---|---|---|---|---|
| Eye Blink | Eyelid sliding over cornea [2] | >100 µV [2] | Delta & Theta bands (< 4-8 Hz) [2] | Frontal channels [1] [2] |
| Vertical Eye Movement | Change in CRD orientation [1] | Varies with movement amplitude | Delta & Theta bands [2] | Frontal channels; different propagation factor than blinks [1] |
| Lateral Eye Movement (Saccade) | Change in CRD orientation [2] | Varies with movement amplitude | Delta & Theta bands, with effects up to 20 Hz [2] | Channels near temples (e.g., F7/F8) [2] |
Recent research demonstrates that the CRD and eyelid-related artifacts exhibit a high degree of stationarity, making them particularly amenable to correction algorithms. Studies confirm these artifacts can be considered stationary for at least 1 to 1.5 hours, validating the use of calibration-based correction methods for both offline and online applications [3].
The acquisition of high-quality calibration data is a critical first step for many artifact correction methods, including the SGEYESUB algorithm [3].
The Sparse Generalized Eye Artifact Subspace Subtraction (SGEYESUB) algorithm is a state-of-the-art method that offers an optimal trade-off between correcting eye artifacts and preserving brain activity [3].
For research employing machine learning decoding, a specific validation protocol is recommended to assess the impact of artifact handling on performance [4].
Table 2: Key Materials and Tools for Ocular Artifact Research
| Item | Specification / Example | Primary Function in Research |
|---|---|---|
| High-Density EEG System | 64+ channels; active electrode systems (e.g., actiCAP) | Records scalp potentials with high spatial resolution; active systems reduce cable movement artifacts [2]. |
| Electrooculography (EOG) Electrodes | Standard Ag/AgCl electrodes | Placed periocularly to record reference signals for vertical and horizontal eye movements/blinks [2]. |
| Artifact Correction Software | SGEYESUB, ICA implementations (e.g., in EEGLAB) | Implements algorithms to separate and subtract ocular artifacts from the EEG signal [3] [2]. |
| Computational Environment | MATLAB, Python | Provides the framework for implementing custom analysis scripts, signal processing, and MVPA decoding [4]. |
| Calibration Paradigm Software | Presentation, PsychoPy | Displays visual cues to guide participants through standardized sequences of eye movements and blinks for calibration [3]. |
Electroencephalography (EEG) provides a non-invasive window into brain function, yet the signals it captures are perpetually contaminated by non-neural sources. Among these, artifacts generated by eye blinks and movements represent a significant challenge for interpretation. These ocular artifacts not only obscure underlying neural activity but, more critically, can masquerade as genuine neurophysiological phenomena. This misattribution can lead to false conclusions in fundamental neuroscience research, clinical assessment, and pharmaceutical development. Understanding the specific spectral and temporal characteristics of these artifacts is therefore a prerequisite for any rigorous EEG analysis pipeline. This application note details how ocular artifacts mimic brain activity and provides validated protocols for their identification and removal.
Ocular artifacts primarily originate from the corneo-retinal dipole. Eyeball rotation during blinks and saccades creates a large electric field that propagates across the scalp, dominating the EEG signal, particularly in frontal regions [5]. The central problem is that the characteristics of these artifacts often overlap with those of neurogenic signals in both the time and frequency domains, leading to potential misclassification.
Table 1: Spectral and Temporal Characteristics of Ocular Artifacts vs. Neural Signals
| Feature | Ocular Artifacts (Eye Blinks/Movements) | Genuine Neural Signals | Key Differentiators and Risks |
|---|---|---|---|
| Spectral Profile | Broad-spectral, high-amplitude spikes; predominant in very low frequencies (<4 Hz) but power can extend into beta/gamma bands [6] [7]. | Typically oscillatory activity confined to classic bands (e.g., Alpha: 8-13 Hz, Beta: 13-30 Hz). | Broad-spectral power of artifacts can be misinterpreted as broad-band power suppression or induced gamma activity [6]. |
| Temporal Profile | Sharp, spike-like onset with large amplitude (often >100 μV) [6] [7]. Stereotypical, monophasic/biphasic shape for blinks. | Smoother, lower-amplitude waveforms (e.g., <20 μV for ERPs). | High-amplitude spikes can be misidentified as epileptiform activity or large evoked potentials [7]. |
| Spatial Topography | Maximal amplitude over frontal and prefrontal electrodes; volume conduction creates a wide, bi-frontal positive wave for blinks [5] [8]. | Topography varies with neural generator location (e.g., sensorimotor rhythms maximal over central sites). | Frontal cortical activity (e.g., LAN ERP component) can be confused with ocular artifact topography [8]. |
| Response to Manipulation | Can be elicited by sound as a vestigial startle reflex (Postauricular Muscle, PAM); can be experimentally manipulated via eye gaze direction [7]. | Modulated by cognitive tasks, sensory stimuli, or pharmacological agents. | Myogenic reflexes can be misattributed to experience-dependent neuroplasticity (e.g., in musicians) [7]. |
The consequences of this masquerade are particularly acute in specific research contexts. For instance, in studies on neuroplasticity, enhancements in the frequency-following response (FFR) in musicians have been observed. However, recent evidence suggests that upwards of ~50% of the FFR can originate from myogenic contamination of the postauricular muscle (PAM), a vestigial startle reflex, rather than from auditory-neurogenic structures [7]. Similarly, in pharmaco-EEG trials, the choice of ocular artifact correction method can significantly alter the estimated pharmacokinetic-pharmacodynamic (PK-PD) relationships of a drug, directly impacting conclusions about its central activity and bioavailability [5].
Objective: To determine the proportion of the FFR that originates from postauricular muscle (PAM) artifact rather than neural sources [7].
Materials:
Procedure:
Objective: To assess the impact of different ocular artifact removal methods on the assessment of drug-induced EEG changes and PK-PD modeling [5].
Materials:
Procedure:
Objective: To replace error-prone manual selection of blink-related Independent Component Analysis (ICA) components with a validated, automated approach [8].
Materials:
icablinkmetrics, ADJUST, and EyeCatchProcedure:
icablinkmetrics: Relies on time-series features, correlating component activity with blink artifacts in the raw EEG.ADJUST: Uses combined stereotypical spatial and temporal features (e.g., temporal kurtosis, spatial average difference).EyeCatch: Relies on spatial features by correlating component scalp maps with a database of eye-activity templates.icablinkmetrics reduces false positives) can be integrated into the standard preprocessing pipeline for objective and reproducible blink removal.The selection of an artifact handling strategy should be guided by the research question. For studies where eye movements or head motions may be of interest (e.g., cybersickness research), minimal preprocessing may be beneficial to retain these potentially informative signals [9]. Conversely, for investigating precise neural oscillatory dynamics or drug effects, aggressive artifact removal using advanced methods like BSS/ICA is necessary.
The following workflow diagram outlines a decision process for handling artifacts that can masquerade as neural activity:
The Scientist's Toolkit table below details key resources for implementing the described protocols.
Table 2: Research Reagent Solutions for Ocular Artifact Research
| Tool Name | Type | Primary Function in Research | Key Application Note |
|---|---|---|---|
| EEGLAB | Software Toolbox | Provides an extensible environment for EEG processing, including ICA decomposition and visualization. | Core platform for running ADJUST, EyeCatch, and icablinkmetrics plugins for automated component selection [8]. |
| ADJUST Plugin | Algorithm/Plugin | Automates ICA component selection for artifacts using spatio-temporal features. | Effectively identifies blink, eye movement, and cardiac artifacts but can be prone to false positives with frontal brain activity [8]. |
| icablinkmetrics() | Algorithm/Plugin | Automates blink component selection based on time-series correlation with artifact. | Reduces false positives; effective in preserving frontally-maximal neural components while removing blink artifacts [8]. |
| Blind Source Separation (BSS) | Signal Processing Method | Separates mixed signals into statistically independent sources (components). | Superior to regression for ocular correction in pharmaco-EEG as it preserves cerebral activity in EOG signals [5]. |
| LORETA | Software Tool | Performs tomographic source localization of EEG signals. | Used to validate artifact removal; BSS preprocessing leads to more neurophysiologically plausible source maps [5]. |
| Postauricular Muscle (PAM) Recording | Electrophysiological Method | Directly records myogenic activity from the vestigial muscle behind the ear. | Critical control for FFR/ABR studies to quantify and correct for myogenic contamination masquerading as neural plasticity [7]. |
Ocular and myogenic artifacts present a formidable challenge in EEG research due to their capacity to imitate genuine neurogenic activity in spectral, temporal, and spatial dimensions. Failure to adequately address these artifacts jeopardizes the validity of findings across diverse fields, from cognitive neuroscience to clinical drug development. By applying the quantitative characteristics, experimental protocols, and methodological recommendations outlined in this application note, researchers can enhance the rigor and reliability of their EEG analyses, ensuring that the signals they interpret truly reflect the workings of the brain.
Electroencephalography (EEG) pre-processing, particularly ocular artifact removal, is a critical foundational step that profoundly influences all subsequent neural signal analysis. The choices made during this pre-processing phase have a high-stakes impact on the validity, reliability, and interpretability of results in event-related potential (ERP) analysis, functional connectivity mapping, and source localization. Research demonstrates that conventional artifact removal approaches can inadvertently introduce significant biases, artificially inflate effect sizes, and distort the spatial localization of neural activity [10] [11] [12]. This Application Note examines the consequences of pre-processing decisions across multiple analytical domains and provides evidence-based protocols to enhance methodological rigor in EEG research, with particular relevance for clinical trials and drug development studies where accurate biomarker quantification is essential.
ERP analysis is particularly vulnerable to pre-processing decisions, with direct consequences for decoding accuracy and interpretability. A comprehensive multiverse analysis of seven experiments revealed that pre-processing choices significantly influence decoding performance across diverse ERP components [12].
Table 1: Impact of Pre-processing Choices on ERP Decoding Performance
| Pre-processing Step | Option A | Option B | Impact on Decoding Performance | Magnitude of Effect |
|---|---|---|---|---|
| Ocular Artifact Correction | ICA-based removal | No correction | Decreased performance (except when artifacts are systematically related to condition) | -1% to -5% [12] |
| Muscle Artifact Correction | ICA-based removal | No correction | Decreased performance (especially for motor tasks) | Up to -5% [12] |
| High-Pass Filter Cutoff | 1.0 Hz | 0.1 Hz | Increased performance | Consistent increase [12] |
| Low-Pass Filter Cutoff | 20 Hz | 40 Hz | Increased performance (time-resolved decoding only) | Variable [12] |
| Baseline Correction | Longer interval | Shorter interval | Increased performance (EEGNet) | Variable [12] |
| Linear Detrending | Applied | Not applied | Increased performance (time-resolved decoding) | Variable [12] |
The observed performance changes stem from the complex interaction between neural signals and structured noise. Artifacts systematically correlated with experimental conditions can create false decodable patterns. For instance, in the N2pc experiment, ocular artifacts contained information about target position in the visual field, as participants made involuntary saccades toward targets. Removing these artifacts reduced decoding performance because the classifier was exploiting these systematic ocular movements rather than neural signals [12]. Similarly, in lateralized readiness potential (LRP) experiments, muscle artifacts from hand movements were predictive of response conditions [12].
These findings present a critical methodological dilemma: while uncorrected artifacts may increase decoding performance, this comes at the expense of interpretability and validity, as models may exploit structured noise rather than neural signals [12]. The RELAX pipeline addresses this challenge through targeted artifact reduction that specifically targets artifact periods of eye movement components and artifact frequencies of muscle components, thereby reducing artificial inflation of effect sizes while preserving neural signals [10] [11].
Functional connectivity (FC) analysis examines the statistical dependencies between neural signals from different brain regions, but its reliability is strongly influenced by pre-processing decisions and measurement approaches.
Table 2: Reliability of EEG Functional Connectivity and Complexity Measures
| Metric Category | Specific Measure | Reliability (ICC) | State-Dependency | Recommended Applications |
|---|---|---|---|---|
| Complexity Measures | Permutation Entropy (PE) | Good-to-excellent (ICC > 0.75-0.90) | Low | Biomarker development [13] |
| Phase-Based FC | Phase Lag Index (PLI) | Moderate to good | Moderate | Robust connectivity analysis [13] |
| Amplitude-Based FC | AECc | Poor to excellent | High | Requires caution [13] |
| Network Topology | Minimum Spanning Trees (MST) | Poor to good | High | Limited stability [13] |
| Complexity-Based FC | Joint Permutation Entropy (JPEINV) | Not fully established | Not established | Emerging measure [13] |
Conventional FC analyses often assume temporal stationarity, but recent research reveals that FC is dynamic and fluctuates with cognitive functioning [14]. Tensor-based approaches for dynamic FC (dFC) can identify change points signifying substantial shifts in network connectivity across participants. In studies with tinnitus patients, this method effectively captured connectivity shifts, revealing that visual stimulation alone produced insignificant connectivity changes, while combined visual stimulation with simultaneous High-Definition tDCS significantly altered brain networks [14].
The Grassmann distance provides a method for comparing differences between subspaces or states in tensor analysis, enabling quantification of how different two network states are from each other [14]. This approach maintains network structure integrity through tensor representations to identify dynamic changes where significant alterations to network structure occur across all subjects due to interventions [14].
Source localization transforms EEG signals from sensor space to neural source space, but this transformation is highly sensitive to pre-processing decisions, particularly artifact handling.
Conventional ICA-based artifact removal can bias source localization estimates due to imperfect component separation that removes neural signals along with artifacts [10] [11]. The subtraction of entire artifactual components often eliminates valuable neural information embedded within those components, leading to mislocalization of neural sources [11]. This bias is particularly problematic for clinical applications and drug development studies where precise spatial localization is essential for target engagement biomarkers.
Template-based source localization approaches provide a viable alternative when subject-specific structural MRI is unavailable. Validation across two independent datasets (HBN and COGBCI) demonstrated that established pipelines using eLORETA with the ICBM 2009c template and CerebrA atlas can produce neurophysiologically plausible activation patterns, showing expected differences between resting state and video-watching conditions, and progressive activation increases with cognitive workload difficulty [15].
The ecological validity of source localization can be established through permutation testing that compares source space amplitudes across different tasks or conditions [15]. This approach has revealed that:
The RELAX pipeline provides a targeted alternative to conventional ICA-based artifact removal [10] [11]:
Data Acquisition: Record EEG using standard protocols with sufficient channels for adequate spatial sampling (minimum 64 channels recommended for reliable source localization) [13] [15]
Initial Processing: Apply band-pass filtering (0.5-40 Hz) and segment data into epochs time-locked to events of interest
ICA Decomposition: Perform ICA to identify artifactual components, but do NOT remove entire components
Targeted Cleaning: For ocular components, remove only the artifact periods rather than entire components. For muscle components, remove only the artifact frequencies [10] [11]
Validation: Compare effect sizes with and without targeted cleaning; be suspicious of artificially inflated effects from residual artifacts [12]
Targeted Artifact Reduction Workflow
For analyzing time-varying connectivity patterns [14]:
Data Preparation: Preprocess high-density EEG (256 channels recommended) and compute source signals
Tensor Construction: Create a multi-mode tensor representing FC networks across participants, time, and conditions
Tensor Decomposition: Apply Tucker decomposition to identify latent patterns while preserving network structure
Change Point Detection: Use analysis of variance (ANOVA) to identify significant change points in connectivity patterns
State Summarization: Compute representative network states for intervals between change points using tensor-matrix projections
For verifying source localization accuracy without ground truth [15]:
Pipeline Setup: Implement an end-to-end pipeline with automatic pre-processing, eLORETA source estimation, and template head model (ICBM 2009c with CerebrA atlas)
Condition Comparison: Record data across different conditions (e.g., resting state vs. task engagement)
Permutation Testing: Perform non-parametric permutation testing on source space amplitudes between conditions
Plausibility Assessment: Verify that activation patterns align with established neurophysiological knowledge (e.g., visual cortex activation during visual tasks)
Cross-Validation: Apply identical pipelines to multiple independent datasets to confirm consistency
Table 3: Essential Research Reagents and Tools for EEG Pre-processing Research
| Tool/Resource | Type | Function | Availability |
|---|---|---|---|
| RELAX Pipeline | Software Plugin | Targeted artifact reduction for EEG | Freely available as EEGLAB plugin [10] [11] |
| MNE-Python | Software Library | EEG preprocessing, source localization, and analysis | Open-source Python package [13] [15] [12] |
| ERP CORE Dataset | Data Resource | Standardized ERP dataset with multiple components | Publicly available for methodological validation [12] |
| ICBM 2009c Template | Anatomical Atlas | Standardized head model for source localization | Publicly available [15] |
| CerebrA Atlas | Anatomical Atlas | Cortical labeling for source reconstruction | Publicly available [15] |
| BESA Connectivity | Commercial Software | Optimized workflows for connectivity analysis | Commercial license [18] |
EEG Pre-processing Impact and Solution Framework
Electroencephalography (EEG) aims to capture the brain's electrical activity, but the recorded signals are invariably contaminated by intruding signals from non-neural sources, known as artifacts. These artifacts obscure the genuine neural signals of interest, introducing uncontrolled variability that can reduce statistical power and confound experimental observations [2]. The amplitude of neural signals typically ranges in tens of microvolts, making them particularly susceptible to being blurred by artifacts, which can often have much larger amplitudes [2]. Best efforts are always made to prevent artifacts during recording; however, a certain level of intrusion remains unavoidable, especially with the growing use of mobile EEG in out-of-lab settings [2]. Effective handling of these artifacts during preprocessing is therefore an invaluable skill for EEG researchers. Artifacts can be broadly classified into two primary categories based on their origin: physiological artifacts, generated by the human body but not by brain activity, and technical artifacts, which originate from the EEG equipment or the recording environment [2]. This classification is fundamental, as the nature of the artifact often dictates the optimal strategy for its identification and remediation.
Understanding the distinct properties of different artifact types is the first step in developing an effective cleaning protocol. The tables below provide a detailed comparison of common physiological and technical artifacts, summarizing their key characteristics, impact on the signal, and common handling methods. This structured overview serves as a quick reference for researchers diagnosing contamination in their datasets.
Table 1: Physiological Artifacts in EEG Recordings
| Artifact Type | Typical Morphology | Spectral Profile | Topographic Distribution | Primary Handling Methods |
|---|---|---|---|---|
| Eye Blink | High-amplitude, slow positive deflection [2] | Delta/Theta bands [2] | Primarily frontal channels [2] | ICA, Regression-based subtraction [2] |
| Eye Movements | Box-shaped deflection with opposite polarity on hemispheres [2] | Delta/Theta bands, up to 20 Hz [2] | Lateral frontal & temporal channels [2] | ICA, Regression-based subtraction [2] |
| Muscle (EMG) | High-frequency, erratic, sharp activity [2] | Broad spectrum, up to 300 Hz, prominent >20 Hz [2] | Localized near muscle groups (e.g., temples, neck) [2] | Artifact rejection, Filtering, ICA for persistent artifacts [2] |
| Pulse (Cardiac) | Rhythmical, small spike pattern [2] | Overlaps with EEG rhythms [2] | Often affects mastoids and temporal regions [2] | ICA, Average reference, ECG-based removal [2] |
| Sweating/Skin Potentials | Very slow drifts and fluctuations [2] | Peak amplitude <1 Hz [2] | Global, across all channels [2] | High-pass filtering, Dry lab environment [2] |
| Body Movement | Large, slow shifts or complex-shaped deflections [2] | Low frequencies [2] | Global, across all channels [2] | Artifact rejection, Filtering, ICA for complex shapes [2] |
Table 2: Technical Artifacts in EEG Recordings
| Artifact Type | Typical Morphology | Spectral Profile | Topographic Distribution | Primary Handling Methods |
|---|---|---|---|---|
| Line Noise | Oscillatory, sinusoidal pattern [2] | 50 Hz or 60 Hz peak (and harmonics) [2] | Global, but can vary by electrode [2] | Notch filter, Band-rejection filters [2] |
| Loose Electrode Contact | Slow drifts or sudden, large "pops" [2] | Low frequencies for drifts [2] | Localized to the specific unstable electrode [2] | Artifact rejection, Channel interpolation/rejection [2] |
| Cable Movement | Transient signal alterations or oscillations [2] | Frequency of cable swing [2] | Often affects groups of channels [2] | Artifact rejection, Filtering (if non-overlapping) [2] |
The RELAX pipeline is a novel method designed to mitigate the pitfalls of standard artifact removal techniques. Traditional approaches, like blindly subtracting components identified as artifactual by Independent Component Analysis (ICA), can remove neural signals alongside artifacts. Counterintuitively, this can artificially inflate event-related potential and connectivity effect sizes and bias source localisation estimates [10] [11].
A comprehensive multiverse analysis has quantified how different preprocessing choices influence the performance of EEG-based decoding models, which is critical for brain-computer interfaces and multivariate pattern analysis.
The following workflow, implemented in software like BrainVision Analyzer, provides a structured approach for common artifact scenarios based on their type [2].
Table 3: Key Software Tools and Methods for EEG Artifact Handling
| Tool/Solution | Primary Function | Application Context | Key Considerations |
|---|---|---|---|
| RELAX (EEGLAB Plugin) | Targeted artifact reduction within ICA components [10] | Mitigating effect size inflation and source localization bias [11] | Freely available; targets artifact periods/frequencies instead of whole components [10] |
| Independent Component Analysis (ICA) | Blind source separation to isolate neural and artifactual components [2] | Removing ocular, muscle, and cardiac artifacts [2] | Imperfect separation can remove neural signal; basis for advanced methods like RELAX [10] |
| Autoreject Package | Automated algorithm for epoch rejection and channel interpolation [12] | Handling transient artifacts and bad channels [12] | Can reduce decoding performance if artifacts are task-related [12] |
| Notch Filter | Attenuates a very narrow frequency band [2] | Removing 50/60 Hz line noise from power lines [2] | Should be used sparingly; can distort waveform if applied excessively [2] |
| High-Pass Filter | Attenuates very low-frequency content [2] | Reducing slow drifts from sweat or movement [2] | Higher cutoffs (e.g., 1 Hz) can improve decoding performance [12] |
| BrainVision Analyzer | Commercial software suite for EEG analysis [2] | Implementing a full preprocessing pipeline including filtering and ICA [2] | Provides a guided workflow for artifact handling protocols [2] |
The rigorous classification of contaminants into physiological and technical artifacts is a cornerstone of robust EEG pre-processing. While traditional methods like ICA have been widely adopted, emerging research underscores their limitations, demonstrating that non-targeted component subtraction can inadvertently remove neural signals and create false positive effects. Modern solutions, such as the targeted cleaning implemented in the RELAX pipeline, offer a more nuanced approach that better preserves neural signals and enhances analytical validity. Furthermore, the relationship between preprocessing choices and downstream analysis goals, such as decoding performance, is complex and non-intuitive. Researchers must therefore balance the pursuit of high signal quality with the imperative of interpretable and valid results, carefully selecting and documenting their artifact handling protocols within the broader context of their research objectives.
Within electroencephalography (EEG) research, regression-based methods represent a foundational approach for the critical task of ocular artifact removal. These techniques are designed to separate neural signals from unwanted ocular interference, such as blinks and eye movements, which can significantly distort EEG readings. The core principle involves modeling the relationship between the recorded EEG signals and reference electrooculography (EOG) channels to estimate and subtract the artifact component. However, a major challenge known as bidirectional contamination can occur during this process, where the procedure inadvertently removes parts of the genuine neural signal along with the artifacts, or fails to fully remove the artifact, thus compromising the integrity of the cleaned EEG data. The performance of these methods is highly dependent on the accuracy of the reference signals and the underlying assumptions of linearity and stationarity between the EOG and EEG recordings. Effective application requires careful consideration of these factors to minimize signal distortion while maximizing artifact removal, ensuring the validity of subsequent neuroscientific or clinical interpretations. [19]
Regression methods operate on the principle of using a reference signal to estimate and subtract artifact contamination from the EEG data. The technique primarily relies on setting a reference channel and using a linear transformation to subtract the estimated artifact from the contaminated EEG, thus obtaining artifact-free EEG. [19] The fundamental model assumes that the recorded EEG signal is a linear combination of true brain activity and the artifact, which is captured by the EOG reference.
A significant limitation of this approach is its performance dependency. The performance of regression methods for artifact removal significantly decreases in the absence of a clean, well-defined reference signal. [19] Furthermore, obtaining a reference signal often requires a separate channel for recording, which increases the operational difficulty and cost of EEG acquisition. [19]
Bidirectional contamination is a critical problem in regression-based artifact removal, manifesting in two primary forms:
The risk of bidirectional contamination is heightened by the overlapping frequency spectra of physiological artifacts like EMG and EOG with the effective components of EEG, making clean separation difficult. [19] Preprocessing choices can inadvertently influence this balance; for instance, while artifact correction steps like ICA may reduce decoding performance, they are crucial for interpretability as uncorrected artifacts allow models to exploit structured noise rather than the neural signal. [12] [20]
Table 1: Comparison of EEG Artifact Removal Techniques
| Technique | Underlying Principle | Advantages | Limitations | Suitability for Ocular Artifacts |
|---|---|---|---|---|
| Regression [19] | Linear subtraction of artifact using reference EOG channels. | Conceptually simple, computationally efficient. | Prone to bidirectional contamination; requires separate reference channel. | Moderate |
| Filtering [19] | Removal of signal components in specific frequency bands. | Effective for non-physiological noise and narrowband artifacts. | Limited use due to frequency overlap between artifacts and EEG. | Low |
| Blind Source Separation (BSS) [19] | Separation of signals into independent components for manual/automatic artifact rejection. | Can be highly effective without a reference signal. | Requires manual inspection; may need many channels; risk of neural signal loss. | High |
| Deep Learning (e.g., CLEnet) [19] | End-to-end extraction of morphological/temporal features to separate EEG from artifacts. | Automated; can handle unknown artifacts and multi-channel data. | Requires large datasets for training; complex model architecture. | High |
This protocol details the steps for implementing a standard regression-based method for removing ocular artifacts, such as blinks and eye movements.
Primary Applications: Preprocessing of EEG data for cognitive neuroscience, clinical monitoring, and brain-computer interface (BCI) applications.
Materials and Reagents:
Procedure:
This protocol provides a methodology for comparing the performance of traditional regression against modern deep learning approaches like CLEnet, focusing on efficacy and the problem of bidirectional contamination.
Primary Applications: Evaluation and selection of artifact removal algorithms for specific research questions; validation of new methods.
Materials and Reagents:
Procedure:
Table 2: Quantitative Performance Comparison of Artifact Removal Methods
| Method | Artifact Type | SNR (dB) | Correlation Coefficient (CC) | RRMSEt | RRMSEf | Notes |
|---|---|---|---|---|---|---|
| Regression [19] | EOG | Not Reported | Not Reported | Not Reported | Not Reported | Known performance decrease without clean reference. [19] |
| CLEnet (Proposed) [19] | Mixed (EMG+EOG) | 11.498 | 0.925 | 0.300 | 0.319 | Best performance on semi-synthetic data. |
| CLEnet (Proposed) [19] | Multi-channel (Unknown) | - | - | - | - | SNR & CC ↑ 2.45% & 2.65% vs. DuoCL; RRMSEt & RRMSEf ↓ 6.94% & 3.30%. |
Table 3: Essential Research Reagent Solutions for EEG Artifact Removal Research
| Item | Function/Application | Example/Notes |
|---|---|---|
| ERP CORE Dataset [12] [20] | A public, standardized EEG dataset with event-related potentials (ERPs). | Used as a source of clean EEG for creating semi-synthetic benchmark data or for method validation. [12] [20] |
| EEGdenoiseNet [19] | A semi-synthetic benchmark dataset for EMG and EOG artifact removal. | Provides clean EEG, EMG, and EOG signals, enabling controlled evaluation of artifact removal algorithms. [19] |
| MNE-Python [12] [20] | A comprehensive open-source Python package for exploring, visualizing, and analyzing human neurophysiological data. | Used for implementing preprocessing pipelines, including filtering, referencing, and ICA. [12] [20] |
| CLEnet Model [19] | A deep learning model integrating dual-scale CNN and LSTM for end-to-end artifact removal. | Designed to handle multi-channel EEG data and various artifact types, including unknown artifacts. [19] |
| Independent Component Analysis (ICA) | A blind source separation technique used to isolate and remove artifact components from EEG data. | A common baseline/comparison method; effectiveness depends on manual inspection and sufficient channels. [19] |
Blind Source Separation (BSS) represents a cornerstone of modern electroencephalography (EEG) preprocessing, enabling the decomposition of complex scalp recordings into constituent cerebral and artifactual sources without prior knowledge of the mixing process. Within the framework of ocular artifact removal research, BSS techniques are indispensable for isolating and eliminating contaminants such as blinks and eye movements, which can severely obscure underlying neural signals. The fundamental assumption of BSS is that the recorded EEG is a linear mixture of statistically independent components originating from distinct neural and non-neural sources [21]. This principle allows researchers to segregate artifacts from brain activity based on their unique statistical properties, spatial topography, and temporal characteristics.
Independent Component Analysis (ICA) stands as the most widely adopted BSS method in EEG research. ICA operates on the premise that the time courses of underlying cerebral and artifactual sources are statistically independent of one another [22]. The algorithm computes an "unmixing" matrix that decomposes the multichannel scalp data into a sum of temporally independent and spatially fixed components [21]. The resulting independent components (ICs) can then be systematically examined for artifact characteristics, allowing researchers to reject contaminating sources while preserving neural activity of interest. The applicability of ICA, however, depends on several conditions: that the sources are spatially stable, that potential summation at electrodes is linear, and that propagation delays are negligible—assumptions generally considered reasonable for EEG data [21].
Recent advancements have expanded the BSS landscape beyond traditional ICA. Newer approaches include recurrent neural network (RNN) methods that transform input signals into corresponding ERP difference waveforms, achieving an interpretable, sparse source representation through L1 regularization [23]. Similarly, sparsity-based methods like Morphological Component Analysis (MCA) decompose signals sparsely in a predefined dictionary, showing particular promise for separating sources with distinct geometrical properties [24]. For ocular artifact removal specifically, hybrid methodologies such as REG-ICA have emerged, combining the strengths of BSS and regression techniques to address the limitation that artifactual components may contain residual neural activity [22].
The mathematical foundation of BSS rests on the linear mixture model, where observed EEG signals are represented as:
X = AS
Where X is the matrix of recorded EEG signals (electrodes × time points), A is the mixing matrix describing how sources combine at the electrodes, and S contains the time courses of the underlying sources [21]. The objective of ICA is to estimate a separating matrix W such that:
U = WX
yields statistically independent components, where U represents the independent components [21]. The mixing matrix A is the inverse of W, and its columns represent the scalp topographies of the corresponding components, providing crucial evidence for their physiological origins [21].
ICA achieves separation by optimizing for statistical independence between components, typically by maximizing non-Gaussianity through measures like kurtosis or negentropy, or by minimizing mutual information [23]. For the separation to be feasible, several assumptions must be met: (1) the source signals must be statistically independent at each time point, (2) the mixing process must be linear and instantaneous, (3) the number of observed mixtures must be at least equal to the number of sources, and (4) at most one source may follow a Gaussian distribution [21].
Various ICA algorithms have been developed, each with distinct optimization approaches:
Table 1: Common ICA Algorithms and Their Characteristics
| Algorithm | Optimization Approach | Advantages | Limitations |
|---|---|---|---|
| Extended-Infomax | Maximizes entropy using a flexible nonlinearity | Can separate both sub- and super-Gaussian sources; effective for EEG | Requires careful parameter tuning [22] |
| FastICA | Fixed-point iteration to maximize non-Gaussianity | Fast convergence; computationally efficient | Assumption of non-Gaussianity may not hold for all sources [25] |
| JADE | Joint approximate diagonalization of eigenmatrices | Robust for non-Gaussian sources; good separation performance | High computational complexity with many channels [26] |
Beyond ICA, other BSS approaches include second-order methods that exploit time structure and decorrelation, as well as sparsity-based methods that leverage the fact that sources may have sparse representations in certain domains [24]. The MODMAX algorithm, for instance, exploits the constant-envelope property of certain signals by minimizing envelope variance, offering reduced computational complexity suitable for hardware implementation [26].
Ocular artifacts manifest in EEG recordings with distinctive spatial, temporal, and spectral signatures that facilitate their identification. Eye blinks typically generate high-amplitude, biphasic potentials with characteristic frontal dominance due to the cornea's positive charge relative to the retina [21]. These artifacts appear as large, low-frequency deflections predominantly observed at frontal electrodes (Fp1, Fp2, F7, F8) with diminishing magnitude toward posterior sites [21]. The typical duration of a blink artifact ranges from 200-400 milliseconds, with a spectral focus below 4 Hz [27].
Horizontal eye movements produce smoother, more sustained potentials with alternating polarity based on direction, while vertical eye movements show more symmetric frontal distributions. These temporal and spatial patterns provide critical heuristics for component classification:
While visual inspection remains common, quantitative metrics enhance objectivity and reproducibility in component classification:
Table 2: Quantitative Metrics for Component Classification
| Metric | Application | Typical Values for Ocular Artifacts |
|---|---|---|
| Spatial Kurtosis | Measures peakedness of scalp topography | High values (>2) indicate focal frontal projections [28] |
| Power Spectral Density | Quantifies frequency content | Dominance in low frequencies (<4 Hz) [27] |
| Dispersion Entropy | Measures signal complexity | Lower values indicate more predictable artifact patterns [27] |
| Artifact to Signal Ratio | Quantifies contamination level | Higher values (>1) in artifactual ICs [22] |
Research demonstrates that artifactual Independent Components (ICs) contain significantly more ocular and less cerebral activity compared to contaminated electrode signals (artifact-to-signal ratio: 1.83 ± 3.65 for ICs vs. 0.69 ± 3.40 for electrode signals; p < 0.04) [22]. This concentration of artifacts in fewer components enables more targeted removal strategies.
The established protocol for ICA-based ocular artifact removal follows a systematic sequence:
Preprocessing Requirements: Before ICA, EEG data must undergo specific preprocessing: bandpass filtering (typically 1-30 Hz), bad channel identification and interpolation, and data segmentation [29]. Proper preprocessing is critical, as dimension reduction via Principal Component Analysis (PCA) before ICA can reduce the quality of the subsequent independent component decomposition [28].
ICA Computation: The ICA algorithm (e.g., Extended-Infomax) is applied to the preprocessed data to derive the unmixing matrix W and independent components [22]. The number of components should match the number of channels for optimal decomposition.
Component Rejection and Reconstruction: After identifying ocular components, clean data is reconstructed by projecting the remaining components back to sensor space using the mixing matrix A (the inverse of W):
cleandata = A[:,cleancomponents] × S[clean_components,:]
where clean_components represents the indices of non-artifactual components [21].
For challenging cases or specific research needs, advanced protocols offer enhanced artifact removal:
REG-ICA Protocol: This hybrid methodology addresses the limitation that ocular ICs may contain residual neural activity by applying regression to ICs rather than directly to EEG signals [22]. The approach involves:
Validation shows REG-ICA removes ocular artifacts more successfully than W-ICA (p < 0.01) or LMS (p < 0.01) while distorting brain activity less in time and frequency domains [22].
Single-Channel Approaches: For portable EEG systems with limited channels, techniques like Fixed Frequency Empirical Wavelet Transform (FF-EWT) combined with Generalized Moreau Envelope Total Variation (GMETV) filtering can effectively separate EOG artifacts by identifying contaminated components using kurtosis, dispersion entropy, and power spectral density metrics [27].
Table 3: Essential Tools for BSS Research in EEG
| Tool/Resource | Function | Application Notes |
|---|---|---|
| EEGLAB | MATLAB toolbox for EEG analysis | Provides ICA implementation, visualization, and component rejection tools [21] |
| ICA Toolkit | Specialized ICA functions | Includes multiple algorithms (FastICA, Extended-Infomax) and visualization utilities [21] |
| ICLabel | Automated component classification | Uses neural networks to label components as brain, ocular, muscle, etc. [23] |
| REG-ICA Plugin | Hybrid artifact removal | Implements regression-based correction of artifactual components [22] |
| RELICA | Reliability estimation for ICs | Assesses stability of independent components through resampling [28] |
Robust validation is essential for evaluating the efficacy of BSS-based artifact removal. Performance should be assessed using multiple metrics:
Table 4: Performance Metrics for BSS Artifact Removal
| Metric | Formula | Interpretation |
|---|---|---|
| Signal-to-Artifact Ratio (SAR) | SAR = 10log₁₀(σ²signal/σ²artifact) | Higher values indicate better artifact suppression [27] |
| Relative Root Mean Square Error (RRMSE) | RRMSE = √(∑(xclean - xprocessed)²/∑(x_clean)²) | Lower values indicate better preservation of neural signals [27] |
| Correlation Coefficient (CC) | CC = cov(xclean, xprocessed)/(σcleanσprocessed) | Values closer to 1 indicate better signal preservation [27] |
| Mean Absolute Error (MAE) | MAE = (1/N)∑⎮xclean - xprocessed⎮ | Lower values indicate less distortion of original signal [27] |
Studies evaluating automated BSS artifact correction algorithms report successful removal of 88% of artifacts in continuous EEG, with varying efficacy by type: ocular (81%), cardiac (84%), muscle (98%), and powerline (100%) [30]. These approaches outperform state-of-the-art algorithms in both artifact reduction rates and computation time [30].
The consequences of ICA preprocessing extend to subsequent EEG analyses. Research investigating EEG microstates—temporally stable patterns of brain activity—reveals that skipping ocular artifact removal affects microstate topography stability and reduces statistical power in eyes-open/eyes-closed comparisons [28]. However, provided that ocular artifacts are removed, microstate topographies and features remain robust to varying levels of additional preprocessing, enabling automated extraction pipelines [28].
Critical Parameters and Settings:
Effective implementation requires vigilant quality control:
Verification of Ocular Component Identification: Compare component topographies and time courses with simultaneous EOG recordings when available. Verify that blink-locked averages show characteristic frontal distribution [21].
Assessment of Neural Preservation: Examine power spectra of cleaned data to ensure preservation of physiologically relevant oscillations (alpha, beta, theta bands). Check that known neural responses (e.g., event-related potentials) remain intact after processing [22].
Troubleshooting Common Issues:
By adhering to these structured protocols and validation frameworks, researchers can implement BSS and ICA methodologies with confidence, ensuring effective ocular artifact removal while preserving the neural signals essential for both basic research and clinical applications.
Electroencephalogram (EEG) signals are fundamental tools for diagnosing neural disorders and understanding brain function, but their interpretation is often compromised by ocular artifacts (OAs), particularly those originating from eye blinks. These artifacts introduce high-amplitude, low-frequency components that obscure underlying neural activity, posing a significant challenge for accurate analysis [27] [31]. This challenge is especially acute in the context of single-channel EEG systems, which are essential for portable and wearable brain-computer interface (BCI) applications due to their minimalistic design and user comfort [27] [32]. In these resource-constrained environments, wavelet transform techniques, specifically the Discrete Wavelet Transform (DWT) and Stationary Wavelet Transform (SWT), have emerged as powerful solutions for ocular artifact removal. Their effectiveness stems from an ability to perform time-frequency analysis, which is crucial for handling the non-stationary nature of EEG signals and for selectively targeting artifact components without significant loss of cerebral information [31] [33]. This document details application notes and protocols for optimizing DWT and SWT with advanced thresholding techniques, providing a structured methodology for researchers and scientists engaged in EEG pre-processing for ocular artifact removal.
The selection between DWT and SWT is a critical first step in designing an EEG pre-processing pipeline. The table below summarizes the core characteristics, advantages, and limitations of each transform in the context of single-channel EEG artifact removal.
Table 1: Comparison of DWT and SWT for Single-Channel Ocular Artifact Removal
| Feature | Discrete Wavelet Transform (DWT) | Stationary Wavelet Transform (SWT) |
|---|---|---|
| Core Principle | Decomposes signal using filtering and downsampling, reducing coefficient length at each level [32]. | Employs filtering without downsampling, maintaining constant coefficient length across decomposition levels [31] [34]. |
| Translation Invariance | Not translation-invariant; small shifts in input signal can cause major changes in coefficients [34]. | Translation-invariant; provides stability against shifts in the input signal [31] [34]. |
| Computational Efficiency | Higher due to data reduction via downsampling [32]. | Lower, as it avoids data reduction, leading to increased computational load [31]. |
| Artifact Localization | May be less precise due to downsampling artifacts [34]. | Superior for precise artifact localization, as it avoids translation variance [31] [34]. |
| Best-Suited For | Applications where computational efficiency is prioritized and some signal shift sensitivity is acceptable. | Scenarios requiring high-fidelity artifact removal and preservation of signal integrity, such as clinical diagnostics [34]. |
A key challenge in deploying these transforms is the automatic selection of the appropriate decomposition level. The following table outlines a proven criterion based on the spectral characteristics of ocular artifacts.
Table 2: Guidelines for Automatic Decomposition Level Selection
| Transform | Sampling Frequency (Hz) | Targeted Artifact Frequency | Proposed Automatic Selection Method | Reference |
|---|---|---|---|---|
| DWT | Not Specified | Eye Blink Artifacts (0.5-12 Hz) | Select decomposition level where the approximation coefficients' frequency range encompasses the artifact. For ocular artifacts, this is typically a low-frequency band (< 4 Hz) [31] [33]. | [33] |
| SWT | 256 | Beta Band (~12-30 Hz) | Use the difference in skewness between approximation coefficients of consecutive levels. Decomposition stops when this difference stabilizes, indicating the artifact-dominated level has been reached [31]. | [31] |
The following table catalogues the key computational tools and materials required to implement the wavelet-based artifact removal protocols described in this document.
Table 3: Research Reagent Solutions for Wavelet-Based EEG Pre-processing
| Item Name | Function/Application | Specification Notes |
|---|---|---|
| Daubechies Wavelet (db4) | Mother Wavelet Function | Commonly selected for biomedical signal processing due to its morphological similarity to the shape of eye blink artifacts [31]. |
| Kurtosis (KS) | Statistical Metric for Component Identification | Used to identify components with high peakiness, a characteristic of artifact-dominated IMFs or wavelet coefficients [27]. |
| Dispersion Entropy (DisEn) | Nonlinear Metric for Component Identification | Measures the complexity and irregularity of signal components, aiding in the separation of artifactual patterns from neural signals [27]. |
| Sample Entropy | Metric for Component Rejection | Used to identify and reject intrinsic mode functions (IMFs) or independent components (ICs) corresponding to EOG artifacts based on signal complexity [32]. |
| Generalized Moreau Envelope Total Variation (GMETV) Filter | Advanced Filtering Technique | Integrated with decomposition methods like FF-EWT to effectively separate and remove artifact sources from single-channel EEG [27]. |
| Local Maximal and Minimal (LMM) | Thresholding Algorithm | An algorithm proposed for performing wavelet thresholding to remove artifacts without disturbing the information related to brain activity [33]. |
Objective: To remove ocular artifacts from single-channel EEG while addressing the overcomplete problem and mode aliasing associated with standard empirical mode decomposition. Background: Direct application of Independent Component Analysis (ICA) to single-channel EEG is not feasible. This protocol uses DWT and Complete Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) to create a multi-dimensional input for ICA, enabling effective source separation [32].
Step-by-Step Methodology:
db4) to a pre-determined level. This yields one set of approximation coefficients (A) and multiple sets of detail coefficients (D1, D2, ... Dn).Objective: To implement an efficient, hardware-friendly DWT thresholding method for portable BCI devices. Background: This protocol focuses on determining the optimal decomposition level and applying a novel Local Maximal and Minimal (LMM) thresholding technique to remove ocular artifacts with minimal loss of cerebral information [33].
Step-by-Step Methodology:
db4 wavelet to the level determined in Step 1.Objective: To automatically remove eye blink artifacts using SWT by leveraging a skewness-based criterion for decomposition level selection. Background: SWT's translation invariance makes it superior for precise artifact localization. This protocol automates the process of identifying the decomposition level where ocular artifacts are most prominent [31].
Step-by-Step Methodology:
db4 mother wavelet. Begin the process without a pre-fixed maximum level.i, calculate the skewness of the approximation coefficients.i) and the previous level (i-1).Objective: To utilize SWT for feature extraction to enable machine learning-based identification of artifactual components. Background: This protocol combines SWT with other signal processing techniques to create a robust framework for artifact removal, suitable for scenarios with multiple artifact types [36].
Step-by-Step Methodology:
The following diagrams illustrate the logical workflows for the core DWT and SWT protocols described in this document.
Diagram 1: DWT-CEEMDAN-ICA workflow for single-channel EEG.
Diagram 2: Automatic SWT (ASWT) with skewness-based decomposition.
Electroencephalography (EEG) is a fundamental tool in neuroscience and clinical diagnostics, providing non-invasive, high-temporal-resolution insights into brain activity. However, its utility is severely compromised by ocular artifacts (OAs)—unwanted signals originating from eye movements and blinks. These artifacts can exhibit amplitudes an order of magnitude greater than neural signals, drastically obscuring underlying brain activity and complicating analysis in both research and clinical settings [37] [38]. The pressing need for automated, precise, and robust ocular artifact removal (OAR) methods has catalyzed a shift from traditional techniques towards sophisticated deep learning (DL) paradigms. This document details the application notes and experimental protocols for emerging hybrid DL architectures, specifically focusing on Autoencoders and CNN-LSTM networks, which represent the cutting edge in EEG pre-processing for ocular artifact removal.
Deep learning models have demonstrated superior performance in removing various EEG artifacts, including ocular artifacts. The table below summarizes the quantitative performance of several state-of-the-art architectures, providing a benchmark for researchers.
Table 1: Performance Metrics of Deep Learning Models for EEG Artifact Removal
| Model Name | Core Architecture | Primary Application | Key Metrics and Performance |
|---|---|---|---|
| AnEEG [39] | LSTM-based GAN | General Artifact Removal | Lower NMSE & RMSE; Higher CC, SNR, and SAR vs. wavelet techniques |
| LSTEEG [40] | LSTM-based Autoencoder | Multi-channel Artifact Detection & Correction | Accurate artifact detection via anomaly detection; High correction performance |
| CLEnet [41] | Dual-Scale CNN + LSTM + EMA-1D | Multi-channel, Unknown Artifacts | CC: 0.925; SNR: 11.498 dB (Mixed artifacts); Outperforms 1D-ResCNN, NovelCNN, DuoCL |
| MSCGRU [42] | Multi-Scale CNN + BiGRU (GAN) | EMG, EOG, ECG Artifacts | RRMSE: 0.277 ± 0.009; CC: 0.943 ± 0.004; SNR: 12.857 ± 0.294 (for EMG) |
| ART [43] | Transformer (ICA-enhanced data) | Multi-channel, Multiple Artifacts | Outperforms other DL models in MSE, SNR; Improves BCI performance |
| EEGIFNet [42] | CNN + BiGRU (Dual-branch) | Recovering clean EEG & artifacts | Effective local and global feature capture |
Objective: To remove ocular and other physiological artifacts from multi-channel EEG data using a hybrid dual-scale CNN and LSTM architecture.
Figure 1: CLEnet End-to-End Workflow
Materials:
Procedure:
Model Construction:
Model Training:
Model Evaluation:
Objective: To detect ocular artifacts in multi-channel EEG data without clean labels, using an anomaly detection approach with an LSTM-Autoencoder.
Figure 2: LSTEEG Anomaly Detection Logic
Materials:
Procedure:
Autoencoder Training:
Anomaly (Artifact) Detection:
Model Evaluation:
Table 2: Essential Resources for EEG Ocular Artifact Removal Research
| Category | Item | Specification / Example | Primary Function in Research |
|---|---|---|---|
| Benchmark Datasets | EEGdenoiseNet [41] | Semi-synthetic; includes clean EEG, EOG, EMG. | Model training & benchmarking for specific artifact types. |
| EEG Eye Artefact Dataset [39] | Real EEG data with labeled ocular artifacts. | Model training & testing on real-world ocular artifacts. | |
| LEMON Dataset [40] | Pre-processed, clean EEG data. | Training autoencoders for unsupervised anomaly detection. | |
| Software & Libraries | ICLabel [37] | CNN-based ICA component classifier. | Automating the labeling of ICA components for training data generation. |
| PyTorch / TensorFlow | Deep Learning Frameworks. | Building, training, and deploying complex hybrid neural networks. | |
| MNE-Python | EEG processing toolbox. | Standard EEG I/O, filtering, preprocessing, and ICA computation. | |
| Evaluation Metrics | Correlation Coefficient (CC) [41] | Measures waveform similarity. | Quantifying preservation of original neural signal. |
| Signal-to-Noise Ratio (SNR) [41] | Measures denoising effectiveness. | Quantifying the level of artifact suppression. | |
| Root Mean Square Error (RMSE) [39] | Measures amplitude difference. | Quantifying the deviation from the ground-truth signal. |
The integration of hybrid deep learning architectures, particularly CNN-LSTM networks and advanced autoencoders, marks a significant leap forward in EEG pre-processing. These models directly address the core challenge of ocular artifact removal by leveraging the complementary strengths of spatial feature extraction and temporal sequence modeling. The provided protocols for CLEnet and LSTEEG offer actionable blueprints for researchers to implement these advanced techniques. As the field evolves, the integration of explainable AI (XAI) and self-supervised learning will further enhance the transparency, efficiency, and adaptability of these models, solidifying their role as indispensable tools in both neuroscience research and clinical drug development.
Electroencephalography (EEG) is a fundamental tool in clinical and cognitive neuroscience due to its non-invasive nature and high temporal resolution. However, the utility of EEG data is critically dependent on effective pre-processing, particularly the removal of artifacts that can obscure neural signals of interest. Artifacts originate from multiple sources, including ocular movements (EOG), muscle activity (EMG), cardiac signals (ECG), electrode noise, and environmental interference. The expansion of EEG into new domains such as wearable devices and real-world neuroimaging has further intensified the need for robust, standardized artifact removal pipelines. This document presents a comprehensive, practical workflow for integrating advanced artifact removal techniques into a complete EEG pre-processing pipeline, framed within the context of ocular artifact removal research.
EEG artifact removal presents unique challenges due to the overlapping spectral and temporal characteristics of neural signals and artifacts. Traditional methods have relied on regression, filtering, and blind source separation techniques, but these approaches often require manual intervention or perform poorly with unknown artifacts or low-density electrode arrays [19] [44]. The emergence of deep learning has transformed this landscape, enabling automated, data-driven approaches that can learn complex feature representations for superior artifact separation [45] [19].
A critical consideration in pipeline design is the intended application context. Wearable EEG systems with dry electrodes and reduced channel counts (often below 16) present distinct challenges compared to traditional high-density systems, including increased vulnerability to motion artifacts and limitations in applying source separation methods like Independent Component Analysis (ICA) [44]. Furthermore, different artifact types necessitate specialized removal approaches, as no single method performs optimally across all categories [45] [44].
This protocol integrates both established and emerging techniques to create a robust, semi-automated workflow for EEG pre-processing with a specific focus on artifact removal. The pipeline is designed to remove major artifacts while preserving neural signals, with particular emphasis on ocular artifacts.
Step 1: Data Import and Channel Localization
Step 2: Filtering and Line Noise Removal
Step 3: Bad Channel Detection and Interpolation
Step 4: Robust Re-referencing
Step 5: Ocular Artifact Removal using ICA
Step 6: Handling Large-Amplitude Transient Artifacts
Step 7: Final Quality Assessment and Export
Table 1: Performance Metrics of Deep Learning Models for Artifact Removal
| Model | Artifact Type | SNR (dB) | Correlation Coefficient | RRMSE (Temporal) | RRMSE (Spectral) |
|---|---|---|---|---|---|
| CLEnet [19] | Mixed (EOG+EMG) | 11.50 | 0.925 | 0.300 | 0.319 |
| Complex CNN [45] | tDCS | 9.80 | 0.910 | 0.315 | 0.335 |
| M4 (SSM) [45] | tACS/tRNS | 10.20 | 0.901 | 0.305 | 0.322 |
| DuoCL [19] | Mixed (EOG+EMG) | 9.85 | 0.899 | 0.325 | 0.342 |
| 1D-ResCNN [19] | Mixed (EOG+EMG) | 9.45 | 0.885 | 0.335 | 0.351 |
Deep learning models have demonstrated remarkable performance in artifact removal, particularly for handling unknown artifacts and multi-channel EEG data.
CLEnet Architecture: This dual-branch network integrates dual-scale Convolutional Neural Networks (CNN) with Long Short-Term Memory (LSTM) networks and an improved EMA-1D attention mechanism. The CNN components extract morphological features at different scales, while the LSTM captures temporal dependencies in EEG signals. The EMA-1D module enhances feature extraction through cross-dimensional interactions [19].
Model Selection Guidelines: Based on comparative studies:
Implementation Considerations: Deep learning approaches require appropriate training data, either from semi-synthetic datasets (where clean EEG is artificially contaminated with artifacts) or from real recordings with corresponding reference signals [19].
Artifact management in wearable EEG demands specialized approaches due to dry electrodes, motion artifacts, and low channel counts.
Adapted Pipelines: For wearable systems with limited channels:
Real-time Processing: Select computationally efficient algorithms for real-time applications, with deep learning approaches showing promising results despite higher computational demands [44].
Table 2: Research Reagent Solutions for EEG Artifact Removal
| Tool/Resource | Type | Primary Function | Application Context |
|---|---|---|---|
| PREP Pipeline [46] | Standardized Pipeline | Robust referencing, bad channel detection | Large-scale EEG standardization, batch processing |
| EEGLAB [47] | MATLAB Toolbox | ICA, PCA, basic preprocessing | General research, flexible processing workflows |
| qEEGt Toolbox [48] | Normative Database | Age-corrected normative SPM of source spectra | Clinical applications, quantitative analysis |
| CLEnet [19] | Deep Learning Model | Multi-artifact removal for single/multi-channel EEG | Complex artifacts, unknown noise sources |
| EEGDenoiseNet [19] | Benchmark Dataset | Semi-synthetic data for algorithm development | Method development, performance validation |
| LightningChart [49] | Visualization Library | Real-time EEG display, power spectrum visualization | Application development, data monitoring |
Diagram 1: Comprehensive EEG Pre-processing Workflow with Integrated Artifact Removal
Diagram 2: Artifact-Type Specific Removal Strategy Selection
This application note presents a comprehensive, practical workflow for integrating advanced artifact removal techniques into EEG pre-processing pipelines. By combining established methods like ICA and PCA with emerging deep learning approaches, researchers can address the complex challenge of artifact removal across diverse experimental contexts. The provided protocols, performance metrics, and implementation tools offer a foundation for standardized, reproducible EEG pre-processing with particular relevance to ocular artifact removal research. As EEG applications continue to expand into real-world environments and clinical applications, robust artifact removal pipelines will remain essential for extracting meaningful neural signals from contaminated recordings.
Electroencephalography (EEG) preprocessing is a critical step for isolating neural signals from non-neural artifacts. Paradoxically, common artifact removal techniques can inadvertently inflate effect sizes and introduce bias into subsequent analyses, potentially compromising the validity of neuroscientific and clinical findings. This phenomenon arises primarily from the imperfect separation of neural and artifactual sources in component-based cleaning methods. When artifact-dominated independent components are subtracted from the data, residual neural elements are also removed, while some artifact signals may remain. This imperfect decomposition can create structured noise that correlates with experimental conditions, leading to false positive effects or artificially enhanced event-related potential (ERP) components [11]. The paradox lies in the fact that steps taken to improve data quality can, under specific circumstances, reduce the very interpretability they are meant to enhance.
Recent empirical investigations have quantified this phenomenon across multiple experimental paradigms. Table 1 summarizes key findings from studies that systematically evaluated how artifact removal impacts downstream analysis metrics. The evidence suggests that the relationship between cleaning intensity and data quality is not monotonic, with potential costs emerging after a certain threshold.
Table 1: Quantitative Evidence of Cleaning Effects on EEG Metrics
| Study | Cleaning Method | Impact on Decoding Performance | Impact on ERP/Effect Size | Paradigms Tested |
|---|---|---|---|---|
| Communications Biology (2025) [12] | ICA, Autoreject | Decreased performance across 7 experiments | Not Reported | N170, MMN, N2pc, P3b, N400, LRP, ERN |
| NeuroImage (2025) [50] | ICA + Artifact Rejection | No significant improvement in most cases | Not Reported | Binary and multi-way classification tasks |
| Clinical Neurophysiology (2025) [11] | Traditional ICA | Not Reported | Artificial inflation of ERP effect sizes, biased source localization | Go/No-go, N400 |
| RELAX Targeted Cleaning [11] | Targeted period/frequency cleaning | Not Reported | Reduced effect size inflation, minimized source localization biases | Go/No-go, N400 |
The neurophysiological basis for this paradox involves the complex interaction between artifact properties and neural signals. Ocular artifacts, for instance, exhibit systematic spatial and temporal patterns that may overlap with cognitively-relevant neural activity. When these artifacts are correlated with experimental conditions—as in the N2pc paradigm where eye movements systematically relate to target position—their removal can actually reduce decoding performance by eliminating predictive information [12]. Similarly, muscle artifacts associated with motor responses in lateralized readiness potential (LRP) paradigms may contain condition-relevant information that improves decoding when preserved [12].
The relationship between artifact removal and decoding performance presents a particularly complex challenge. Multiple studies have demonstrated that common cleaning approaches often fail to improve—and sometimes even diminish—the performance of multivariate classifiers. A comprehensive investigation of seven common ERP paradigms revealed that artifact correction steps consistently reduced decoding performance across experiments and models [12]. This counterintuitive finding was observed for both neural network-based (EEGNet) and time-resolved logistic regression classifiers.
The mechanisms underlying this phenomenon include:
Notably, the impact varies across experimental contexts. In paradigms where artifacts are tightly coupled to the cognitive process of interest (e.g., eye movements in visual attention tasks), their preservation may be particularly consequential for decoding accuracy [12].
For traditional univariate analyses, the risks predominantly involve effect size inflation and spatial distortion. When components containing mixed neural and artifactual content are subtracted, the reconstruction process can create amplitude enhancements in remaining components, artificially inflating ERP effect sizes [11]. This inflation is particularly problematic for clinical applications where effect magnitude may inform diagnostic or therapeutic decisions.
Source localization represents another vulnerability, as the spatial fingerprints of artifacts can bias estimates of neural generators. Traditional ICA subtraction has been shown to systematically displace the apparent sources of neural activity, potentially leading to erroneous conclusions about the neural substrates of cognitive processes [11]. Targeted cleaning approaches that selectively remove artifact-dominated periods or frequencies demonstrate promise in mitigating these localization biases while preserving genuine neural signals.
The following workflow diagrams the comprehensive assessment of preprocessing impacts on EEG decoding performance:
Table 2: Research Reagent Solutions for EEG Cleaning Validation
| Item | Specification | Function/Purpose |
|---|---|---|
| EEG Recording System | 64-channel active electrode systems (e.g., BioSemi, BrainProducts) | High-quality data acquisition with minimal technical artifacts |
| ERP CORE Dataset [12] | 7 standard paradigms (N170, MMN, N2pc, P3b, N400, LRP, ERN) | Standardized benchmark for comparing preprocessing effects |
| MNE-Python | Version 1.4.0 or later | Comprehensive EEG processing with consistent implementation |
| EEGLAB + RELAX Plugin [11] | Latest version from GitHub repository | Targeted artifact cleaning for comparison studies |
| EEGDenoiseNet [41] | Publicly available benchmark dataset | Evaluation of deep learning artifact removal methods |
| Classification Frameworks | EEGNet [12], Time-resolved Logistic Regression [12] | Standardized assessment of decoding performance |
Data Acquisition and Selection
Multiverse Preprocessing Implementation
Table 3: Preprocessing Factors for Systematic Evaluation
| Preprocessing Step | Options/Variations | Implementation Details |
|---|---|---|
| High-Pass Filter | 0.1 Hz, 0.5 Hz, 1.0 Hz | Zero-phase Butterworth filter, order 4 |
| Low-Pass Filter | 30 Hz, 40 Hz, 70 Hz | Zero-phase Butterworth filter, order 4 |
| Ocular Artifact Correction | No correction, ICA, RELAX [11] | ICA: ICLabel for component classification |
| Muscle Artifact Handling | No correction, ICA, Autoreject [12] | Autoreject: per-subject threshold optimization |
| Reference Scheme | Average, Cz, REST | Apply consistency across pipeline |
| Baseline Correction | None, Pre-stimulus interval | -200 to 0 ms pre-stimulus |
| Detrending | None, Linear | Whole-trial linear detrending |
Decoding Analysis
Performance Quantification
Validation and Interpretation
The targeted cleaning approach focuses on preserving neural signals while specifically addressing artifact contamination:
Data Preparation and ICA Decomposition
Targeted Artifact Reduction
Signal Reconstruction and Validation
Source Localization Assessment
To mitigate the risks associated with artifact cleaning, researchers should implement the following methodological controls:
Based on empirical evidence, the following decision framework is recommended:
The paradoxical relationship between cleaning intensity and analytical validity necessitates careful consideration of preprocessing strategies in the context of specific research questions and analytical approaches. By implementing systematic validation protocols and adopting targeted cleaning methods, researchers can navigate the delicate balance between artifact removal and signal preservation.
Electroencephalography (EEG) is a vital tool in clinical and cognitive neuroscience, but the recorded signals are invariably contaminated by artifacts, with ocular (EOG) artifacts being one of the most pervasive challenges. Traditional preprocessing methods, such as the wholesale rejection of independent components (ICs) identified as artifactual, introduce a significant confound: the inadvertent removal of neural activity alongside the noise [10]. This can lead to the artificial inflation of effect sizes in event-related potentials and connectivity analyses, as well as biases in source localisation estimates [10]. In response, a paradigm shift towards targeted cleaning strategies is emerging. These strategies focus suppression efforts specifically on the temporal periods and frequency bands dominated by artifacts, thereby maximizing the preservation of the underlying neural signal [10]. This application note details the principles, protocols, and practical implementation of these targeted strategies, framing them within the critical context of preserving data integrity for research and drug development.
The limitation of conventional artifact removal, particularly Independent Component Analysis (ICA), is its all-or-nothing approach to components classified as artifactual. Due to imperfect component separation, subtracting an entire component inevitably removes some neural information [10]. This non-specific cleaning has demonstrable consequences:
Targeted cleaning mitigates these issues by operating on the principle that artifacts are often transient (e.g., an eye blink) or confined to specific frequency bands (e.g., muscle artifacts in the high-frequency range). Instead of rejecting entire components, these methods apply corrections only to the contaminated segments or frequencies, leaving the rest of the neural signal intact [10].
A foundational step in targeted cleaning is understanding the adversary. The table below summarizes key artifacts, their spectral and temporal characteristics, and the primary risk they pose to neural data.
Table 1: Characteristics of Major EEG Artifacts and Their Impact
| Artifact Type | Primary Source | Spectral Signature | Temporal Signature | Potential Impact on Neural Data |
|---|---|---|---|---|
| Ocular (EOG) | Eye blinks & movements [51] | Low-frequency (< 4 Hz) [51] | Transient, high-amplitude pulses [27] | Obscures delta/theta bands; volume conduction distorts frontal signals [51] |
| Muscle (EMG) | Head, jaw, neck muscle activity [51] | Broad spectrum (0 - >200 Hz) [51] | Can be transient or sustained | Masks high-frequency neural activity (e.g., gamma); complicates connectivity analysis |
| Cardiac (ECG) | Heartbeat [51] | ~1.2 Hz (pulse) [51] | Periodic, regular pattern [51] | Introduces periodic, non-neural spikes, particularly in datasets with few trials |
| Line Noise | Power supply (50/60 Hz) [52] | Very narrow band (50/60 Hz & harmonics) | Continuous | Can be mistaken for steady-state neural oscillations |
The efficacy of targeted cleaning can be quantified using standardized metrics. The following table compares the performance of various modern artifact removal methods, highlighting the quantitative advantage of targeted and advanced automated approaches.
Table 2: Performance Comparison of Contemporary Artifact Removal Methods
| Method Name | Core Approach | Reported Performance Metrics | Key Advantage |
|---|---|---|---|
| Targeted ICA Reduction [10] | Targets artifact periods/frequencies of ICs | Reduced effect size inflation & source localisation bias [10] | Preserves neural signals outside artifact epochs [10] |
| FF-EWT + GMETV [27] | Fixed-frequency decomposition & filtering | Lower RRMSE, higher CC on synthetic data; improved SAR & MAE on real EEG [27] | Effective for single-channel EEG; preserves low-frequency info [27] |
| NeuroClean [52] | Unsupervised ML pipeline with clustering-based IC rejection | >97% accuracy in motor task classification (vs. 74% for raw data) [52] | Fully automated; generalizes across tasks and subjects [52] |
| NeuXus [53] | LSTM-based R-peak detection & average subtraction for EEG-fMRI | Execution times <250 ms; performance on par with established offline tools [53] | Open-source, real-time capability for EEG-fMRI [53] |
This protocol is adapted from methods proven to mitigate effect size inflation in ERP analyses [10].
1. Prerequisite Preprocessing:
2. Component Classification & Targeting:
1 corresponds to samples within the identified artifact periods and 0 corresponds to clean neural data periods.3. Targeted Correction and Reconstruction:
This protocol provides a structured, step-by-step approach for effective artifact correction, emphasizing quality control at each stage [29].
1. Bandpass Filtering and Bad Channel Interpolation:
2. Ocular Artifact Correction using ICA:
3. Large-Amplitude Transient Artifact Correction using PCA:
4. Quality Checking and Data Export:
The following workflow diagram illustrates the key decision points in these protocols, contrasting the traditional and targeted approaches.
Successful implementation of targeted cleaning strategies requires a suite of robust software tools. The following table catalogs key open-source platforms that facilitate these advanced preprocessing workflows.
Table 3: Essential Open-Source Software for Targeted EEG Cleaning
| Tool Name | Platform | Primary Function | Relevance to Targeted Cleaning |
|---|---|---|---|
| MNE-Python [54] | Python | Comprehensive EEG/MEG processing & analysis. | Provides the core infrastructure (objects, filtering, ICA) for building custom targeted cleaning pipelines. Highly extensible. |
| EEGLAB [54] | MATLAB | Interactive EEG processing with GUI & plugins. | Ideal for initial exploration and manual IC classification; the RELAX plugin implements targeted cleaning [10]. |
| FieldTrip [54] | MATLAB | Advanced M/EEG analysis (code-driven). | Offers high flexibility for building custom, scripted analysis pipelines including sophisticated artifact handling. |
| Brainstorm [54] | Standalone (MATLAB/Java) | User-friendly M/EEG application. | Lowers the barrier to entry for advanced processing like source localization after effective artifact removal. |
| NeuXus [53] | Python | Real-time EEG processing. | Demonstrates the application of targeted methods (LSTM for pulse artifact) in real-time scenarios like neurofeedback. |
| NeuroClean [52] | Python | Unsupervised ML preprocessing pipeline. | Embodies the trend towards fully automated, machine-learning driven artifact removal that generalizes across datasets. |
The move towards targeted cleaning strategies represents a maturation of EEG preprocessing methodology. By shifting from the gross rejection of entire signal components to the precise suppression of artifacts in their specific temporal and frequency domains, researchers can directly address the critical confound of neural signal loss. The protocols and tools outlined in this application note provide a practical roadmap for implementing these strategies. Adopting these methods enhances the reliability, validity, and ultimately, the scientific value of EEG research, ensuring that conclusions are drawn from a data stream that reflects true brain activity as faithfully as possible.
The pursuit of clean electroencephalography (EEG) data for neuroscientific research and clinical applications necessitates the effective removal of ocular artifacts, which include blinks and eye movements. These artifacts pose a significant challenge as their frequency band (3–15 Hz) overlaps with crucial neural signals like theta and alpha rhythms, and their amplitude can be many times greater than that of cerebral activity [55] [51]. While a multitude of artifact removal techniques exist, the selection of an appropriate method for real-time applications—such as brain-computer interfaces (BCIs), neurofeedback, or affective human-robot interaction—hinges on a critical trade-off: achieving sufficient signal fidelity without introducing prohibitive computational latency [56]. This application note details current methodologies and provides structured protocols for researchers navigating this essential compromise.
The following sections and comparative tables summarize the core characteristics, performance metrics, and ideal use cases of the predominant ocular artifact removal techniques.
Table 1: Characteristics of Primary Ocular Artifact Removal Techniques
| Method | Core Principle | Key Advantages | Primary Limitations | Suitability for Real-Time Processing |
|---|---|---|---|---|
| Regression-Based [55] [51] | Linear subtraction of an artifact template derived from EOG or frontal EEG channels. | - Simple, computationally lightweight- Low processing time | - Assumes linearity and stationarity- Risk of over-cleaning and signal loss- Requires reference channels | Excellent |
| Independent Component Analysis (ICA) [57] [51] | Blind source separation to isolate and remove artifact components. | - High fidelity preservation for non-linear artifacts- Effective for multi-channel data | - Computationally intensive- Requires manual component inspection in traditional form- Decomposition quality reduced by motion artifacts | Challenging (without optimized, automated versions) |
| Artifact Subspace Reconstruction (ASR) [57] [55] | Sliding-window PCA identifies and reconstructs high-variance artifact subspaces based on a clean calibration period. | - Automated, requires minimal user input- Effective for large-amplitude motion artifacts | - Performance depends on calibration data and threshold (k)- Aggressive thresholds can "overclean" |
Good |
| iCanClean [57] | Canonical Correlation Analysis (CCA) to subtract noise subspaces correlated with reference (physical or pseudo) noise signals. | - Highly effective for motion and ocular artifacts- Improves ICA decomposition quality | - Requires noise reference- Performance depends on parameter tuning (R², window) | Good |
| Adaptive Filtering with Optimization [58] | Neural network-based adaptive filtering (e.g., NARX) with hybrid optimization algorithms (e.g., Firefly + LM) to find optimal weights. | - High performance in artifact removal- Achieves high Signal-to-Noise Ratio (SNR) | - High computational complexity- Longer computation time | Moderate to Challenging |
Table 2: Quantitative Performance Comparison of Featured Methods
| Method | Reported Performance Metrics | Key Computational Considerations |
|---|---|---|
| ICA | Considered a benchmark for quality; improves with high channel counts (e.g., >40) [55]. | High computational load for matrix decomposition and inversion; slower processing times. |
| ASR | Significantly reduces power at gait frequency and harmonics; improves component dipolarity [57]. | Processing time depends on k parameter and data window size; generally faster than ICA. |
| iCanClean | Superior to ASR in recovering dipolar brain components and capturing expected P300 effects; effective power reduction at artifact frequencies [57]. | Processing time influenced by CCA computation and window size; more effective but can be computationally heavier than ASR. |
| FLM-Optimized Adaptive Filter | Achieves high SNR (e.g., 42.042 dB) and low MSE [58]. | High computational cost due to hybrid optimization algorithm and neural network training. |
| Regression-Based | Foundational method; performance is acceptable with good reference signals [55] [51]. | Very low computational load and processing time, enabling the fastest real-time execution. |
ASR is an automated, component-based method suitable for removing large-amplitude artifacts in real-time or near-real-time scenarios [57] [55].
Workflow Overview
Step-by-Step Procedure
Calibration Phase: Establish Reference Data
Processing Phase: Real-Time Cleaning
k) compared to the reference data. A higher k value (e.g., 20-30) is less aggressive, while a lower value (e.g., 10) is more aggressive [57].iCanClean leverages canonical correlation analysis to remove artifact subspaces that are correlated with a noise reference, making it highly effective for motion and ocular artifacts [57].
Workflow Overview
Step-by-Step Procedure
Create Pseudo-Reference Noise Signal
Canonical Correlation Analysis (CCA)
Threshold and Subtract
Table 3: Essential Materials and Software for EEG Artifact Removal Research
| Item | Function & Application | Notes for Selection |
|---|---|---|
| High-Density EEG System (>40 channels) | Records scalp potentials. Essential for spatial methods like ICA, which benefit from a higher channel count [55]. | Systems with active electrodes and high input impedance are preferred for motion-rich environments. |
| Electrooculogram (EOG) Electrodes | Records eye movement and blink activity. Serves as a dedicated reference signal for regression-based and other reference-dependent methods [55] [51]. | Typically placed above/below the eye and at the outer canthi. |
| Dual-Layer Noise Sensors | Physically separate sensors that are mechanically coupled to EEG electrodes but not in contact with the scalp, capturing only motion artifact. The ideal reference for iCanClean [57]. | Not all EEG systems support these; pseudo-references can be used as an alternative. |
| Processing Software (EEGLAB, BCILAB) | Provides open-source environments for implementing ICA, ASR, and other processing algorithms. Essential for protocol development and testing [57]. | Plugins for ASR and iCanClean are available for EEGLAB. |
| Artifact Subspace Reconstruction (ASR) | A ready-to-implement algorithm for cleaning continuous EEG. Integrated into toolboxes like EEGLAB [57] [55]. | The key parameter k must be optimized for the specific study (recommended range 10-30). |
| iCanClean Algorithm | A specialized algorithm for artifact removal, particularly effective with motion. Available as a plugin or standalone code [57]. | Can be used with dual-layer or pseudo-reference signals. The R² threshold and window length are critical parameters. |
Electroencephalography (EEG) pre-processing is a critical determinant of data quality and interpretability, significantly influencing subsequent analysis and decoding outcomes. This article examines the effects of three pivotal pre-processing choices—filtering, referencing, and segmentation—on final EEG output, with particular emphasis on ocular artifact removal. Within a broader thesis on EEG pre-processing techniques, we systematically evaluate how methodological variations at these stages can either preserve neural signals or introduce systematic biases. Drawing on recent multiverse analyses and comparative studies, we demonstrate that preprocessing choices can considerably influence decoding performance, effect sizes, and the topological interpretation of functional brain networks. The article provides structured quantitative comparisons and detailed experimental protocols to guide researchers, scientists, and drug development professionals in constructing optimized, reproducible EEG pre-processing pipelines.
The selection of pre-processing parameters has measurable, and sometimes counterintuitive, effects on downstream EEG metrics. The tables below summarize key quantitative findings from systematic evaluations.
Table 1: Impact of Filtering Choices on Decoding Performance [12]
| Filter Type | Parameter Range | Effect on Decoding Performance | Notes |
|---|---|---|---|
| High-Pass Filter (HPF) | 1.0 Hz cutoff | Baseline performance | Lower cutoffs preserve slow neural signals but may include more drift. |
| Higher cutoffs (e.g., >1.0 Hz) | Consistent increase | Effect observed across experiments and models (EEGNet & time-resolved LR). | |
| Low-Pass Filter (LPF) | Lower cutoffs (e.g., <40 Hz) | Increased performance for time-resolved classifiers | Helps focus on relevant frequency bands for specific cognitive components. |
| Higher cutoffs (e.g., >40 Hz) | Mixed or decreased performance | May include high-frequency muscle noise that harms decoding. |
Table 2: Impact of Referencing and Other Steps on Output [12] [59]
| Pre-processing Step | Common Choices | Effect on Signal/Performance | Quantitative Findings |
|---|---|---|---|
| Re-Referencing | Common Average Reference (CAR) | Common baseline; affects global signal topography. | Similar topographical representations obtained with CAR, REST, and RESIT. [59] |
| Robust CAR (rCAR) | Reduces impact of outlier channels. | Showed the most different Event-Related Spectral Perturbation (ERSP) pattern. [59] | |
| REST / RESIT | Models an infinite reference; theory-driven. | Considered effective for standardizing data across studies. [59] | |
| Data Segmentation | Epoch length & alignment | Found to significantly affect the cleaning procedure. [59] | A critical step that can alter the effectiveness of subsequent artifact removal. |
| Baseline Correction | Varying time windows | Longer baseline windows were beneficial for decoding performance in most experiments. [12] | Particularly improved performance for EEGNet. |
| Linear Detrending | Application vs. Non-application | Had a positive effect on decoding for most experiments and frameworks. [12] | Helps remove slow linear drifts that can confound signal analysis. |
Table 3: Comparative Performance of Artifact Removal Methods [60] [11]
| Method | Primary Use | Key Performance Metrics | Trade-offs and Considerations |
|---|---|---|---|
| ICA-based Subtraction | General artifact removal (Ocular, Muscle) | Can artificially inflate ERP effect sizes and bias source localization. [11] | Imperfect component separation removes neural signals alongside artifacts. |
| Targeted Artifact Reduction (RELAX) | Ocular & Muscle artifacts | Effective cleaning while reducing effect size inflation and source localization biases. [11] | Targets artifact periods (eye movements) and frequencies (muscle) within components. |
| Average Artifact Subtraction (AAS) | BCG artifact in EEG-fMRI | Best signal fidelity (MSE = 0.0038, PSNR = 26.34 dB). [60] | Template-based; may not handle artifact variability well. |
| Optimal Basis Set (OBS) | BCG artifact in EEG-fMRI | Best structural similarity (SSIM = 0.72). [60] | Captures dominant variations in artifact structures. |
This protocol outlines a systematic grid-search approach to evaluate the combined impact of multiple pre-processing choices, as exemplified in recent literature [12].
1. Experimental Data Selection:
2. Defining the "Forking Paths" (Multiverse Construction):
3. Outcome Measurement:
4. Statistical Modeling:
This protocol tests the hypothesis that targeted artifact cleaning outperforms broad component subtraction, preserving neural signal integrity [11].
1. Data Preparation:
2. Independent Component Analysis (ICA):
3. Artifact Removal Interventions:
4. Outcome Comparison:
Table 4: Essential Software and Methodological Resources for EEG Pre-processing Research
| Tool / Resource | Type | Primary Function in Research | Application Notes |
|---|---|---|---|
| MNE-Python | Software Library | Provides a comprehensive, open-source toolkit for EEG/MEG data analysis, including filtering, ICA, epoching, and source localization. [12] | Enables the scripted, reproducible implementation of multiverse pre-processing pipelines. |
| EEGLAB | Software Environment | An interactive graphical and scriptable environment for processing EEG data, with a strong focus on ICA. [11] | Extensive plugin ecosystem (e.g., for the RELAX pipeline). Useful for component inspection and manual labeling. |
| RELAX Pipeline | EEGLAB Plugin | Implements targeted artifact reduction methods to clean ICA components without fully subtracting them. [11] | Crucial for testing hypotheses on effect size inflation and minimizing neural signal loss. |
| ERP CORE Dataset | Reference Dataset | A publicly available dataset containing seven classic ERP paradigms recorded from 40 participants. [12] [11] | Serves as an ideal benchmark for evaluating and comparing the efficacy of different pre-processing pipelines. |
| Autoreject | Software Library | An algorithm for the automated, data-driven rejection and interpolation of bad EEG epochs. [12] | Can be incorporated as a step in a multiverse analysis to assess its impact on decoding performance. |
| EEGNet | Neural Network Model | A compact convolutional neural network for EEG-based brain-computer interfaces and decoding tasks. [12] | Useful as a standardized decoder to objectively measure how pre-processing choices affect classification accuracy. |
Electroencephalography (EEG) preprocessing is a critical step for ensuring the validity of neural data, particularly in ocular artifact removal. Inadequate preprocessing can lead to three major pitfalls—over-correction, under-correction, and component misclassification—that compromise data integrity and interpretation. Over-correction occurs when neural signals are mistakenly removed alongside artifacts, reducing genuine brain activity. Under-correction leaves artifacts in the data, creating structured noise that can be mistakenly decoded as neural signals. Component misclassification happens when automated algorithms incorrectly label neural components as artifactual or vice versa. This application note examines these pitfalls within the context of multivariate pattern analysis (MVPA), provides quantitative assessments of their impact, and offers detailed protocols for optimizing artifact correction pipelines.
Recent empirical studies have systematically evaluated how preprocessing decisions affect decoding performance in EEG analysis. The tables below summarize key quantitative findings.
Table 1: Impact of Artifact Correction on Decoding Performance Across Multiple ERP Paradigms
| Preprocessing Approach | Effect on Decoding Performance | Key Findings | Paradigms Assessed |
|---|---|---|---|
| Artifact Correction (ICA) + Artifact Rejection | No significant improvement in most cases [50] [4] | Combination did not enhance performance for SVM/LDA classifiers | N170, MMN, N2pc, P3b, N400, LRP, ERN |
| Artifact Correction Alone | May decrease performance [12] | Reduces accuracy by removing predictive structured noise | Multiple experiments from ERP CORE |
| No Artifact Correction | Can artificially inflate performance [50] [12] | Classifiers may learn artifact patterns rather than neural signals | All paradigms assessed |
Table 2: Performance Trade-offs Between ICA and Artifact Blocking in Infant EEG
| Performance Metric | Independent Component Analysis (ICA) | Artifact Blocking (AB) |
|---|---|---|
| Sensitivity (Artifact Removal) | Higher effectiveness in correcting eye-movement segments [61] | Lower sensitivity to artifacts [61] |
| Specificity (Signal Preservation) | Causes more distortion to clean neural signals [61] | Higher specificity; distorts clean EEG less [61] |
| Signal-to-Noise Ratio (SNR) | Corrects artifacts effectively [61] | Corrects artifacts equally effectively as ICA [61] |
| Recommended Use Case | When complete artifact removal is prioritized | When preserving neural signal integrity is critical |
This protocol evaluates whether artifact correction improves or harms decoding accuracy for a specific research paradigm.
Materials and Setup
Procedure
Expected Outcomes: For paradigms where artifacts are condition-related (e.g., N2pc), Pipeline C (minimal preprocessing) may show highest accuracy, indicating potential over-correction in other pipelines [12].
This protocol directly compares the sensitivity and specificity of ICA versus Artifact Blocking for ocular artifact correction.
Materials and Setup
Procedure
Expected Outcomes: ICA is expected to show higher sensitivity (better artifact correction) while AB should demonstrate higher specificity (less distortion of clean signals) [61].
The following diagram illustrates the decision pathway for selecting and validating an artifact correction approach to avoid common pitfalls:
Artifact Correction Decision Workflow
Table 3: Essential Resources for Ocular Artifact Correction Research
| Resource Category | Specific Tool/Solution | Function in Research | Application Notes |
|---|---|---|---|
| Software Libraries | MNE-Python [12] | Comprehensive EEG processing | Implements filtering, ICA, and decoding |
| EEGLAB [61] | Component analysis | ICA implementation and ICLabels | |
| Benchmark Datasets | ERP CORE [12] | Method validation | Contains seven standard ERP paradigms |
| TUH EEG Artifact Corpus [62] | Algorithm training | Expert-annotated artifact labels | |
| International Infant EEG Data Integration Platform [61] | Specialized validation | Infant EEG with eye movement artifacts | |
| Classification Tools | SVM & LDA [50] [4] | Decoding performance assessment | Standard classifiers for EEG MVPA |
| EEGNet [12] | Neural network decoding | Deep learning approach for comparison | |
| Validation Metrics | Signal-to-Noise Ratio [61] | Correction effectiveness | Quantifies artifact reduction |
| Multiscale Entropy [61] | Signal integrity | Measures neural signal preservation | |
| T-sum/Average Accuracy [12] | Decoding performance | Quantifies classification effectiveness |
Effective management of over-correction, under-correction, and component misclassification requires paradigm-aware preprocessing strategies. Artifact correction should not be viewed as a one-size-fits-all process but rather as a carefully calibrated procedure that balances signal preservation with artifact removal. Researchers should routinely perform control analyses comparing decoding performance with and without artifact correction, particularly when studying paradigms where artifacts might be systematically related to experimental conditions. When implementing automated component classification, visual inspection of critical components remains essential to avoid component misclassification, especially when applying algorithms trained on adult data to specialized populations like infants. Future work should focus on developing more specific artifact detection methods that leverage deep learning approaches tailored to distinct artifact classes [62], particularly for wearable EEG systems where traditional methods like ICA may be less effective [44].
In electroencephalography (EEG) research, the removal of ocular artifacts is a critical preprocessing step to ensure the integrity of neural data. The establishment of robust, quantitative performance metrics is paramount for objectively evaluating and comparing the efficacy of various artifact removal algorithms [44]. Within the broader context of advanced EEG pre-processing techniques, this document details the application and protocol for four cornerstone metrics: Correlation Coefficient (CC), Mutual Information (MI), Signal-to-Artifact Ratio (SAR), and Normalized Mean Square Error (NMSE). These metrics provide a multifaceted assessment, evaluating aspects ranging from signal fidelity and information preservation to the sheer magnitude of artifact suppression [63] [64] [65]. Their standardized application is essential for driving reproducible research and technological advancements in the field, particularly for the growing domain of wearable EEG systems [44].
The following table summarizes the core mathematical definitions and ideal value ranges for each key performance metric.
Table 1: Definition and Interpretation of Key Performance Metrics
| Metric | Mathematical Formulation | Ideal Value | Primary Assessment Focus |
|---|---|---|---|
| Correlation Coefficient (CC) | (\rho = \frac{C(t1, t2)}{\sigma{x(t1)} \sigma{x(t2)}}) [63] | +1 or -1 | Linear relationship and waveform similarity between clean and processed signals [39] [63]. |
| Mutual Information (MI) | (I(U, V) = \sum \sum P(u, v) \log \frac{P(u, v)}{P(u)P(v)}) [63] | Higher values indicate better performance. | Shared information content, including non-linear dependencies [63] [65]. |
| Signal-to-Artifact Ratio (SAR) | (SAR = 10 \log{10} \frac{\sum (x{art}(n) - x(n))^2}{\sum (x_{proc}(n) - x(n))^2}) [63] | Higher values indicate better performance. | Degree of artifact removal achieved by the processing algorithm [63]. |
| Normalized Mean Square Error (NMSE) | (NMSE = 10 \log{10} \frac{\sum (x(n) - x{proc}(n))^2}{\sum x(n)^2}) [63] | 0 (or lower dB values) | Overall error magnitude and difference between the original clean and processed signals [39] [63]. |
This protocol outlines a standard procedure for quantitatively comparing the performance of different ocular artifact (OA) removal techniques using the defined metrics.
1. Objective: To evaluate and compare the efficacy of multiple OA removal algorithms (e.g., Wavelet Transform, ICA, FF-EWT) on a standardized dataset. 2. Materials and Datasets:
coif3 or bior4.4), apply a statistical threshold to detail coefficients, and reconstruct the signal [63].Table 2: Example Results from a Comparative Study of Wavelet-Based Techniques [63]
| Algorithm | Basis Function | Threshold | CC | MI | SAR (dB) | NMSE (dB) |
|---|---|---|---|---|---|---|
| DWT | coif3 | Statistical | 0.97 | 1.42 | 12.45 | -25.10 |
| DWT | bior4.4 | Statistical | 0.96 | 1.38 | 11.98 | -24.50 |
| SWT | sym3 | Universal | 0.93 | 1.25 | 9.85 | -21.70 |
| SWT | haar | Universal | 0.91 | 1.18 | 8.90 | -20.50 |
This protocol is tailored for assessing the performance of deep learning-based artifact removal models, such as the AnEEG network [39].
1. Objective: To validate the performance of a novel deep learning model (e.g., AnEEG) for ocular and muscular artifact removal. 2. Materials and Datasets:
Table 3: Essential Tools and Algorithms for EEG Artifact Removal Research
| Tool/Solution | Function in Research | Example Use Case |
|---|---|---|
| Discrete Wavelet Transform (DWT) | Decomposes single-channel EEG into time-frequency components for artifact identification and removal via thresholding [63] [65]. | Unsupervised ocular artifact removal using coif3 or bior4.4 wavelets with a statistical threshold [63]. |
| Fixed Frequency EWT (FF-EWT) | An adaptive signal decomposition technique that separates EEG into predefined frequency sub-bands, facilitating the isolation of artifact components [64]. | Automated identification and removal of EOG-related components using kurtosis and dispersion entropy [64]. |
| Independent Component Analysis (ICA) | A blind source separation method that decomposes multi-channel EEG into independent components, allowing for manual or automated rejection of artifact-related components [44] [66] [69]. | Removal of ocular and muscular artifacts from multi-channel EEG data by identifying and subtracting artifact-contributed components [44] [69]. |
| Non-Local Means (NLM) Filter | A denoising filter that reduces noise by averaging similar patterns (patches) from different parts of the signal, effective for non-stationary noise like EMG [65]. | Correcting wavelet coefficients corrupted by muscle artifacts in a hybrid WPD-NLM framework [65]. |
| Generative Adversarial Network (GAN) | A deep learning framework where a generator produces clean EEG from artifact-contaminated input, and a discriminator distinguishes it from real clean EEG [39]. | Artifact removal through adversarial training, often combined with LSTM layers to model temporal dependencies in EEG signals [39]. |
The following diagram illustrates the standard experimental workflow for evaluating artifact removal algorithms, from data preparation to performance assessment and validation.
Experimental Workflow for Metric Evaluation
The logical relationship between the performance metrics and the goals of artifact removal research is crucial for interpreting results. The diagram below maps this signaling pathway.
Logical Relationships Between Metrics and Research Goals
Electroencephalography (EEG) is a fundamental tool in neuroscience and clinical diagnostics due to its high temporal resolution and non-invasive nature. However, the utility of EEG is often compromised by physiological artifacts, with ocular artifacts (OAs) from eye blinks and movements being a primary source of contamination that can lead to misleading conclusions and diminish the performance of brain-computer interfaces (BCIs) [70] [71]. The selection of an optimal artifact removal strategy is therefore a critical preprocessing step. This application note provides a structured benchmark of three predominant methodological families for ocular artifact removal: Regression-based methods, Independent Component Analysis (ICA), and Wavelet-based techniques. Furthermore, we evaluate the emerging superiority of hybrid and deep learning approaches that integrate the strengths of these classical methods. Based on quantitative performance metrics across simulated and real EEG datasets, hybrid methods such as ICA-Regression and Wavelet-ICA consistently outperform individual techniques by more effectively separating neural activity from artifacts while preserving critical brain signal information [70] [72].
Table 1: Key Performance Metrics of EEG Artifact Removal Methods
| Method | Key Principle | Primary Strength | Primary Limitation | Reported SNR (dB) | Reported MSE |
|---|---|---|---|---|---|
| Regression | Linear subtraction of EOG from EEG [71] | Computational simplicity [71] | Removes neural activity correlated with EOG; requires reference signal [71] [41] | Information Not Provided | Information Not Provided |
| ICA | Blind source separation of EEG into independent components [70] | Effective separation without reference EOG [71] | Manual component selection; potential loss of neural info [70] [72] | 86.44 (Normal), 78.69 (ASD) [73] | Information Not Provided |
| Wavelet (DWT) | Time-frequency decomposition and thresholding [73] | Preserves signal characteristics; good feature retention [73] | Complex parameter selection (wavelet basis, threshold) [72] | Information Not Provided | 309,690 (ASD) [73] |
| Hybrid (ICA-Regression) | ICA decomposition followed by regression on artifact components [70] | High artifact removal; preserves neural activity [70] | Higher computational complexity [70] | Information Not Provided | Information Not Provided |
| Deep Learning (CLEnet) | End-to-end feature learning via neural networks [41] | Automated; adapts to various artifacts; high performance [41] | Requires large datasets for training [41] | 11.50 (on mixed artifacts) [41] | Information Not Provided |
Ocular artifacts are a major contaminant in EEG signals, characterized by their high amplitude and overlap with the frequency spectrum of neuronal signals, making them difficult to remove with conventional filters [72] [74]. The presence of these artifacts can significantly reduce the classification accuracy of brain-computer interfaces and lead to invalid conclusions in neuroscientific research [70] [71]. Effective preprocessing is therefore not merely a technical step but a foundational requirement for data integrity. While numerous techniques exist, the field lacks a universal standard, and the choice of method involves critical trade-offs between the completeness of artifact removal and the preservation of underlying neural signals [12]. This document addresses this gap by providing a quantitative benchmark and standardized protocols for the most prominent methodologies.
A systematic comparison of methods on standardized datasets reveals clear performance differences. A hybrid ICA–Regression method demonstrated significantly lower Mean Square Error (MSE) and Mean Absolute Error (MAE), alongside higher mutual information between the reconstructed and original artifact-free EEG, when compared to standalone ICA, regression, Wavelet-ICA (wICA), or Regression-ICA (REG-ICA) [70].
Table 2: Comprehensive Benchmarking of Artifact Removal Techniques
| Method | Residual Motion Artifact | Neural Signal Preservation | Computational Cost | Automation Level | Key Performance Findings |
|---|---|---|---|---|---|
| Regression | High (Bidirectional contamination) [71] | Low (Removes correlated neural activity) [71] | Low | Semi-Automatic | Outperformed by ICA and hybrid methods [70] |
| ICA | Moderate [70] | Moderate (Loss from component rejection) [72] | Medium | Low (Manual IC inspection) [70] | Increased detection performance by 10-20% over regression [72] |
| Wavelet (DWT) | Information Not Provided | High (Robust feature preservation) [73] | Medium | High | Achieved lowest MAE (4785.08) and MSE (309,690) in ASD data [73] |
| Hybrid (ICA-Regression) | Low [70] | High (Recovers neural activity from artifactual ICs) [70] | High | High | Superior performance in lower MSE/MAE and higher mutual information [70] |
| Deep Learning (ART/CLEnet) | Low [41] | High (End-to-end reconstruction) [43] [41] | Very High (for training) | Full End-to-End | Surpassed other DL models; improved BCI performance [43] [41] |
The choice of preprocessing method directly influences the success of subsequent analyses. A 2025 multiverse study demonstrated that artifact correction steps, particularly ICA, generally reduced trial-wise decoding performance in EEG-based classifiers [12]. This counterintuitive result occurs because artifacts can be systematically correlated with task conditions; for example, in a visual attention task (N2pc), ocular artifacts related to target-directed saccades are informative for the decoder. Removing these artifacts thus removes predictive information, potentially reducing accuracy but improving the interpretability and neurological validity of the model [12].
This protocol details the automated method proven effective in [70].
The following workflow diagram illustrates this multi-stage process:
This protocol, adapted from [72] [75], combines ICA with wavelet denoising to recover neural activity from artifactual components.
Table 3: Essential Tools for EEG Artifact Removal Research
| Tool / Resource | Function / Description | Relevance in Research |
|---|---|---|
| EEGdenoiseNet [41] | A semi-synthetic benchmark dataset containing clean EEG, EOG, and EMG. | Provides standardized data for training and fair evaluation of new artifact removal algorithms, especially deep learning models. |
| ERP CORE [12] | A public dataset containing event-related potentials from seven classic paradigms. | Ideal for assessing how preprocessing choices affect downstream decoding performance of neural signals. |
| ICA Algorithms (e.g., FastICA) | A family of blind source separation algorithms implemented in toolboxes like EEGLAB. | The core engine for decomposing multi-channel EEG into independent sources for manual or automated artifact removal. |
| Stationary Wavelet Transform (SWT) | A wavelet transform that is translation-invariant, unlike the Discrete Wavelet Transform (DWT). | Preferred over DWT for signal denoising as it avoids aliasing artifacts and provides more stable performance [72]. |
| Deep Learning Models (e.g., CLEnet [41], ART [43]) | End-to-end neural networks (CNN, LSTM, Transformer) that map noisy EEG to clean EEG. | Represents the state-of-the-art for fully automated, high-performance removal of multiple artifact types without manual intervention. |
The benchmarking data and protocols presented herein lead to the following application notes:
In conclusion, while classical methods like regression and standalone ICA provide foundational techniques, the field is moving toward sophisticated hybrid and deep learning approaches that offer a more nuanced and effective separation of brain and artifact, thereby enhancing the reliability of EEG-based research and clinical applications.
Within electroencephalography (EEG) research, the imperative for robust validation methodologies is paramount, particularly when developing and deploying pre-processing techniques for ocular artifact removal. Application-centric validation moves beyond theoretical performance to assess how methods perform under real-world conditions, ensuring that enhancements in signal processing translate into genuine improvements in data quality and subsequent analysis. This framework is essential for evaluating techniques across key domains of EEG analysis: the integrity of Event-Related Potentials (ERPs), the fidelity of spectral analysis, and the accuracy of source localization. This document provides detailed application notes and experimental protocols to guide researchers in the pharmaceutical and neuroscientific communities in implementing rigorous, application-centric validation for their EEG research.
ERPs are a cornerstone of cognitive neuroscience and neuropharmacology, providing insights into brain function in response to specific stimuli. However, ocular artifacts (blinks and eye movements) introduce substantial noise that can obscure or mimic genuine neural activity [51]. The validation of artifact removal techniques must therefore demonstrate not only the reduction of noise but, crucially, the preservation of underlying neural signals. Recent evidence indicates that preprocessing choices, including artifact correction, can significantly influence downstream decoding performance, underlining the need for rigorous validation [12].
A comparative analysis of common artifact removal methods reveals a trade-off between noise removal and signal preservation. The following table summarizes key performance metrics from the literature, providing a benchmark for validation studies.
Table 1: Performance Comparison of Ocular Artifact Removal Methods for ERP Studies
| Method | Principle | Advantages | Limitations | Reported Impact on ERP Decoding Performance |
|---|---|---|---|---|
| Regression (Time/Frequency Domain) | Estimates and subtracts artifact contribution based on EOG reference channels [51] | Simple, computationally efficient; requires reference channel | Prone to over-correction; can remove neural activity; bidirectional contamination [51] | Not explicitly tested in recent multiverse study [12] |
| Blind Source Separation (BSS) - ICA | Decomposes EEG into components; artifactual components are manually or automatically removed [51] | Effective for multiple artifact types; does not require reference channel | Subjective component selection; risk of removing neural data; computationally intensive [51] | Generally decreased decoding performance across multiple ERP components (e.g., ERN, N2pc) [12] |
| Wavelet Transform | Decomposes signal in time-frequency domain; thresholds artifact coefficients | Good for non-stationary artifacts; localized processing | Parameter selection (wavelet type, thresholds) is critical | Not explicitly tested in recent multiverse study [12] |
| Hybrid Methods | Combines multiple approaches (e.g., ICA-Wavelet) to enhance performance [51] | Can leverage strengths of individual methods; potentially higher accuracy | Increased complexity; more parameters to optimize | Not explicitly tested in recent multiverse study [12] |
This protocol outlines a comprehensive procedure for validating the efficacy of an ocular artifact removal method on ERP data.
Aim: To determine the impact of a specific artifact removal technique on the amplitude, morphology, and discriminative power of Event-Related Potentials.
Materials and Reagents:
Procedure:
Data Preprocessing (Pre-Artifact Removal):
Artifact Removal Implementation:
Validation & Outcome Measures:
Spectral analysis of EEG, encompassing power in delta, theta, alpha, beta, and gamma bands, is critical for investigating brain states in basic research and as a biomarker in clinical trials. Ocular artifacts introduce low-frequency noise that disproportionately affects lower frequency bands (e.g., delta, theta) [51]. Validation must ensure that observed spectral changes are neural in origin and not a residual artifact of the cleaning process.
This protocol assesses the impact of artifact removal on subsequent power spectral density and connectivity analyses.
Aim: To evaluate the efficacy of an ocular artifact removal method in preserving the integrity of neural oscillatory activity across the frequency spectrum.
Materials and Reagents:
Procedure:
Data Preprocessing & Artifact Removal:
Validation & Outcome Measures:
Figure 1: Workflow for validating the impact of artifact removal on spectral analysis.
Source localization algorithms, such as LORETA and MUSIC, infer the intracranial generators of scalp-recorded EEG activity. This is particularly valuable for pre-surgical epilepsy evaluation and cognitive neuroscience [76]. Ocular artifacts create widespread, non-neural电场分布 that severely distort the forward model, leading to spurious source estimates. Validation in this context is the most challenging, as it requires a "ground truth" for the source location.
This protocol leverages simultaneous intracranial and extracranial recordings to provide a direct validation of source localization accuracy before and after artifact removal.
Aim: To quantify the improvement in source localization accuracy achieved by an ocular artifact removal method, using intracranial recordings as a ground truth reference.
Materials and Reagents:
Procedure:
Data Preprocessing & Artifact Removal:
Source Localization:
Validation & Outcome Measures (Using Ground Truth iEEG):
Table 2: Essential Tools and Software for EEG Pre-processing and Validation Research
| Item Name | Function/Application | Example/Note |
|---|---|---|
| EEGLAB | An interactive MATLAB toolbox for processing continuous and event-related EEG, MEG, and other electrophysiological data. | Essential for performing ICA and visualizing components for manual rejection. |
| MNE-Python | Open-source Python software for exploring, visualizing, and analyzing human neurophysiological data. | Highly flexible for building automated preprocessing pipelines, including filtering, ICA, and source localization [12]. |
| FieldTrip | MATLAB toolbox for the analysis of EEG, MEG, intracranial EEG, and other electrophysiological data. | Particularly strong for advanced spectral analysis and source reconstruction methods. |
| ERPLAB | A MATLAB plugin for EEGLAB, specifically designed for analyzing ERP data. | Provides robust tools for averaging, quantifying, and statistically analyzing ERPs. |
| LORETA/Kyoto | Software for localizing the sources of EEG activity in the brain. | A widely used implementation of the Low-Resolution Brain Electromagnetic Tomography algorithm. |
| AutoReject | A Python-based tool for automated, data-driven artifact rejection and correction of ERP/EEG data. | Can be integrated into MNE-Python pipelines to objectively handle artifact rejection [12]. |
| Cartool | A software for the functional mapping and analysis of the brain's electrical activity from EEG/MEG. | Known for its powerful and user-friendly visualization of EEG source imaging results. |
| BCI2000 | A general-purpose platform for brain-computer interface (BCI) research. | Useful for real-time EEG acquisition and processing experiments. |
Figure 2: Multi-modal validation framework integrating ERP, spectral, and source localization assessments.
Electroencephalography (EEG) source localization enhances the spatial resolution of neural electrical activity, bridging the gap between its excellent temporal resolution and traditionally limited spatial accuracy [15]. However, a significant challenge arises in real-world research and clinical settings where subject-specific structural magnetic resonance imaging (MRI) is often unavailable due to cost, practicality, or clinical constraints. Template-based approaches that use standardized head models present a viable solution, but their validity critically depends on robust and validated pre-processing pipelines [15]. This challenge is particularly acute in ocular artifact removal research, where artifacts can constitute high-amplitude signals that dominate the EEG recording [55]. The pre-processing steps chosen to mitigate these artifacts directly shape the signal quality upon which source localization depends [12]. This Application Note provides a structured framework for validating pre-processing choices specifically for template-based EEG source localization, ensuring neurophysiologically plausible results without subject-specific anatomical information.
Template-based source localization employs standardized head models derived from population-averaged anatomical templates, such as the ICBM 2009c Nonlinear Symmetric template, to solve the EEG inverse problem [15]. The fundamental challenge is an ill-posed problem where infinitely many possible electrical source configurations could explain the recorded scalp potentials [15]. While individual anatomical variations can influence precision, research demonstrates that template-based approaches can achieve localization accuracies comparable to individual MRI-based solutions in many scenarios [15]. The viability of these methods hinges on the pre-processing pipeline's ability to preserve neural signals while removing non-neural contaminants that can distort source estimates.
Ocular artifacts, including blinks and saccades, present a particular challenge for source localization due to three key characteristics:
Table 1: Characteristics of Major EEG Artifacts Affecting Source Localization
| Artifact Type | Spectral Characteristics | Spatial Distribution | Primary Impact on Source Localization |
|---|---|---|---|
| Ocular Blinks | 3-15 Hz (overlaps theta/alpha) | Primarily frontal regions | Frontal source displacement, reduced accuracy in anterior regions |
| Eye Movements | 3-15 Hz | Frontal to central regions | Mislocalization to frontal eye fields, distorted anterior patterns |
| Muscle Artifacts | 0 Hz to >200 Hz | Focal or generalized | Widespread noise, false sources near muscle groups |
| Cardiac Artifacts | ~1.2 Hz (pulse) | Often focal near vessels | Periodic false sources, particularly in temporal regions |
Validating pre-processing pipelines for template-based source localization requires data-driven performance metrics that assess both reproducibility and biological plausibility. The NPAIRS (Nonparametric Prediction, Activation, Influence, and Reproducibility Resampling) framework provides two essential metrics [77]:
For template-based source localization specifically, permutation testing of source space amplitudes across different task conditions provides a robust validation approach [15]. This method tests whether the reconstructed source activations exhibit expected neurophysiological patterns despite the absence of subject-specific anatomical information.
When using functional templates for validation, different template types demonstrate varying performance characteristics. A comparative study of language templates found significant differences in their ability to capture primary language areas [78]:
Table 2: Performance Comparison of Language Template Types for Capturing Primary Language Areas
| Template Type | Basis of Generation | Anterior PLA Inclusion | Posterior PLA Inclusion | False Inclusion Rate | Best Use Case |
|---|---|---|---|---|---|
| Anatomical Template | Brain anatomy | Highest (PIF: 0.95) | Lower | Moderate | Screening for anterior language areas |
| Task-fMRI Template | Healthy subject task-fMRI | Moderate | High (PIF: 0.91) | Low (FIF: 0.061) | Posterior language localization |
| Resting-state fMRI Template | Healthy subject rs-fMRI | Lower | Moderate | Lowest (FIF: 0.054) | Specificity-focused applications |
| Meta-analysis Template | Literature synthesis | Balanced (DSC: 0.30 anterior) | High (PIF: 0.90) | Moderate | General-purpose applications |
The following workflow diagram illustrates the comprehensive pipeline for template-based EEG source localization with integrated validation checkpoints:
EEG Source Localization Validation Pipeline
ICA remains a cornerstone technique for ocular artifact removal, particularly in high-density EEG systems (≥40 channels) [55]. The standard approach involves:
For EEG systems with fewer channels, regression-based methods provide an alternative approach [55]:
To establish neurophysiological plausibility without ground truth, implement the following validation protocol:
Table 3: Essential Tools for Template-Based EEG Source Localization Research
| Tool/Category | Specific Examples | Function/Purpose | Implementation Considerations |
|---|---|---|---|
| Pre-processing Software | MNE-Python, EEGLAB, RELAX Plugin | Signal filtering, artifact removal, data visualization | RELAX specifically targets artifact periods to minimize signal loss [10] |
| Source Localization Algorithms | eLORETA, sLORETA, Beamforming | Solving the inverse problem to estimate neural sources | eLORETA shows good performance with template head models [15] |
| Head Models | ICBM 2009c Nonlinear Symmetric Template, MNI Colin27 | Standardized anatomical reference for source estimation | ICBM 2009c compatible with CerebrA atlas for precise ROI definition [15] |
| Validation Frameworks | NPAIRS, Permutation Testing | Assessing reproducibility and biological plausibility | NPAIRS provides SPR and PE metrics without ground truth [77] |
| Deep Learning Artifact Removal | ART (Artifact Removal Transformer), Custom CNNs | Advanced denoising using neural networks | ART uses transformer architecture for end-to-end denoising [43] |
| Template Libraries | CerebrA Atlas, Automated Anatomical Labeling (AAL) | Defining regions of interest for source analysis | Template selection significantly impacts results (see Table 2) [78] |
Recent advances in deep learning offer promising alternatives to traditional artifact removal methods:
Artifact Removal Transformer (ART): A transformer-based, end-to-end model that effectively removes multiple artifact types simultaneously from multichannel EEG data [43]. ART is trained on pseudo clean-noisy data pairs generated via ICA and demonstrates superior performance in restoring EEG signals compared to other deep learning approaches [43].
Convolutional Neural Networks (CNNs): Specialized lightweight CNNs can be optimized for specific artifact types with distinct temporal window sizes [62]:
A systematic multiverse analysis reveals that pre-processing choices significantly influence downstream analysis outcomes [12]:
The following diagram illustrates the decision pathway for optimizing pre-processing parameters based on research goals:
Pre-processing Optimization Decision Pathway
Template-based source localization without subject-specific MRI presents a feasible approach for EEG research when coupled with rigorously validated pre-processing pipelines. The key to success lies in selecting artifact removal methods appropriate for the specific research context—whether prioritizing decodability for BCI applications or interpretability for clinical research—and systematically validating the entire pipeline using data-driven performance metrics. Targeted artifact reduction approaches that clean only artifact periods preserve neural signals more effectively than complete component subtraction, minimizing biases in subsequent source localization. As deep learning methods continue to advance, they offer promising avenues for more precise artifact removal while preserving neural signals essential for accurate source reconstruction.
Electroencephalography (EEG) is a fundamental tool in neuroscience research and clinical applications, but its signal integrity is consistently compromised by ocular artifacts—unwanted signals generated primarily by eye blinks and movements. These artifacts pose a significant challenge because their frequency spectrum overlaps with neuronal signals of interest, potentially obscuring genuine brain activity and leading to misinterpretation of data [33]. The development of effective artifact removal strategies is particularly crucial for emerging applications in portable and wearable brain-computer interface (BCI) devices, where signal quality must be maintained under less controlled conditions [33] [79]. This document establishes a comprehensive framework for objectively selecting ocular artifact removal methods based on specific research goals and EEG setup constraints, providing researchers with a systematic approach to this critical pre-processing decision.
The challenge of ocular artifacts extends beyond simple signal contamination; different artifact types exhibit distinct spatial, temporal, and spectral characteristics that require tailored detection and removal strategies [79]. Without clear classification and appropriate processing, pipelines risk applying overly generic solutions that can be ineffective or may even compromise the integrity of neurophysiological components of interest [79]. This framework addresses these complexities by integrating technical performance metrics with practical implementation considerations across diverse research scenarios.
Traditional approaches to ocular artifact removal encompass several methodological families, each with distinct operational principles and limitations. Regression-based methods typically require a separate electro-oculogram (EOG) reference channel and use linear transformation to subtract estimated artifacts from contaminated EEG. While conceptually straightforward, their performance significantly decreases without a reference signal, and the additional recording channel increases operational complexity [19]. Filtering techniques face fundamental limitations due to the significant spectral overlap between ocular artifacts and neuronal signals, making clean separation in the frequency domain challenging [19].
Blind Source Separation (BSS) methods, particularly Independent Component Analysis (ICA), map artifact-contaminated signals into a new data space where artifactual components can be identified and removed. ICA is especially effective for isolating ocular artifacts without a reference signal [47] but requires multi-channel EEG data and sufficient prior knowledge for component selection. A significant limitation emerges when processing single-channel EEG, where ICA may remove cerebral information along with artifacts during signal reconstruction [33]. Wavelet-based approaches like Discrete Wavelet Transform (DWT) employ decomposition and thresholding techniques on wavelet coefficients to detect and remove artifacts while preserving cerebral information. These methods effectively retain the original shape of EEG in both frequency and temporal domains [33], though selecting appropriate threshold values and decomposition levels presents algorithmic challenges [33].
Table 1: Comparison of Traditional Ocular Artifact Removal Methods
| Method | Key Mechanism | Channel Requirements | Advantages | Limitations |
|---|---|---|---|---|
| Regression | Linear transformation using EOG reference | Requires EOG channel | Simple implementation | Performance depends on reference quality |
| Filtering | Frequency-domain separation | Single or multi-channel | Computationally efficient | Limited by spectral overlap |
| ICA | Blind source separation | Multi-channel preferred | Effective without EOG reference | Requires manual component inspection |
| Wavelet Transform | Time-frequency decomposition & thresholding | Single or multi-channel | Preserves temporal structure | Threshold selection critical |
| Hybrid (DWT-LMM) | Wavelet + Local Maximal/Minimal thresholding | Single-channel compatible | Automated thresholding; Hardware efficient | Limited validation across diverse artifacts |
Machine learning classifiers have shown promising results in detecting ocular artifacts within EEG signals. Comparative studies reveal that Artificial Neural Networks (ANN) paired with scalp topography features currently represent the most powerful feature-classifier combination for eye-blink artifact detection, outperforming other algorithmic approaches [80]. These supervised learning methods require substantial labeled training data but offer automated detection capabilities that reduce manual intervention.
Deep learning architectures have transformed EEG artifact removal approaches by learning deep-level features of both artifacts and neural signals. The CLEnet model integrates dual-scale Convolutional Neural Networks (CNN) with Long Short-Term Memory (LSTM) networks and an improved attention mechanism (EMA-1D) to extract morphological and temporal features simultaneously [19]. This approach has demonstrated superior performance in removing multiple artifact types—including EMG, EOG, and mixed artifacts—while maintaining compatibility with multi-channel EEG inputs, addressing a significant limitation of earlier DL models [19]. Unlike traditional methods, deep learning approaches can adapt to unknown artifacts, making them particularly valuable for real-world applications where artifact characteristics may be unpredictable [19].
Table 2: Performance Comparison of Artifact Removal Methods Across Experimental Setups
| Method | Average Correlation Coefficient | RMSE | SNR Improvement | Hardware Area Consumption | Power Consumption |
|---|---|---|---|---|---|
| DWT-LMM [33] | 0.9369 | 2.2252 | Not reported | 6490.07 μm² | 792.18 μW |
| CLEnet (Mixed artifacts) [19] | 0.925 | 0.300 (temporal) | 11.498 dB | Not reported | Not reported |
| ICA-Wavelet [33] | ~0.89 (estimated) | ~2.5 (estimated) | Not reported | Higher than DWT-LMM | Higher than DWT-LMM |
| Regression [81] | Moderate | Moderate | Not reported | Low | Low |
| Adaptive Filtering [33] | Moderate | Moderate | Not reported | Medium | Medium |
Objective comparison of artifact removal methods requires standardized quantitative metrics that capture different aspects of performance. The average correlation coefficient measures how well the cleaned signal preserves the original brain activity by quantifying similarity between processed and clean reference signals, with values closer to 1.0 indicating superior preservation of neural information [33] [19]. Root Mean Square Error (RMSE) evaluates the magnitude of difference between processed and reference signals, where lower values indicate more accurate artifact removal [33]. Some studies further differentiate between temporal domain (RRMSEt) and frequency domain (RRMSEf) errors to capture both dimensions of signal fidelity [19].
The signal-to-noise ratio (SNR) quantifies the improvement in signal quality after processing, with higher values indicating better separation of neural signals from artifacts [19]. For applications where hardware implementation is considered, area consumption (measured in μm² for silicon chips) and power consumption (measured in μW) become critical metrics for evaluating practical feasibility in portable devices [33]. Additionally, computational complexity directly impacts processing speed and suitability for real-time applications, making it an important consideration for brain-computer interface systems [33] [79].
Recent evidence suggests that preprocessing choices, including artifact correction methods, can significantly influence downstream analysis outcomes. A 2025 multiverse study demonstrated that artifact correction steps generally reduce decoding performance across experiments and models, likely because uncorrected artifacts may contain systematic information correlated with experimental conditions [12]. For instance, in paradigms where eye movements are systematically related to task conditions (e.g., visual field experiments), removing ocular artifacts may eliminate predictive information, thereby decreasing classification accuracy [12].
These findings highlight the critical importance of aligning artifact removal strategies with specific research goals. While removing artifacts is essential for isolating genuine neural activity, excessive correction might eliminate behaviorally relevant signals. Researchers must therefore consider whether their primary objective is maximizing signal purity for neurophysiological interpretation or optimizing decoding performance for brain-computer interface applications, as these goals may necessitate different approaches to artifact management [12].
This protocol provides a standardized procedure for semi-automatic EEG preprocessing incorporating Independent Component Analysis (ICA) and Principal Component Analysis (PCA) with step-by-step quality checking to ensure removal of large-amplitude artifacts [47]:
Step-by-Step Implementation:
Bandpass Filtering and Bad Channel Interpolation: Apply bandpass filtering (1-40 Hz recommended) to remove slow drifts and high-frequency noise. A relatively high cutoff high-pass filter (1-2 Hz) is crucial for obtaining good ICA decomposition [47]. Identify and interpolate bad channels using statistical criteria (e.g., excessive noise, flat-lining).
ICA Decomposition for Ocular Artifact Removal: Select a stationary segment of data containing ocular artifacts (eye blinks and movements) for ICA decomposition. Stationarity is critical for proper ICA performance [47]. Run ICA decomposition on this segment and visually inspect components to identify those representing ocular artifacts based on their topography (frontally dominant) and time course (correlation with eye blink events).
Apply ICA Weights and PCA Correction: Apply the computed ICA weights to the entire dataset to remove ocular artifacts. Follow with PCA-based correction to address large-amplitude, non-specific transient artifacts such as muscle vibrations and line noise that may remain after ICA processing [47].
Quality Checking and Data Export: Visually inspect the processed data to verify artifact removal while preserving neural signals of interest. Compare data before and after processing to ensure major artifacts have been adequately addressed. Export the cleaned data for subsequent analysis.
This protocol outlines the implementation procedure for deep learning-based artifact removal, specifically using architectures like CLEnet that integrate CNN and LSTM networks [19]:
Step-by-Step Implementation:
Data Preparation and Preprocessing: Utilize semi-synthetic datasets created by combining clean EEG segments with recorded artifact signals (EOG, EMG) or real EEG datasets with annotated artifacts. For multi-channel approaches, ensure proper formatting of spatial relationships between electrodes. Partition data into training, validation, and test sets maintaining temporal independence.
Model Architecture Configuration: Implement the CLEnet architecture integrating dual-scale CNN for morphological feature extraction at different spatial scales, LSTM networks for capturing temporal dependencies, and an improved EMA-1D attention mechanism to enhance feature representation [19]. Adjust hyperparameters based on specific EEG characteristics and artifact types.
Model Training and Validation: Train the model using mean squared error (MSE) or similar loss functions between cleaned and target signals. Employ appropriate regularization techniques to prevent overfitting. Monitor performance on validation sets to determine convergence and select optimal epochs for testing.
Performance Evaluation: Evaluate model performance using standardized metrics including SNR, correlation coefficient (CC), and RMSE in both temporal and frequency domains (RRMSEt, RRMSEf) [19]. Compare against traditional methods and other deep learning architectures to establish comparative performance.
Selecting the optimal ocular artifact removal method requires careful consideration of multiple constraints related to EEG setup, research objectives, and practical implementation factors. The following decision framework aligns method capabilities with specific research scenarios:
Table 3: Decision Framework for Ocular Artifact Removal Method Selection
| Research Scenario | Recommended Method | Rationale | Implementation Considerations |
|---|---|---|---|
| Single-channel EEG | DWT-LMM [33] | Specifically designed for single-channel applications; hardware-efficient | Automated decomposition level selection; Minimal parameter tuning |
| Multi-channel research-grade EEG | ICA-based approaches [47] | Established effectiveness with sufficient spatial information | Requires stationarity; Manual component inspection needed |
| Real-time BCI applications | Adaptive filtering [33] or Deep Learning [19] | Fast computation; Suitable for online processing | DL methods require pre-training; Adaptive filters need reference |
| Wearable/Portable EEG | DWT-LMM [33] or Optimized DL models [19] | Balance of performance and computational efficiency | Consider power consumption and processing latency |
| Maximizing signal purity | Combined ICA-Wavelet [33] or CLEnet [19] | Superior artifact removal and signal preservation | Increased computational demands; More complex implementation |
| Unknown artifact types | Deep Learning (CLEnet) [19] | Adaptability to various unseen artifacts | Requires diverse training data; Computational intensive training |
For researchers navigating multiple constraints simultaneously, the following integrated decision pathway provides a systematic selection process:
Assess Channel Availability: For single-channel systems, prioritize wavelet-based methods (DWT-LMM) or specialized single-channel adaptations [33]. For multi-channel setups (≥8 channels), consider ICA-based approaches that leverage spatial information [47].
Evaluate Real-time Requirements: For offline analysis, computational complexity is less critical, enabling use of more demanding methods like deep learning or combined approaches. For real-time applications, prioritize methods with low latency such as adaptive filtering or optimized wavelet techniques [33].
Characterize Artifact Properties: For well-defined ocular artifacts with characteristic topographies, ICA and regression methods are sufficient. For complex, unknown, or multiple co-occurring artifact types, deep learning approaches offer superior adaptability [19].
Consider Hardware Constraints: For portable or wearable applications with limited processing capabilities and power budgets, implement hardware-efficient methods like DWT-LMM with demonstrated low area and power consumption [33].
Align with Research Objectives: If maximizing decoding accuracy is the primary goal, carefully evaluate the impact of artifact removal on classification performance, as some artifacts may contain condition-relevant information [12]. For neurophysiological interpretation, prioritize methods that maximize signal purity and preserve genuine neural dynamics.
Successful implementation of ocular artifact removal methods requires both computational tools and appropriate experimental materials. This toolkit summarizes essential components for establishing an effective artifact processing pipeline:
Table 4: Essential Research Reagents and Solutions for EEG Artifact Research
| Tool/Category | Specific Examples | Function/Purpose | Implementation Notes |
|---|---|---|---|
| Software Platforms | EEGLAB [82], MNE-Python [82] [68], FieldTrip [82] | Comprehensive EEG analysis environments | Provide implementations of major algorithms (ICA, wavelet, etc.) |
| Reference Datasets | EEGdenoiseNet [19], ERP CORE [12] | Benchmarking and validation | Semi-synthetic datasets enable controlled performance evaluation |
| Deep Learning Frameworks | TensorFlow, PyTorch with MNE-Python integration [82] | Implementation of DL artifact removal | Pre-trained models available for some architectures |
| Hardware Solutions | Carbon-Wire Loops (CWL) [83], Portable EEG systems | Reference artifact recording; Mobile data collection | CWL provides dedicated MR artifact capture for simultaneous EEG-fMRI |
| Quality Metrics | Correlation coefficient, RMSE, SNR [33] [19] | Performance quantification | Essential for objective method comparison |
| Visualization Tools | Topographic maps, Component inspectors [82] [68] | Result interpretation and validation | Critical for manual component classification in ICA |
Selecting appropriate ocular artifact removal methods requires careful consideration of research goals, EEG setup constraints, and practical implementation factors. Traditional methods like ICA and wavelet transforms offer established approaches with understood limitations, while emerging deep learning techniques provide adaptability to complex artifacts at increased computational cost. This objective comparison framework equips researchers with standardized protocols, performance metrics, and decision guidelines to align method selection with specific research requirements. As EEG applications expand into wearable devices and real-time systems, continued development of efficient, automated artifact removal strategies will remain essential for advancing both basic neuroscience and applied brain-computer interface technologies.
The effective removal of ocular artifacts is not a one-size-fits-all endeavor but a critical, nuanced step that profoundly influences the validity of EEG findings. As this review synthesizes, the choice of method—whether regression, ICA, wavelet transforms, or emerging deep learning approaches—must be guided by the specific research context, balancing the removal of contaminants with the preservation of genuine neural signals. The move towards targeted, non-invasive cleaning strategies and robust validation frameworks is paramount. Future directions should prioritize the development of standardized, open-source pipelines that enhance reproducibility, the creation of methods effective for single-channel and mobile EEG systems for ecological monitoring, and the careful integration of these techniques into clinical trial protocols to ensure that biomarkers and drug efficacy assessments are based on clean, interpretable neural data.