This article provides a comprehensive analysis of real-time artifact removal techniques critical for reliable Human-Robot Interaction (HRI) systems.
This article provides a comprehensive analysis of real-time artifact removal techniques critical for reliable Human-Robot Interaction (HRI) systems. Targeting researchers and development professionals, we explore the foundational challenges of physiological and motion artifacts in EEG and other biosignals, review state-of-the-art methodological approaches including ICA, ASR, iCanClean, and adaptive filtering, and offer practical troubleshooting strategies for optimizing performance in ecological settings. Through comparative validation of techniques and discussion of emerging trends like deep learning and auxiliary sensors, this work synthesizes key principles for developing robust, real-time artifact removal pipelines that ensure accurate emotion recognition, intent decoding, and seamless brain-robot communication.
In human-robot interaction (HRI) research, the accurate interpretation of biological signals is paramount for developing responsive and adaptive systems. Electroencephalography (EEG) and other biosignals provide a non-invasive window into a user's cognitive and emotional state, enabling robots to respond to human intention and affect. However, these signals are invariably contaminated by artifacts—unwanted signals from non-neural sources—which can severely degrade system performance. Effective artifact removal is particularly critical in real-time HRI applications, where the fidelity of the processed signal directly impacts the robot's ability to make appropriate and timely decisions. This note defines the major categories of artifacts, summarizes removal performance data, and provides detailed experimental protocols for their mitigation within HRI research contexts.
Artifacts in biosignal recordings are broadly classified into three categories based on their origin: physiological, motion-related, and environmental. The table below delineates their sources, characteristics, and impact on signals.
Table 1: Classification and Characteristics of Common Artifacts
| Artifact Category | Specific Type | Origin | Frequency Band | Key Characteristics | Impact on HRI |
|---|---|---|---|---|---|
| Physiological | Ocular (EOG) | Eye blinks and movements [1] | Slow frequencies, below 5 Hz [2] | High amplitude, slow waves [1] | Obscures prefrontal cortex signals critical for emotion estimation [2] [3] |
| Muscle (EMG) | Muscle contractions (face, neck, jaw) [1] | Broad spectrum (20–300 Hz) [2] [1] | High-frequency, transient spikes [1] | Masks neural activity in beta/gamma bands, crucial for intention estimation [2] [4] | |
| Cardiac (ECG/PPG) | Heartbeat and pulse [1] | ~1.2 Hz (pulse) [1] | Regular, periodic waveform [1] | Can be mistaken for rhythmic brain activity; introduces periodic noise [1] | |
| Motion | Head Movement | Cable sway, electrode displacement [5] [6] | Overlaps with neural bands [5] | Baseline shifts, amplitude bursts [5] | Causes significant signal distortion during mobile HRI tasks [5] [6] |
| Gait-Related | Heel strike during walking [5] | Low frequency | Arrhythmic amplitude bursts [5] | Corrupts signals in mobile brain-imaging scenarios [5] | |
| Environmental | Power Line Noise | Electrical mains [2] [1] | 50/60 Hz and harmonics [2] | Stationary, narrowband interference [1] | Obscures neural signals in the gamma band [2] |
| Technical | Faulty electrodes, cable movement [1] | Varies | Sudden signal drops or spikes [1] | Creates non-physiological signal patterns, leading to misinterpretation [1] |
Selecting an appropriate artifact removal method requires an understanding of its performance under specific conditions. The following table synthesizes quantitative results from recent studies for easy comparison.
Table 2: Performance Comparison of Selected Artifact Removal Methods
| Method Name | Algorithm Type | Artifacts Targeted | Key Performance Metrics | Reported Performance | Suitability for Real-Time HRI |
|---|---|---|---|---|---|
| iCanClean [6] | Real-time capable filtering | Motion, Muscle, Eye, Line-noise | Data Quality Score (0-100%) [6] | Improved from 15.7% to 55.9% (all artifacts) [6] | High (validated on mobile data) [6] |
| Motion-Net [5] | Subject-specific CNN | Motion | Artifact Reduction (η): 86% ±4.13; SNR Improvement: 20 ±4.47 dB [5] | 86% ±4.13 artifact reduction [5] | Medium (subject-specific training required) [5] |
| Mutual Information (Epanechnikov) [7] | Blind Source Separation (BSS) | General (tested for emotion recognition) | Classification Accuracy [7] | 80.13% accuracy [7] | Medium (computational cost depends on implementation) [7] |
| Adaptive Filtering [4] | Modified LMS Adaptive Filter | TENS feedback artifact in sEMG | Signal-to-Noise Ratio (SNR) Improvement [4] | SNR increase of 10.3 dB [4] | High (designed for real-time prosthetic control) [4] |
| 1D-CNN with Penalty [8] | Convolutional Neural Network | Motion in fNIRS | Signal-to-Noise Ratio (SNR) Improvement [8] | SNR improvement > 11.08 dB [8] | High (processing time: 0.53 ms/sample) [8] |
This protocol is adapted from methodologies optimized for real-time emotion estimation in HRI, using a minimal set of electrodes for wearability [2] [3].
1. Objective: To acquire clean EEG signals for emotion estimation in real-time by removing ocular and motion artifacts. 2. Materials and Setup:
3. Procedure:
The following workflow diagram illustrates the real-time processing pipeline:
This protocol uses the Motion-Net deep learning model for high-fidelity removal of motion artifacts, ideal for scenarios with repetitive but subject-specific movements [5].
1. Objective: To train a subject-specific convolutional neural network (CNN) for removing motion artifacts from mobile EEG data. 2. Materials and Setup:
3. Procedure:
The workflow for this subject-specific approach is outlined below:
Table 3: Essential Materials and Tools for Artifact Removal Research
| Item Name | Function/Application | Specific Example/Note |
|---|---|---|
| Mobile EEG System with Active Electrodes | Records brain activity with reduced susceptibility to motion artifacts compared to passive electrodes. | Essential for any HRI study involving movement [6]. |
| Electrooculogram (EOG) Electrodes | Records eye movements and blinks to serve as a reference signal for regression-based or adaptive filtering methods. | Placed above, below, and to the side of the eyes [1]. |
| Inertial Measurement Unit (IMU) | Measures head acceleration and movement, providing a reference signal for motion artifact removal algorithms. | Used in Motion-Net and adaptive filtering approaches [5]. |
| Transcutaneous Electrical Nerve Stimulation (TENS) Unit | Provides sensory feedback in prosthetic HRI studies; also a source of known artifact for testing removal algorithms. | Artifacts can be removed using adaptive filters as shown in prosthetic hand research [4]. |
| iCanClean Algorithm | A real-time capable, generalized framework for removing multiple artifact sources without separate noise sensors. | Outperformed ASR, Auto-CCA, and Adaptive Filtering in phantom head tests [6]. |
| Motion-Net Deep Learning Model | A subject-specific CNN for high-accuracy motion artifact removal when sufficient per-subject training data is available. | Effective with smaller datasets when using Visibility Graph features [5]. |
| Visibility Graph (VG) Feature Extraction | A method to convert time-series signals into graph structures, improving deep learning model performance on smaller datasets. | Used to enhance the accuracy of Motion-Net [5]. |
In human-robot interaction (HRI), real-time processing is a fundamental requirement rather than a luxury. Delays in robotic decision-making directly compromise system responsiveness, safety, and user trust, particularly in dynamic service contexts where computational efficiency is critical [9]. The core challenge lies in designing intelligent computing architectures that can rapidly process multimodal inputs—from queries to biological signals—while maintaining high accuracy and low failure rates [9].
This document provides application notes and experimental protocols for implementing real-time processing systems in HRI research, with particular emphasis on artifact removal techniques for brain-computer interfaces and query-processing frameworks. The quantitative data and methodologies presented serve as essential references for researchers developing next-generation interactive robotic systems.
Table 1: Performance Comparison of HRI Computing Models [9]
| Performance Metric | HICM Framework | CDS Model | DGTA Model | CCS Model |
|---|---|---|---|---|
| Calculation Time Reduction | 8.67% improvement | Baseline | Not specified | Not specified |
| Service Time Reduction | 15.09% improvement | Baseline | Not specified | Not specified |
| Failure Rate Reduction | 7.87% improvement | Baseline | Not specified | Not specified |
| Success Factor | 11.8% higher | Baseline | Not specified | Not specified |
| Matching Ratio | 14.88% higher | Baseline | Not specified | Not specified |
| Failure Rates | 6.22% lower | Baseline | Not specified | Not specified |
Table 2: Real-Time EEG Processing Specifications for HRI [3] [10]
| Processing Parameter | Target Specification | Application Context |
|---|---|---|
| Temporal Resolution | ≤ 1 ms | EEG signal acquisition |
| Artifact Removal | Real-time EOG/blink removal | Prefrontal cortex signals |
| Frequency Bands | Delta (0.1-3.5 Hz), Theta (4-7.5 Hz), Alpha (8-13 Hz), Beta (14-30 Hz), Gamma (30-50 Hz) | Emotion estimation, intentional command detection |
| Electrode Placement | 10-20 system, emphasis on AF3, T7, TP7, P7, AF4, T8, TP8, P8 | Emotion estimation, eye artifact detection |
| Key Electrode Positions | F7, F8 channels for lateral eye movement detection | Phase difference analysis for eye movement |
Objective: Implement and validate the Hybrid Intelligent Computing Model (HICM) for robotic query processing and service provision [9].
Materials:
Procedure:
Performance Benchmarking
Validation Testing
Data Analysis:
Objective: Implement optimized real-time electro-oculographic (EOG) artifact removal and emotion estimation for human-robot interaction applications [3].
Materials:
Procedure:
Real-Time Artifact Removal
Feature Extraction and Emotion Classification
System Integration and Validation
Data Analysis:
Objective: Develop and validate a BCI for robot control using eye artifacts for users with neurodegenerative disorders [10].
Materials:
Procedure:
Command Structure Implementation
Real-Time System Integration
Validation Methodology
Data Analysis:
Table 3: Essential Research Materials for Real-Time HRI Systems
| Category | Specific Solution | Function in HRI Research |
|---|---|---|
| Computing Framework | Hybrid Intelligent Computing Model (HICM) | Reduces calculation time by 8.67% and service time by 15.09% through combined annealing and Tabu Search optimization [9] |
| EEG Acquisition System | TMSi SAGA 64+ (64+ channels) | Provides high-temporal resolution (≤1ms) EEG acquisition with 10-20 electrode placement for non-invasive brain activity monitoring [10] |
| Artifact Removal Algorithm | Dual-threshold blink detection with phase analysis | Enables real-time EOG artifact removal by combining characteristic shape recognition (blinks) and opposite-phase signals at F7/F8 (lateral movements) [10] |
| Signal Processing Toolbox | Real-time ICA with wavelet analysis | Removes electro-oculographic artifacts while preserving valuable EEG information through component separation and reconstruction [3] |
| Emotion Estimation Model | Multi-feature classification with smoothing | Combines stable features from beta/gamma bands (16-50 Hz) with feature smoothing (LDS or Savitzky-Golay) for real-time emotion classification [3] |
| Robot Control Interface | Eye artifact-based BCI command system | Translates detected eye artifacts (blinks, lateral movements) into robot control commands through virtual timestamps and state machines [10] |
| Performance Validation Suite | Standardized query sets and task metrics | Enables comparative performance analysis across computational frameworks using calculation time, service time, and failure rates [9] |
Artifacts—unwanted signals originating from non-neural or non-behavioral sources—represent a significant challenge in human-robot interaction (HRI) research. These intrusive signals can severely degrade the performance of systems designed for emotion estimation and intent decoding, ultimately undermining the naturalness and effectiveness of human-robot collaboration. The pursuit of real-time HRI necessitates robust artifact removal pipelines that can operate under strict temporal constraints without compromising signal integrity. This application note synthesizes current research and provides detailed protocols for addressing artifact-related challenges in key HRI applications, with particular emphasis on neuroscientific and multimodal interaction contexts.
The tables below summarize empirical findings on the effects of artifacts and the performance of various mitigation strategies across different HRI modalities.
Table 1: Impact of Metal Artifacts on CT Image Quality and BMD Measurement Accuracy (Adapted from [11])
| Reconstruction Method | Mean Attenuation (HU) | Signal-to-Noise Ratio (SNR) | Artifact Index (AI) | Bone Mineral Density (BMD) Accuracy |
|---|---|---|---|---|
| Conventional Imaging (CI) | 583.6 | Highest | Highest | Low |
| O-MAR | Significantly Reduced | Moderate | Reduced | High (Comparable to CI*) |
| VMI (200 keV) | Significantly Reduced | Lowest | Lowest | Low |
| VMI + O-MAR | Significantly Reduced | Lowest | Lowest | Moderate |
| CI* (Gold Standard) | — | — | — | Reference |
Table 2: Emotion and Gesture Decoding Accuracy in Multimodal HRI (Data from [12])
| Modality | Model Used | Classification Target | Accuracy |
|---|---|---|---|
| Touch + Sound | Support Vector Machine (SVM) | 10 Emotions | 40.0% |
| Touch Only | Not Specified | 10 Emotions | Lower than multimodal |
| Sound Only | Not Specified | 10 Emotions | Lower than multimodal |
| Touch + Sound | CNN-LSTM | 6 Social Touch Gestures | 90.74% |
Table 3: Key Artifact Types and Their Characteristics in HRI Research
| Artifact Type | Primary Source | Impacted HRI Modality | Common Mitigation Strategies |
|---|---|---|---|
| Electro-oculographic (EOG) | Eye movements, blinking | EEG-based Emotion Estimation | Real-time ICA, Wavelet analysis [3] |
| Metal Artifacts | Orthopedic implants, screws | CT-based Bone Density Assessment | O-MAR algorithm, Virtual Monoenergetic Imaging (VMI) [11] |
| Muscle Artifacts | Head/body movements | EEG-based Emotion Estimation | Frequency filtering (e.g., 1-50 Hz band) [3] |
| Background Noise | Electrical interference | EEG, Tactile Sensing | Notch filters (e.g., 50Hz) [3] |
Objective: To reliably decode human emotions and social gestures through tactile and auditory signals in HRI [12].
Materials:
Procedure:
Objective: To remove artifacts from EEG signals in real-time for robust emotion estimation in affective HRI [3].
Materials:
Procedure:
Objective: To improve image quality and accuracy of Bone Mineral Density (BMD) measurements in QCT scans affected by metal implant artifacts [11].
Materials:
Procedure:
HRI Artifact Management Workflow
EEG Artifact Removal Pipeline
Table 4: Key Research Reagent Solutions for HRI Artifact Management Studies
| Item | Function/Application | Example/Specification |
|---|---|---|
| Dual-Layer Spectral Detector CT | Enables Virtual Monoenergetic Imaging (VMI) for metal artifact reduction in peri-implant bone assessment [11]. | Philips IQon system |
| O-MAR Algorithm | Projection-based correction algorithm that segments and "inpaints" metal-corrupted sinogram data to reduce artifacts in CT [11]. | Philips Healthcare implementation |
| Piezoresistive Pressure Sensor | Custom tactile sensor for measuring pressure, location, and duration of touch gestures in social HRI [12]. | Custom-built for robot integration |
| Wearable EEG System | For acquiring neural signals for emotion estimation; requires specific electrode placement for optimal results [3]. | Minimum 8 electrodes (AF3, AF4, T7, T8, TP7, TP8, P7, P8) |
| Real-Time ICA Software | Independent Component Analysis for blind source separation and removal of EOG artifacts from EEG streams [3]. | Adaptive Mixture ICA (AMICA) |
| Support Vector Machine (SVM) | A lightweight, effective classifier for real-time emotion estimation from EEG or multimodal features [12] [3]. | Linear or RBF kernel |
| CNN-LSTM Model | Hybrid deep learning architecture for temporal sequence classification, effective for social touch gesture recognition [12]. | Used for gesture classification from tactile-audio data |
Wearable electroencephalography (EEG) is revolutionizing neurotechnology by enabling brain monitoring in real-world environments, from clinical trials to human-robot interaction (HRI) research. However, this transition from controlled laboratory settings to dynamic ecological environments introduces two persistent technical challenges: the use of dry electrodes and the management of motion artifacts. Dry electrodes, while offering superior portability and ease of use, often compromise signal quality due to higher and more unstable electrode-skin impedance. Simultaneously, motion artifacts—induced by subject movement, environmental noise, and the inherent limitations of mobile acquisition—corrupt the neurological signal of interest, complicating real-time analysis. For HRI applications, where low-latency, reliable brain signal decoding is crucial for seamless and safe interaction, overcoming these challenges is paramount. This document provides application notes and experimental protocols to address these specific issues, framed within the context of real-time artifact removal for HRI research.
Recent benchmarking studies provide critical quantitative data on the performance of dry-electrode EEG systems compared to standard wet EEG in a clinical trial context. The following tables summarize key findings on system burden and signal quality.
Table 1: Comparison of Operational Burden between Dry-Electrode and Standard EEG Systems
| Metric | Standard EEG (Wet) | Dry-EEG Device A (DSI-24) | Dry-EEG Device B (Quick-20r) | Dry-EEG Device C (zEEG) |
|---|---|---|---|---|
| Median Set-up Time | Benchmark (Longest) | ~50% faster than Standard EEG [13] | Significantly faster [13] | Significantly faster [13] |
| Median Clean-up Time | Benchmark (Longest) | Significantly faster [13] | Significantly faster [13] | Significantly faster [13] |
| Technician Ease of Set-up (0-10) | 7 | 9 [13] | 7 [13] | 7 [13] |
| Technician Ease of Clean-up (0-10) | 5 | 9 [13] | 9 [13] | 9 [13] |
| Participant Comfort | Highest (Benchmark) | Matched or lower than Standard EEG [13] | Matched or lower than Standard EEG [13] | Matched or lower than Standard EEG [13] |
Table 2: Signal Quality Performance of Dry-Electrode EEG Across Functional Domains
| EEG Application / Signal Aspect | Dry-EEG Performance vs. Standard EEG | Notes and Challenges |
|---|---|---|
| Resting-State Quantitative EEG | Adequately captured [13] | Suitable for power spectrum analysis. |
| P300 Evoked Potential | Adequately captured [13] | Reliable for event-related potential studies. |
| Low-Frequency Activity (< 6 Hz) | Notable challenges [13] | Prone to contamination from motion and drift. |
| Induced Gamma Activity (40-80 Hz) | Notable challenges [13] | Susceptible to muscle artifact contamination. |
| Action Anticipation (MRCP/BP) | Feasible with advanced processing [14] [15] | Early BP (~1.5s before movement) is low amplitude and difficult to detect. |
| Individual Finger Movement Decoding | Feasible with deep learning [15] | Achieved ~80% accuracy for binary classification in real-time BCI [15]. |
Objective: To quantitatively evaluate the performance of a dry-electrode EEG device against a standard wet-EEG system in a context relevant to HRI research.
Materials:
Procedure:
Objective: To implement and validate a pipeline for the real-time detection and removal of motion and ocular artifacts during a dynamic human-robot interaction task.
Materials:
Procedure:
Table 3: Essential Research Tools for Wearable EEG and HRI Experiments
| Item | Function/Application | Examples & Notes |
|---|---|---|
| Dry-Electrode EEG Systems | Mobile EEG acquisition for real-world HRI. | CGX Quick-20r, Wearable Sensing DSI-24, Zeto zEEG [13]. Vary in set-up speed, comfort, and signal performance. |
| Standard Wet-EEG System | Gold-standard benchmark for validation studies. | Systems with QuikCap Neo Net and Grael amplifier [13]. Essential for controlled comparison of dry-EEG performance. |
| Inertial Measurement Units (IMUs) | Monitoring head and body movement for motion artifact detection. | Can be integrated or external. Provides objective data to correlate with motion artifacts in EEG [16]. |
| Robotic Platforms | End-effector for HRI tasks and BCI control. | Collaborative robot arms (e.g., UR3e), dexterous robotic hands [15]. Platform choice depends on the HRI paradigm. |
| Real-Time Processing Software | Platform for implementing artifact removal pipelines. | MATLAB/Simulink, LabStreamingLayer (LSL), Python with MNE-real-time [3] [15]. |
| Public EEG Datasets | Algorithm development and benchmarking. | SEED (for emotion) [3], MI-based BCI datasets [14]. Crucial for training machine learning models like artifact classifiers. |
| Deep Learning Frameworks | Decoding motor commands and improving artifact removal. | EEGNet [15], Convolutional Neural Networks (CNNs) [14] [18]. Enable end-to-end decoding of complex intentions (e.g., individual finger movements). |
Blind Source Separation (BSS) represents a cornerstone of modern signal processing, enabling the recovery of underlying source signals from their observed mixtures without prior knowledge of the mixing process. In the context of real-time artifact removal for human-robot interaction (HRI) research, BSS techniques are indispensable for processing electrophysiological signals, particularly electroencephalography (EEG). HRI environments present unique challenges for signal acquisition, where artifacts originating from muscle activity, eye movements, cable swings, and magnetic induction significantly compromise signal integrity [19]. These artifacts must be effectively separated and removed to ensure accurate interpretation of neural signals for brain-computer interface (BCI) applications. Among the diverse BSS algorithms, Independent Component Analysis (ICA) and the Second-Order Blind Identification (SOBI) algorithm have emerged as fundamental workhorses, each offering distinct advantages for specific HRI scenarios.
ICA operates on the principle of maximizing the statistical independence of component signals, typically by minimizing mutual information or maximizing non-Gaussianity [20] [21]. The core mathematical model assumes that observed signals ( \mathbf{X} ) are linear mixtures of statistically independent sources ( \mathbf{S} ), related via a mixing matrix ( \mathbf{A} ) (( \mathbf{X} = \mathbf{A} \mathbf{S} )). The objective is to estimate a separating matrix ( \mathbf{W} ) that recovers the original sources (( \mathbf{S} = \mathbf{W} \mathbf{X} )) [20]. In practical terms, ICA effectively solves the "cocktail party problem"—separating individual voices from recorded mixtures—making it exceptionally suitable for isolating neural signals from artifact-contaminated EEG data [21]. Its strength lies in identifying and removing artifacts embedded within the data without discording valuable neurological information, thereby preserving the continuity of brain signals essential for real-time HRI systems [22].
SOBI, in contrast, leverages a different statistical property by exploiting the time structure of sources, specifically their autocorrelation functions [22]. It operates under the assumption that source signals are temporally correlated but mutually uncorrelated with different time lags. By performing joint diagonalization of several covariance matrices computed at different time lags, SOBI can separate sources with distinct spectral characteristics. This methodological approach proves particularly effective for separating artifacts with pronounced periodic components, such as power line interference, channel noise, and certain movement artifacts common in dynamic HRI environments.
Table 1: Core Algorithmic Characteristics of ICA and SOBI
| Feature | Independent Component Analysis (ICA) | Second-Order Blind Identification (SOBI) |
|---|---|---|
| Statistical Principle | Maximizes non-Gaussianity/statistical independence | Exploits temporal structure & autocorrelation |
| Separation Basis | Higher-order statistics | Second-order statistics (covariance matrices) |
| Optimal Use Cases | Artifact removal (eye blinks, muscle activity), source localization [20] [22] | Periodic noise, channel-specific artifacts, colored noise environments [22] |
| Computational Load | Moderate to High | Generally Lower |
| Real-Time Suitability | Moderate (requires adaptations like ORICA) [23] | High |
The implementation of ICA for artifact removal follows a systematic pipeline comprising data preparation, decomposition, component identification, and signal reconstruction. A critical preprocessing step for most ICA algorithms is whitening (or sphering), which removes correlations between channels and normalizes their variances, effectively transforming the data into an uncorrelated unit-variance space [20]. Geometrically, this process restores the data's spherical structure, simplifying the subsequent ICA step to identifying an appropriate rotation of the whitened data. Whitening not only standardizes the data but also dramatically improves the convergence speed and stability of ICA algorithms.
Following whitening, the core ICA algorithm estimates the separating matrix that maximizes the independence of the resulting components. Multiple ICA variants exist, including Infomax ICA, FastICA, and extended algorithms capable of identifying both sub-Gaussian and super-Gaussian sources [22] [21]. The Infomax algorithm, for instance, implemented in tools like EEGLAB's runica, employs a natural gradient approach to minimize the mutual information between output channels, effectively maximizing their statistical independence [22]. During decomposition, the algorithm iteratively adjusts the separating matrix until a convergence criterion is met, producing a set of independent components along with their corresponding time courses and spatial topographies.
Table 2: ICA Algorithms and Their Applications in HRI Research
| Algorithm | Key Mechanism | Advantages for HRI |
|---|---|---|
| Infomax ICA | Maximizes information transfer (mutual information minimization) via gradient descent | Default in EEGLAB; effective for standard EEG artifact separation [22] |
| FastICA | Fixed-point iteration to maximize negentropy (non-Gaussianity) | Faster convergence; computationally efficient [21] |
| SOBI | Joint diagonalization of covariance matrices at multiple time lags | Effective for correlated noise; robust to certain motion artifacts [22] |
| ORICA (Online ICA) | Recursive, sample-by-sample update of unmixing matrix | Enabled for real-time processing in dynamic HRI settings [23] |
The subsequent workflow for artifact removal involves meticulous analysis of the derived components. Researchers must identify which components correspond to neural activity and which represent artifacts. This classification relies on evaluating multiple characteristics: (1) spatial topography—artifact components often exhibit characteristic scalp distributions (e.g., frontal focus for ocular artifacts, periocular or neck muscle patterns for EMG); (2) temporal dynamics—artifact components typically show time courses reflecting their non-neural origin (e.g., pulse-like patterns for eye blinks, high-frequency bursts for muscle activity); and (3) spectral properties—artifact components often display distinctive power spectral densities (e.g., slow drifts for movement artifacts, high-frequency content for muscle noise) [22]. After identifying artifactual components, they can be subtracted from the original signal by projecting only the neural components back to the sensor space, resulting in cleaned data suitable for subsequent analysis in HRI systems.
Successful implementation of ICA and SOBI for HRI research requires both specialized software tools and appropriate hardware configurations. The computational demands of these algorithms, particularly for real-time applications, necessitate a robust processing environment.
Table 3: Essential Software Tools for BSS Implementation
| Tool/Resource | Function | Implementation Platform |
|---|---|---|
| EEGLAB | Interactive environment for ICA; includes runica, FastICA, SOBI [22] | MATLAB |
| MNE-Python | Open-source Python package for EEG/MEG analysis | Python |
| ORICA Plugin | Enables real-time, recursive ICA for online processing [23] | EEGLAB/MATLAB |
| FastICA Package | Efficient implementation of FastICA algorithm | R, Python, MATLAB [21] |
| SpyICA Toolbox | Contains Python implementation of ORICA | Python [23] |
For experimental data acquisition in HRI contexts, wearable EEG systems with dry or semi-dry electrodes are typically employed due to their practicality and rapid setup [16]. However, these systems present specific challenges for BSS, including a reduced number of channels (typically below 16), which can impair the efficacy of source separation methods like ICA and SOBI [16] [19]. Successful decomposition generally requires the number of sensors to equal or exceed the number of independent sources to be identified; therefore, low-density systems may struggle to separate neural signals from multiple concurrent artifacts. Furthermore, artifacts in wearable EEG exhibit specific features due to dry electrodes, reduced scalp coverage, and subject mobility, necessitating tailored processing pipelines that explicitly address these peculiarities [16].
Step 1: Data Preparation and Preprocessing Acquire EEG data using a wearable system with appropriate channel configuration (≥16 channels recommended for effective decomposition). Apply a high-pass filter (e.g., 1 Hz cutoff) to remove slow drifts and a low-pass filter (e.g., 50-60 Hz) to reduce high-frequency noise. Identify and interpolate severely noisy channels. For ICA, it is crucial to use continuous, unfiltered (or minimally filtered) data, as aggressive filtering can alter the statistical properties essential for successful source separation [22].
Step 2: Execute ICA Decomposition Select an appropriate ICA algorithm (e.g., Infomax for general use, FastICA for speed, ORICA for real-time applications). The data matrix ( \mathbf{X} ) (channels × time points) is decomposed such that ( \mathbf{S} = \mathbf{W} \mathbf{X} ), where ( \mathbf{W} ) is the unmixing matrix and ( \mathbf{S} ) contains the independent components. Ensure sufficient data length is available for stable decomposition; as a rule of thumb, the number of data points (time samples) should be at least the square of the number of channels [22].
Step 3: Component Identification and Classification
Visualize component properties using tools like EEGLAB's iclabel or similar automated classifiers. Inspect:
Step 4: Artifact Removal and Signal Reconstruction After identifying artifactual components (e.g., components 1, 2, and 4 in a hypothetical decomposition), reconstruct the cleaned data by projecting all components except the artifactual ones back to the sensor space. This is mathematically achieved by creating a modified component matrix ( \mathbf{S}{clean} ) where the rows corresponding to artifacts are set to zero, and then computing ( \mathbf{X}{clean} = \mathbf{W}^{-1} \mathbf{S}_{clean} ), where ( \mathbf{W}^{-1} ) is the inverse of the unmixing matrix (equivalent to the mixing matrix ( \mathbf{A} )) [20].
Validating the efficacy of artifact removal is crucial for ensuring signal quality in HRI applications. The following metrics are commonly employed:
The transition from offline analysis to real-time artifact removal presents significant challenges and opportunities for HRI research. Traditional ICA algorithms are computationally intensive and typically operate on complete datasets, making them unsuitable for real-time applications where low latency is critical. However, recent advancements have enabled online ICA implementations that update the decomposition recursively as new data arrives.
The Online Recursive Independent Component Analysis (ORICA) algorithm represents a breakthrough in this domain, enabling real-time source separation suitable for dynamic HRI environments [23]. ORICA employs a recursive update rule for the unmixing matrix, allowing it to adapt to non-stationary signal statistics—a common characteristic in EEG data during human-robot interaction. Implementation options include the original ORICA plugin for EEGLAB and Python implementations within the SpyICA toolbox [23].
For real-time BCI applications in HRI, a typical processing pipeline integrates ORICA with other processing modules:
This integrated approach enables robust BCI performance even in the presence of motion artifacts and other noise sources inherent to interactive environments. Studies have demonstrated that such pipelines can effectively handle artifacts arising from natural movements during HRI tasks, including walking, gesturing, and collaborative manipulation activities [19].
While both ICA and SOBI offer powerful blind source separation capabilities, their relative performance varies across different HRI scenarios. ICA generally excels at separating temporally independent sources with non-Gaussian distributions, making it particularly effective for ocular artifacts, muscle activity, and cardiac interference. SOBI's strength lies in separating sources with distinct temporal autocorrelation profiles, offering advantages for removing periodic noise and certain motion artifacts.
Future developments in BSS for HRI research are likely to focus on several key areas:
The integration of advanced BSS techniques like ICA and SOBI into the HRI research pipeline represents a critical enabling technology for developing robust, real-world brain-computer interfaces. By effectively separating neural signals from contamination sources, these methods pave the way for more natural and reliable human-robot collaboration across diverse application domains, from rehabilitation robotics to industrial human-robot teams.
The advancement of Human-Robot Interaction (HRI) relies heavily on the accurate interpretation of user states through physiological signals like Electroencephalography (EEG). However, these signals are frequently corrupted by artifacts—unwanted noise from muscular activity, eye movements, or motion—which can severely degrade system performance. Real-time artifact removal is, therefore, not merely a preprocessing step but a critical component for enabling robust and reliable HRI systems. Techniques such as Common Average Reference (CAR), Localized Regression Removal (LRR), and Adaptive Filtering have emerged as powerful solutions for this challenge, each offering distinct mechanisms for isolating and removing noise while preserving neural information. Their successful implementation allows for more accurate brain-computer interfaces, adaptive automation, and closed-loop robotic systems that can respond to user intentions with high fidelity [3] [25] [26].
The necessity for these techniques is particularly pronounced in real-world HRI applications. For instance, in robotic-assisted surgery, a surgeon's high mental workload must be accurately detected via EEG to trigger adaptive assistance, a process compromised by motion and physiological artifacts [25]. Similarly, providing sensory feedback in prosthetic hands using Transcutaneous Electrical Nerve Stimulation (TENS) introduces substantial artifacts into surface Electromyography (sEMG) signals used for control, necessitating advanced filtering for closed-loop operation [27]. This document details the application and protocols for CAR, LRR, and Adaptive Filtering, providing a structured framework for their implementation in HRI research.
CAR is a spatial filtering technique that operates on the principle of re-referencing. It assumes that artifacts are common to all electrodes while neural signals are local. By subtracting the average signal of all electrodes from each individual channel, CAR effectively suppresses widespread noise.
N channels, the CAR-transformed signal Y_i(t) for channel i is given by:
Y_i(t) = X_i(t) - (1/N) * Σ_{j=1}^{N} X_j(t)
where X_i(t) is the original signal of the i-th channel at time t.LRR is an advanced version of regression-based artifact removal that targets local spatial correlations. Instead of using a global reference like CAR, LRR estimates and removes artifacts based on the signals from a local subset of electrodes, often surrounding the target channel. This makes it more effective for artifacts with a localized topography, such as ocular artifacts.
Adaptive filters are a class of algorithms that dynamically adjust their parameters to track a non-stationary signal or noise statistics. This is ideal for HRI, where the user and environment are in constant flux. The core component is an adaptive algorithm, such as the Least-Mean-Squares (LMS), that minimizes the error between the filter's output and a desired signal.
d(n) = s(n) + v(n)) and a reference input (x(n)) that is correlated with the artifact v(n). The filter adjusts its weights w to produce an output y(n) that best estimates v(n). This estimate is then subtracted from the primary input to recover the clean signal s(n) [29] [27].Table 1: Comparison of Core Artifact Removal Techniques
| Technique | Underlying Principle | Primary Strength | Best Suited for Artifact Type | Computational Load |
|---|---|---|---|---|
| Common Average Reference (CAR) | Spatial re-referencing | Simplicity and speed | Global, common-mode noise | Low |
| Localized Regression (LRR) | Local linear regression | Effectiveness on localized artifacts | Ocular, localized muscle artifacts | Medium |
| Adaptive Filtering | Dynamic parameter adjustment | Handles non-stationary noise | Motion, TENS, and physiological artifacts | Medium to High |
This protocol outlines the procedure for implementing an adaptive filter to remove TENS artifacts from sEMG signals, enabling simultaneous sensory feedback and precise control of prosthetic hands [27].
w(n+1) = w(n) + μ * sign(e(n)) * sign(x(n)), where μ is the step size.
Real-Time Adaptive Filtering for Prosthetic Control
This protocol describes a framework for using EEG-based mental workload (MWL) assessment to trigger adaptive automation in robotic-assisted surgery (RAS), requiring robust, real-time artifact removal [25].
Table 2: Performance Metrics from Cited HRI Studies
| Experiment / Study | Primary Metric | Reported Performance | Impact on HRI Task |
|---|---|---|---|
| TENS Artifact Removal [27] | Signal-to-Noise Ratio (SNR) Increase | +10.3 dB average | Restored prosthetic control performance to no-TENS levels |
| MWL-AA for Surgery [25] | High MWL Prediction Accuracy | 77.9% | Reduced surgeon workload and improved task performance |
| Emotion Recognition [7] | Classification Accuracy | 80.13% with Epanechnikov kernel | Enhanced emotion estimation for responsive HRI |
| Motor Imagery [28] | Four-Class Classification Accuracy | 98.575% mean accuracy | High-precision control of assistive devices |
Table 3: Essential Materials and Tools for HRI Artifact Removal Research
| Item Name | Specification / Example | Function in Research |
|---|---|---|
| Wearable EEG System | g.Nautilus (32-channel), dry-electrode systems | Acquires neural data in real-time from human users in interactive scenarios [25] [31]. |
| Auxiliary Biosensor Kit | Tobii Pro Glasses 2.0 (eye-tracker), Inertial Measurement Unit (IMU) | Provides reference signals (e.g., gaze, acceleration) for adaptive filtering of ocular and motion artifacts [25] [30]. |
| Stimulation System | Transcutaneous Electrical Nerve Stimulation (TENS) device | Generates sensory feedback in prosthetic applications, also serving as a known source of artifact for validation [27]. |
| Computing Platform | Laptop/PC with real-time operating system (e.g., Ubuntu with ROS) | Runs artifact removal algorithms and the primary HRI task with strict timing constraints. |
| Software Library | MATLAB (Signal Processing Toolbox), Python (MNE, SciPy, PyTorch) | Provides implementations of standard filters, ICA, LMS, and deep learning models for signal processing [29] [7]. |
| Robotic Platform | Da Vinci Research Kit (dVRK), prosthetic hand, humanoid robot | Serves as the interactive endpoint for validating the artifact removal pipeline in a realistic HRI loop [25] [27]. |
The following diagram illustrates a consolidated, advanced workflow for real-time artifact removal in an HRI setting, integrating multiple techniques discussed in this document.
Integrated Real-Time Artifact Removal Workflow
Real-time artifact removal is a critical enabling technology for human-robot interaction (HRI) research, allowing for the study of brain dynamics during natural, whole-body movement. Electroencephalography (EEG) provides the temporal resolution necessary to investigate cortical processes during dynamic HRI tasks, but its signal quality is severely compromised by motion, muscle, and other artifacts. Artifact Subspace Reconstruction (ASR) and iCanClean represent two advanced, real-time capable pipelines that address this challenge through fundamentally different approaches. This application note provides a detailed technical overview of both methods, including quantitative performance comparisons, experimental protocols, and implementation guidelines tailored for HRI research settings.
ASR is an automated, online, component-based artifact removal method for removing transient or large-amplitude artifacts in multi-channel EEG recordings without requiring reference noise signals [32] [33]. Instead, it relies on clean calibration data to establish a baseline brain state and operates as a sliding-window channel interpolation algorithm that identifies contaminated channels and reconstructs them using uncontaminated channels via principal component analysis [33].
iCanClean is a novel cleaning algorithm that uses reference noise recordings (e.g., from IMUs or dedicated noise sensors) to remove noisy EEG subspaces through canonical correlation analysis (CCA) [34] [35]. It functions as a generalized framework for removing multiple artifact sources in real time without requiring clean calibration data or risking accidental removal of brain activity [34].
Table 1: Quantitative Performance Comparison of Artifact Removal Methods
| Method | Data Quality Score (All Artifacts) | Required Reference | Computational Efficiency | Key Artifacts Addressed |
|---|---|---|---|---|
| No Cleaning | 15.7% [34] | None | N/A | N/A |
| ASR | 27.6% [34] | Clean calibration data [33] | High [33] | Motion, muscle, eye [34] |
| Auto-CCA | 27.2% [34] | None | High [34] | Muscle, line noise [34] |
| Adaptive Filtering | 32.9% [34] | Reference noise signals [34] | Medium [34] | Eye artifacts, motion (with modifications) [34] |
| iCanClean | 55.9% [34] | Reference noise signals [34] | High [34] | Motion, muscle, eye, line-noise [34] |
Table 2: iCanClean Parameter Optimization Findings
| Parameter | Optimal Value | Effect of Deviation | Application Context |
|---|---|---|---|
| Window Length | 4 seconds [36] | Shorter windows may miss artifacts; longer windows reduce adaptability | Mobile EEG during walking [36] |
| Cleaning Aggressiveness (r²) | 0.65 [36] | Lower values preserve brain activity; higher values remove more artifacts | General mobile settings [36] |
| Noise Channels | 16-64 channels [36] | Performance gradually decreases with fewer channels | High-density EEG systems [36] |
Application Context: Long-duration EEG recordings with minimal experimenter intervention, such as all-night sleep studies [33].
Required Equipment:
Procedure:
Application Context: Real-time artifact removal during human-robot interaction tasks involving walking, reaching, or other whole-body movements [34] [36].
Required Equipment:
Procedure:
Diagram 1: Comparative Workflows of ASR and iCanClean
Table 3: Essential Research Materials and Tools
| Item | Function | Example Implementation |
|---|---|---|
| High-Density EEG System | Record electrocortical activity with sufficient spatial resolution | 120+ electrode dual-layer systems [36] |
| Inertial Measurement Units (IMUs) | Capture head motion dynamics for reference-based artifact removal | 9-axis IMUs with accelerometer, gyroscope, magnetometer [35] |
| Reference Noise Sensors | Provide dedicated noise recordings for subspace identification | Integrated noise sensors in EEG cap design [34] |
| Artifact Subspace Reconstruction (ASR) | Remove transient artifacts without reference signals | Python (ASRpy) or EEGLAB implementations [32] [33] |
| iCanClean Algorithm | Remove multiple artifact types using reference noise recordings | Custom implementation based on canonical correlation analysis [34] |
| Synchronization System | Align temporal data across multiple sensor modalities | Lab Streaming Layer (LSL) or hardware triggers [35] |
| Mobile Phantom Head | Validate artifact removal performance with ground truth | Electrically conductive phantom with embedded sources [34] |
For human-robot interaction studies where ecological validity and real-time processing are paramount, iCanClean offers significant advantages by effectively handling the complex artifact profiles encountered during whole-body movement and interaction with robotic systems. The method's ability to remove multiple artifact types simultaneously while preserving brain activity makes it particularly suitable for HRI paradigms that involve walking, reaching, or other dynamic movements [34].
When implementing real-time artifact removal for closed-loop HRI systems, consider the following guidelines:
For researchers working with existing stationary datasets or with limited access to reference sensors, ASR provides a powerful alternative that can significantly improve data quality without requiring additional hardware modifications [33]. The choice between these approaches should be guided by specific research questions, available equipment, and the degree of ecological validity required in the HRI paradigm.
The advancement of real-time Human-Robot Interaction (HRI) is critically dependent on the accurate interpretation of user states, such as intention, emotion, and cognitive load. Electroencephalography (EEG) provides a non-invasive, high-temporal-resolution method for monitoring these states. However, EEG signals are persistently contaminated by various artifacts—unwanted signals from non-neural sources—which can severely degrade HRI system performance. These artifacts include those from eye movements (EOG), muscle activity (EMG), cardiac activity (ECG), and environmental noise [37] [2] [38]. Effective artifact removal is therefore a essential preprocessing step to ensure the reliability of subsequent brain signal analysis and interpretation in dynamic HRI applications, such as adaptive robot control and implicit communication [26] [39].
This document outlines emerging deep learning (DL) and hybrid optimization approaches for real-time artifact removal, framed within the context of HRI research. The focus is on providing applicable notes and detailed, reproducible protocols for researchers and scientists. These methodologies are designed to overcome the limitations of traditional techniques like Independent Component Analysis (ICA) and regression, particularly their often-inadequate performance in dynamic, real-world HRI scenarios where computational efficiency and high accuracy are paramount [40] [38].
The table below summarizes the key performance metrics of several state-of-the-art artifact removal methods as reported in recent literature. These quantitative results provide a benchmark for comparing the efficacy of different approaches.
Table 1: Performance Metrics of Advanced Artifact Removal Methods
| Methodology | Reported SNR (dB) | Reported Accuracy (%) | Key Metric 1 (Value) | Key Metric 2 (Value) | Primary Application Context |
|---|---|---|---|---|---|
| FLM-Optimized Adaptive Filtering [40] | 42.042 | - | MSE: Low | RMSE: Low | General EEG Artifact Removal |
| AnEEG (LSTM-GAN) [37] | Improved SNR & SAR | - | NMSE: Lower | RMSE: Lower | General EEG Artifact Removal |
| Mutual Information (Epanechnikov) [7] | - | 80.13 | - | - | Emotion Recognition |
| Hybrid CNN-LSTM (with EMG) [38] | Increased post-processing SNR | - | - | - | SSVEP Preservation during HRI |
| DBGS (Hardware-Software Hybrid) [41] | - | - | SAAF: 12.77 ± 0.85 dB | Correlation: 0.84 ± 0.33 | sEMG Extraction during FES |
Abbreviations: SNR (Signal-to-Noise Ratio), MSE (Mean Square Error), RMSE (Root Mean Square Error), NMSE (Normalized Mean Square Error), SAR (Signal-to-Artifact Ratio), SAAF (Stimulus Artifact Attenuation Factor).
Another critical metric for HRI is the computational performance of these models, which directly impacts their feasibility for real-time application.
Table 2: Computational and Operational Performance
| Model/Frame work | Key Innovation | Computational Performance | Failure Rate Reduction | Service Time Reduction |
|---|---|---|---|---|
| HICM [9] | Hybrid Intelligent Computing Model (Annealing + Tabu Search) | Calculation time reduced by 8.67% | 7.87% lower | 15.09% shorter |
| Real-time EEG Optimization [2] | Lightweight artifact removal & feature smoothing | Suitable for real-time constraints | - | - |
This protocol describes the procedure for implementing the AnEEG model, which uses a Long Short-Term Memory network integrated with a Generative Adversarial Network (LSTM-GAN) for effective artifact suppression [37].
1. Equipment and Software Setup:
2. Data Preparation and Preprocessing:
3. Model Architecture and Training:
4. Validation and Quantitative Analysis:
This protocol is specifically designed for removing muscle artifacts (EMG) from EEG signals in HRI contexts, such as those involving SSVEP, by leveraging additional EMG recordings [38].
1. Equipment and Software Setup:
2. Data Collection and Preprocessing:
3. Model Architecture and Training:
4. Validation and SSVEP-Specific Evaluation:
This diagram illustrates the end-to-end pipeline for processing neural signals to enable robust human-robot interaction, integrating artifact removal as a critical first step.
This diagram details the internal structure of a hybrid CNN-LSTM model that uses an additional EMG reference signal to remove muscle artifacts from EEG.
This section lists key computational tools, algorithms, and data resources that form the essential "reagent solutions" for developing and implementing the discussed artifact removal approaches in HRI research.
Table 3: Essential Research Tools and Resources
| Tool/Resource Name | Type | Primary Function | Application Context |
|---|---|---|---|
| AnEEG (LSTM-GAN) [37] | Deep Learning Model | Removes various artifacts by generating clean EEG signals. | General EEG preprocessing for clinical or BCI applications. |
| Hybrid CNN-LSTM (w/ EMG) [38] | Deep Learning Model | Specifically targets muscle artifacts using EMG reference. | HRI studies involving motion or SSVEP paradigms. |
| FLM Optimization [40] | Hybrid Training Algorithm | Firefly + Levenberg-Marquardt optimizes neural network weights. | Enhancing adaptive filters for artifact removal. |
| Mutual Information (Epanechnikov) [7] | Blind Source Separation | Identifies and removes artifact components based on mutual information. | Emotion recognition from EEG; general artifact removal. |
| DBGS Algorithm [41] | Hybrid Hardware-Software | Real-time template subtraction for stimulation artifacts. | Functional Electrical Stimulation (FES) environments. |
| HOC & Hjorth Features [7] | Feature Extraction | Provides stable, discriminative features from cleaned EEG signals. | Emotion estimation and pattern recognition post-artifact removal. |
| ICA & SOBI [2] [7] | Classical Algorithms | Baseline methods for blind source separation and component rejection. | Benchmarking and comparison against new deep learning methods. |
This application note provides a detailed protocol for establishing a real-time processing pipeline for artifact removal in human-robot interaction (HRI) research. Such pipelines are critical for developing brain-centered HRI experiences that are intuitive, effective, and capable of adapting to human cognitive and emotional states. The document outlines the complete system architecture, step-by-step experimental procedures, and validation methodologies required to implement a robust pipeline capable of processing electroencephalography (EEG) data under real-time constraints, thereby facilitating advanced affective HRI research.
The evolution of Human-Robot Interaction (HRI) is increasingly focused on creating brain-centered experiences where robots can understand and adapt to human states in real-time [42]. A significant challenge in this domain is that current robots often fail to comprehend personalized intents, attentions, and emotions, which prevents them from serving people appropriately across different contexts [42]. Real-time artifact removal from physiological signals like EEG is a foundational technology to overcome this barrier. It enables robots to dynamically interpret a user's emotional and cognitive state, paving the way for co-adaptive joint actions and more natural communication [42] [2].
Affective HRI requires lightweight software and affordable, wearable devices to become practically viable [2]. The real-time estimation of emotions from EEG data presents a particular optimization challenge, balancing processing speed with high accuracy. Traditional offline, supervised artifact removal methods often involve complex deep learning architectures with extensive hyper-parameter tuning, processes that can take days or weeks and are unsuitable for real-time applications [2]. The pipeline described herein addresses these obstacles by integrating optimized artifact removal and emotion estimation methodologies that function within strict real-time constraints, making it possible to conduct HRI studies that are both ecologically valid and scientifically rigorous.
The proposed pipeline is engineered for the continuous processing of EEG data, from acquisition to the final output of a classified emotional state, which can then be used by a robot to modulate its interaction. The entire system must operate with minimal latency to be effective in real-time HRI scenarios.
The logical flow of the real-time processing pipeline, from signal acquisition to robot action, is illustrated below.
Objective: To remove ocular artifacts from continuous EEG data with minimal loss of information and processing time.
Materials:
Methodology:
Validation:
Objective: To accurately classify a subject's emotional state from EEG features in real-time.
Materials:
Methodology:
Validation:
The performance of different components in the pipeline can be evaluated using the following quantitative metrics.
Table 1: Comparison of Real-Time Artifact Removal Techniques
| Technique | Primary Use | Key Parameters | Reported Processing Time | Advantages |
|---|---|---|---|---|
| ICA with Wavelet Analysis [2] | EOG & Muscle Artifact Removal | Components to reject, Wavelet family | Optimized for real-time | High accuracy in artifact separation |
| Regression-Based Methods | EOG Artifact Removal | Regression coefficients | Fast | Computationally lightweight |
| Notch Filter [2] | 50/60 Hz Powerline Noise | Frequency (50/60 Hz), Q-factor | Negligible | Highly effective for target noise |
| Bandpass Filter (1-50 Hz) [2] | Broadband Noise & Muscle Artifact | Low-cut (1Hz), High-cut (50Hz) | Negligible | Preserves key frequency bands for HRI |
Table 2: Emotion Classification Performance on SEED Database (Example)
| Classification Approach | Subject-Dependent Accuracy | Subject-Independent Accuracy | Suitable for Real-Time? | Notes |
|---|---|---|---|---|
| Proposed Methodology (e.g., with LDS smoothing) [2] | High (>90% achievable) | Maintains high performance | Yes | Balanced accuracy and speed |
| Deep Learning Models (e.g., CNNs) [2] | Very High | Can be high with large data | Often No | High computational cost, slow tuning |
| SVM with Raw Features | Moderate | Lower | Yes | Lower accuracy, especially cross-subject |
The workflow for the emotion estimation protocol, from pre-processing to classification, is detailed below.
This section lists the essential hardware, software, and algorithmic "reagents" required to implement the real-time HRI pipeline.
Table 3: Essential Materials and Tools for the HRI Pipeline
| Item Name / Category | Specification / Example | Primary Function in the Pipeline |
|---|---|---|
| Wearable EEG Headset | OpenBCI [42] | Acquires raw brainwave data from the user in a portable format suitable for HRI settings. |
| Real-Time Data Bus | Apache Kafka [43] | Handles the continuous, low-latency streaming of EEG data and messages between pipeline components. |
| Artifact Removal Algorithm | Real-Time ICA [2] | Identifies and removes ocular and muscle artifacts from the EEG stream to clean the data. |
| Feature Set | Beta & Gamma Bandpower [2] | Serves as the input variables for the classifier, capturing the neural correlates of emotion. |
| Feature Smoothing Algorithm | Linear Dynamic Systems (LDS) [2] | Reduces temporal variability in features, improving the stability and accuracy of emotion estimation. |
| Classification Model | Linear SVM [2] | A lightweight, fast model that maps smoothed EEG features to a discrete emotional state. |
| Robotics Middleware | Robot Operating System (ROS) [44] | Provides the software framework to integrate the emotion classification output with robot control logic. |
Implementing this pipeline in live HRI studies requires careful consideration of several factors. The entire system must be designed for low latency to ensure the robot's response is perceived as timely and natural by the human user [2]. Furthermore, the choice between subject-dependent and subject-independent models is crucial. While subject-dependent models can offer higher accuracy, subject-independent models are more practical for applications with new users [2]. Finally, a robust monitoring framework should be established to track system health metrics such as latency, error frequency, and data saturation in real-time, ensuring the pipeline's reliability during extended experimental sessions [43].
Affective human-robot interaction (HRI) requires software that is not only accurate but also computationally efficient to function under real-time constraints. A significant challenge in this domain is processing electroencephalography (EEG) signals, which involves critical steps like artifact removal and emotion estimation without introducing disruptive delays [3]. The optimization of these processes is paramount for creating fluid and natural interactions between humans and robots, particularly when using wearable EEG devices in real-world, uncontrolled environments [31]. This document outlines application notes and protocols for achieving such optimization, focusing on methodologies that balance high accuracy with minimal processing time.
Processing EEG signals for real-time HRI presents two primary, interconnected challenges that must be addressed to ensure system efficacy and user acceptance.
Effective artifact removal is a prerequisite for reliable EEG analysis. The following protocols detail optimized approaches for different artifact types.
Protocol 3.1.1: Real-Time Electro-oculographic (EOG) Artifact Removal
Protocol 3.1.2: Filtering for Muscular and Environmental Artifacts
The following workflow integrates these techniques into a cohesive real-time processing pipeline.
Once artifacts are mitigated, the cleaned signal can be used for state estimation. The protocols below focus on optimizing feature extraction and classification.
Protocol 3.2.1: Optimized Emotion Estimation Workflow
Protocol 3.2.2: Feature-Based Error-Related Potential (ErrP) Detection
To objectively evaluate the computational efficiency and accuracy of the optimized system, the following metrics should be used.
Table 1: Key Performance Metrics for Real-Time EEG Processing
| Metric | Formula | Target for Real-Time HRI |
|---|---|---|
| Artifact Removal Processing Time | N/A | < 100 ms per epoch [3] |
| Classification Accuracy | (TP + TN) / (TP + TN + FP + FN) | > 85% (subject-dependent) [3] |
| Recall (Sensitivity) | TP / (TP + FN) | Maximize (Critical for ErrP detection) [26] |
| F1 Score | 2 × (Precision × Recall) / (Precision + Recall) | > 0.7 (for imbalanced ErrP datasets) [26] |
Abbreviations: TP (True Positive), TN (True Negative), FP (False Positive), FN (False Negative).
Implementing the aforementioned protocols requires a suite of software tools and algorithmic approaches.
Table 2: Essential Research Reagents for Real-Time EEG-HRI
| Tool/Algorithm | Type | Function in the Pipeline |
|---|---|---|
| Independent Component Analysis (ICA) | Algorithm | Blind source separation for isolating ocular and muscular artifacts from neural signals [3] [31]. |
| Wavelet Transform | Algorithm | Multi-resolution analysis for precise, localized correction of artifacts in specific signal components [3] [31]. |
| Artifact Subspace Reconstruction (ASR) | Algorithm/Pipeline | Statistical method for identifying and removing high-variance artifact components; suitable for various artifact types [31]. |
| Linear Discriminant Analysis (LDA) | Algorithm | A lightweight, efficient classifier suitable for subject-independent ErrP detection and emotion classification [26]. |
| SEED Database | Dataset | A publicly available dataset for benchmarking emotion estimation methodologies using EEG [3]. |
The strategic application of these tools and protocols is summarized in the following optimization strategy diagram.
In human-robot interaction (HRI) research, real-time analysis of neural signals via electroencephalography (EEG) provides a critical window into human cognitive and affective states. However, full-body movement, essential to naturalistic HRI, introduces motion artifacts that severely compromise EEG signal quality [45] [46]. A central challenge is optimizing artifact removal parameters to maximize noise suppression without removing neural signals of interest—a problem known as "over-cleaning." This application note synthesizes proven parameter-tuning strategies from Automatic Speech Recognition (ASR) and the iCanClean EEG artifact removal algorithm, providing a framework for developing robust real-time artifact removal pipelines in HRI research.
The following tables consolidate quantitative findings from empirical studies on iCanClean and ASR, providing a basis for informed parameter selection.
Table 1: iCanClean Parameter Sweep Results for Mobile EEG (from Young, Older, and Low-Functioning Older Adults during Walking) [47]
| Window Length | R² Threshold (Aggressiveness) | Average Number of "Good" ICA Components | Performance Change vs. Baseline |
|---|---|---|---|
| Not Applied (Baseline) | Not Applied | 8.4 | Baseline |
| 4 seconds | 0.65 | 13.2 | +57% |
| 4 seconds | 0.60 | ~12.7 | +51% |
| 4 seconds | 0.70 | ~12.2 | +45% |
| 2 seconds | 0.65 | ~12.0 | +43% |
| 1 second | 0.65 | ~11.5 | +37% |
| Infinite | 0.65 | ~10.5 | +25% |
Table 2: Comparative Performance of Artifact Removal Methods on a Phantom Head Model (Data Quality Score %) [34]
| Artifact Condition | Uncleaned | iCanClean | ASR | Auto-CCA | Adaptive Filtering |
|---|---|---|---|---|---|
| Brain (Target) | 57.2% | - | - | - | - |
| Brain + All Artifacts | 15.7% | 55.9% | 27.6% | 27.2% | 32.9% |
| Brain + Walking Motion | 21.5% | 56.5% | 30.1% | 28.8% | 35.2% |
Table 3: Artifact Subspace Reconstruction (ASR) Parameter Impact [46]
| ASR Parameter (k) | Cleaning Aggressiveness | Impact on ICA Decomposition | Recommended Use Case |
|---|---|---|---|
| k = 10 | Very High | Risk of significant "over-cleaning" and signal loss | Not recommended for locomotion |
| k = 20 | High | Moderate risk of over-cleaning; use with caution | Stationary tasks with large artifacts |
| k = 30 | Moderate | Balance between cleaning and preservation | General use [46] |
| k = 50+ | Low | Limited motion artifact removal | Minimal artifact scenarios |
Objective: To empirically determine the optimal iCanClean window length and R² threshold for a specific HRI experimental setup involving human movement [47].
Materials: Dual-layer EEG cap with 120 scalp electrodes and 120 noise electrodes; amplification system; computing setup with iCanClean software.
Procedure:
Objective: To compare the efficacy of iCanClean and ASR in recovering event-related potentials (ERPs) during a dynamic HRI task, such as a Flanker task performed while jogging [46].
Materials: Mobile EEG system; setup for a dynamic cognitive task (e.g., visual stimuli presented during robot interaction).
Procedure:
k parameter of 20-30.
Diagram 1: The iCanClean workflow and parameter influence.
Diagram 2: The consequences of over-cleaning and their causes.
Table 4: Essential Materials and Tools for Real-Time Artifact Removal Research
| Item Name | Function/Description | Application in HRI Context |
|---|---|---|
| Dual-Layer EEG Cap | A specialized cap with scalp electrodes and mechanically coupled, outward-facing noise electrodes. The noise electrodes record artifact signals without brain activity [47] [34]. | Provides ideal reference noise signals for iCanClean during dynamic HRI tasks involving walking, reaching, or head movement. |
| iCanClean Software | An algorithm that uses Canonical Correlation Analysis (CCA) and reference noise signals to detect and remove noise subspaces from EEG data [47] [34]. | The core tool for real-time, all-in-one artifact removal without requiring clean calibration data. |
| Artifact Subspace Reconstruction (ASR) | A real-time-capable algorithm in EEGLAB/BCILAB that uses principal component analysis (PCA) to remove high-variance artifacts based on a clean calibration period [46]. | An alternative cleaning method when dual-layer caps are unavailable; requires careful tuning of the k parameter. |
| ICLabel EEGLAB Plugin | A trained classifier that automatically labels Independent Components (ICs) from ICA based on source type (e.g., brain, muscle, eye, heart) [47] [46]. | Critical for the quantitative assessment of cleaning quality by counting brain-derived components post-ICA. |
| High-Density EEG System (100+ channels) | An EEG recording system with a sufficient number of electrodes to enable high-quality ICA decomposition and source localization [47]. | Ensures adequate spatial sampling for unmixing neural and artifactual sources in complex HRI environments. |
The emergence of low-density, wearable electroencephalography (EEG) systems represents a paradigm shift in neurophysiological monitoring, offering unprecedented opportunities for real-world brain activity assessment. These portable, affordable devices are increasingly overcoming the limitations of traditional high-density EEG labs, which are characterized by high operational costs, limited accessibility, and artificial recording environments [48]. For human-robot interaction (HRI) research, wearable EEG systems enable the investigation of brain dynamics in naturalistic settings, providing crucial insights into emotional and cognitive states during interactive tasks. However, the transition to low-density systems presents significant challenges, particularly regarding signal quality and artifact contamination in real-time applications [3] [2]. This application note outlines comprehensive strategies for implementing low-density, wearable EEG systems, with emphasis on experimental protocols, artifact removal techniques, and validation methods specifically tailored for HRI research contexts.
Modern brain wearables leverage several advanced technologies that enable reliable brain monitoring outside traditional clinical settings. Understanding these core technologies is essential for selecting appropriate systems and optimizing their implementation in HRI research.
Dry Electrode EEG Systems: Conventional EEG systems require skin abrasion, conductive gel application, and trained technicians—processes that are time-consuming and uncomfortable for patients. Dry electrode technology eliminates these requirements, making it suitable for home-based monitoring and real-time HRI applications. QUASAR's dry electrode EEG sensors feature ultra-high impedance amplifiers (>47 GOhms) that handle contact impedances up to 1-2 MOhms, producing signal quality comparable to wet electrodes. This technology enables recordings through hair without skin preparation, while patented mechanical isolation designs stabilize electrodes for artifact-free recordings even during movement [48].
Ear-EEG Systems: Ear-EEG represents a significant breakthrough for long-term monitoring applications, allowing discreet, comfortable brain monitoring. These systems capture EEG signals from within the ear canal using either dry or wet electrodes. The Naox device employs dry-contact electrodes with active electrode technology featuring 13 TΩ input impedance to minimize noise despite higher electrode-skin impedance (approximately 300 kΩ). Recent innovations include user-generic earpieces with dry electrodes that eliminate hydrogels while maintaining signal quality comparable to wet electrode systems [48].
Multimodal Integration: Beyond electrical activity measurement, modern brain wearables increasingly incorporate complementary technologies. Functional near-infrared spectroscopy (fNIRS) measures changes in blood oxygenation and volume in the cortex, providing complementary insights into brain activity patterns. As a non-invasive neuroimaging modality, fNIRS offers several advantages for portable monitoring, including strong agreement with simultaneously acquired fMRI measurements and greater tolerance to noise and movement than EEG [48].
Table 1: Comparison of Wearable EEG Technologies for HRI Research
| Technology | Spatial Resolution | Comfort for Long Sessions | Motion Artifact Resistance | Setup Time | Best Use Cases in HRI |
|---|---|---|---|---|---|
| Dry Electrode Headsets | Medium-High | Medium | Medium | ~4 minutes | Controlled laboratory HRI studies |
| Ear-EEG Systems | Low | High | High | <2 minutes | Long-duration naturalistic interaction studies |
| Multimodal (EEG+fNIRS) | Medium (EEG) + Low (fNIRS) | Medium | Medium-Low | 5-10 minutes | Complex cognitive state assessment |
| Adhesive Patch Systems | Low | Medium-High | High | ~3 minutes | Ambulatory studies with movement |
Implementing robust experimental protocols is essential for collecting valid, reproducible EEG data in HRI contexts. The following protocols address the specific challenges of low-density systems in interactive scenarios.
The HEROIC (Home EEG Recording frOm Interfacing Computer) platform provides an open-source framework for remote EEG data collection during customized neurocognitive tasks. This platform enables participants to independently collect advanced EEG data without expert technician assistance, making it particularly valuable for longitudinal HRI studies [49].
Device Initialization and Setup:
Session Recording:
Post-Session Completion:
Affective HRI requires lightweight software and wearable devices capable of real-time emotion estimation. The following protocol optimizes this process for low-density systems:
Stimuli Presentation:
EEG Data Acquisition Parameters:
Feature Extraction and Processing:
Artifact contamination represents the most significant challenge for reliable EEG analysis in HRI contexts, particularly with low-density systems. Effective artifact removal is essential for accurate emotion estimation and cognitive state classification.
Electro-oculographic (EOG) Artifact Removal:
Mutual Information-Based Approaches: Novel Blind Source Separation algorithms based on Mutual Information (MI) minimization have demonstrated superior artifact removal performance for emotion recognition tasks. These methods utilize:
Optimized Processing Pipeline:
Table 2: Performance Comparison of Artifact Removal Methods for Emotion Recognition
| Method | Computational Efficiency | Classification Accuracy | Information Preservation | Ease of Implementation | Recommended HRI Context |
|---|---|---|---|---|---|
| SOBI | High | 68.15% | Medium | High | Basic real-time applications with limited processing resources |
| MI with Gaussian Kernel | Medium | 78.33% | Medium-High | Medium | Standard laboratory HRI studies with moderate artifact contamination |
| MI with Epanechnikov Kernel | Medium-High | 80.13% | High | Medium | Advanced HRI applications requiring high classification accuracy |
| ICA with Wavelet Analysis | Medium | 75-82% (task-dependent) | High | Medium | Scenarios with prominent EOG artifacts |
Modern artifact handling increasingly leverages machine learning approaches that simultaneously address artifact contamination and state classification:
Feature Selection Optimization:
Channel Selection Strategies: Research demonstrates that strategic channel selection can maintain classification performance while significantly reducing system complexity:
Rigorous validation is essential for establishing the reliability of low-density EEG systems in HRI applications. The following approaches provide comprehensive performance assessment:
Event-Related Potential (ERP) Validation:
Simultaneous Recording Protocols:
Sleep Staging Validation:
Establishing correlation between EEG metrics and behavioral measures strengthens the validity of low-density systems for HRI research:
Cognitive State Correlations:
Clinical Application Validation:
Despite significant advances, several challenges remain in the widespread implementation of low-density wearable EEG systems for HRI research:
Data Quality and Signal-to-Noise Ratio: Low-density systems face inherent signal-to-noise ratio limitations compared to high-density clinical systems. Mitigation strategies include:
Inter-Subject and Cross-Session Variability: EEG signals exhibit substantial variability across individuals and recording sessions, particularly problematic for real-time HRI applications:
Technical and Computational Constraints: Wearable systems balance performance with practical constraints:
Table 3: Essential Research Tools for Low-Density Wearable EEG Studies
| Tool/Platform | Type | Primary Function | Key Features | Compatibility |
|---|---|---|---|---|
| HEROIC | Open-source software platform | Remote EEG data collection | Stimulus synchronization, quality control, multiple device support | Muse, Emotiv, OpenBCI |
| Muse 2 | Consumer-grade EEG headset | Affordable EEG acquisition | 4 electrodes, ~$250 cost, validated for ERP collection | HEROIC, Mind Monitor, Muse Direct |
| Emotiv EPOC X | Research-grade wearable EEG | High-quality mobile acquisition | 14 channels, saline electrodes, research validation | Emotiv Pro, custom MATLAB |
| cEEGrid | Ear-EEG system | Discreet, long-term monitoring | Around-ear electrode array, minimal setup | OpenBCI, custom amplifiers |
| Autoreject | Python library | Automated EEG preprocessing | Bad channel detection, epoch rejection, interpolation | MNE-Python compatible |
| MARA | Automated artifact removal | Component classification | Machine learning-based ICA component rejection | EEGLAB, Python |
Low-density, wearable EEG systems represent a transformative technology for human-robot interaction research, enabling the investigation of brain dynamics in naturalistic environments and real-time adaptive interactions. The strategies outlined in this application note—covering technical implementation, experimental protocols, advanced artifact removal, and validation frameworks—provide researchers with comprehensive guidelines for successful system deployment. While challenges remain in signal quality, computational efficiency, and cross-subject reliability, ongoing advances in dry electrode technology, machine learning approaches, and open-source software platforms continue to enhance the capabilities of these systems. By implementing the detailed protocols and methodologies presented here, HRI researchers can leverage wearable EEG technology to gain unprecedented insights into human cognitive and emotional states during interactive tasks, ultimately advancing the development of more responsive, adaptive robotic systems.
Within human-robot interaction (HRI) research, the accurate measurement of human state and movement is paramount for enabling seamless and safe collaboration. Inertial Measurement Units (IMUs) are crucial auxiliary sensors for this purpose, providing data on human motion and physical interaction. However, their signals are susceptible to various noise sources and artifacts that can degrade performance in real-time systems. This application note details the role of IMUs and the principles of noise reference utilization, providing structured protocols and data to enhance artifact removal in HRI research.
An Inertial Measurement Unit (IMU) is an electromechanical or solid-state device that measures the specific force, angular rate, and, sometimes, the magnetic field surrounding an object. Typically, it contains a suite of sensors orthogonally mounted to provide measurements along three axes [55]:
The proliferation of Micro-Electromechanical Systems (MEMS) technology has made IMUs extremely small, lightweight, low-power, and cost-effective, facilitating their integration into wearable HRI systems [55]. MEMS accelerometers operate on the principle of a sprung proof mass; acceleration causes displacement of this mass, which is measured via changes in electrical capacitance and converted into an acceleration value [55].
IMU data is corrupted by various error sources, which can be characterized as deterministic (bias, scale factor errors) or stochastic (noise). Key noise-related concepts include:
Table 1: Key Performance Metrics for MEMS and FOG IMUs (adapted from [55])
| Performance Metric | MEMS IMU (e.g., Motus) | Fiber-Optic Gyro (FOG) IMU (e.g., Boreas D90) | Unit |
|---|---|---|---|
| Roll/Pitch Accuracy | 0.05 | 0.005 | ° |
| Heading Accuracy | 0.8 (magnetic) | 0.01 | ° |
| Gyro Bias Instability | 0.2 | 0.001 | °/hr |
| Weight | ~26 | ~2500 | g |
| Power Consumption | ~1.4 | ~12 | W |
Effective artifact removal is critical for leveraging IMU data in real-time HRI. Advanced signal processing techniques have demonstrated significant improvements.
A prominent method for inertial sensor denoising involves the Lifting Wavelet Transform (LWT). LWT is a popular denoising technique with advantages over classic wavelets in terms of time and computational complexity [56]. One advanced approach optimizes this process using a Genetic Algorithm (GA) for intelligent averaging of different multi-level LWT outputs [56].
This hybrid method (LWT-GA) has shown substantial performance gains in both static and dynamic scenarios relevant to HRI, as summarized in Table 2.
Table 2: Performance Improvement of LWT-GA Denoising Method for Inertial Sensors (data from [56])
| Data Type | Sensor | Reported Improvement | Notes |
|---|---|---|---|
| Dynamic Data | Gyroscope | 83% | Compared to raw sensor data. |
| Dynamic Data | Accelerometer | 59% | Compared to raw sensor data. |
| Static Data | Gyroscope | 71% | Compared to raw sensor data. |
| Static Data | Accelerometer | 36% | Compared to raw sensor data. |
The principles of artifact removal extend to other sensors used in HRI. In EEG signal analysis, for example, Blind Source Separation (BSS) algorithms are used to identify and remove artifacts. A Mutual Information (MI)-based BSS algorithm using an Epanechnikov kernel for probability density estimation demonstrated superior performance, achieving 80.13% accuracy in an emotion recognition task, outperforming both Gaussian kernel-based MI and classical SOBI algorithms [7]. This underscores the importance of selecting advanced, computationally efficient algorithms for real-time artifact removal in multi-modal HRI systems.
Objective: To integrate an IMU into a multi-modal HRI system and achieve temporal synchronization of sensor data streams for robust artifact analysis.
Materials:
Methodology:
Objective: To quantitatively evaluate the performance of a denoising algorithm (e.g., LWT-GA) on IMU data collected in an HRI context.
Materials:
Methodology:
Improvement (%) = [(Metric_raw - Metric_denoised) / Metric_raw] * 100 [56].The following diagram illustrates the integrated workflow for using auxiliary sensors and noise references in an HRI artifact removal pipeline.
Table 3: Key Research Reagent Solutions for IMU-based HRI Studies
| Item / Solution | Function / Application in Research |
|---|---|
| MEMS IMU (e.g., Advanced Navigation Motus) | The primary sensor for measuring linear acceleration and angular rate. Its small size, weight, and power (SWaP-C) profile make it ideal for wearable HRI applications [55]. |
| High-Precision Motion Capture System (e.g., OptiTrack) | Serves as a ground truth reference system for validating the accuracy of denoised IMU data and derived trajectories. |
| MATLAB or Python with Signal Processing Toolbox | Software environment for implementing and testing denoising algorithms (e.g., Lifting Wavelet Transform, Blind Source Separation). |
| Robot Operating System (ROS) | Middleware framework for synchronizing, logging, and processing multi-sensor data streams in real-time HRI experiments. |
| Epanechnikov Kernel-based BSS Algorithm | A computational tool for artifact removal, particularly effective in separating noise from source signals in EEG and other biophysical data [7]. |
| Genetic Algorithm Optimization Library | Used to fine-tune the parameters of denoising algorithms, such as wavelet threshold levels, to maximize performance [56]. |
In the field of affective Human-Robot Interaction (HRI), the real-time estimation of human emotions from Electroencephalography (EEG) signals presents a critical engineering challenge: achieving an optimal balance between high signal fidelity and low processing speed [3] [2]. The development of automatic systems for patient therapy and evaluation depends on the robot's ability to adapt its behavior dynamically to a patient's changing mood, a process which requires both high accuracy in emotion classification and near-instantaneous processing [3]. This application note details the core trade-offs, quantitative performance metrics, and standardized protocols for implementing real-time artifact removal in HRI research, a cornerstone for building reliable and responsive closed-loop systems.
The primary obstacle in real-time EEG processing is online artifact removal, which must cleanse the signal of noise without discarding valuable neural information or introducing prohibitive computational delays [3]. The most common artifacts are electro-oculographic (EOG) blinks, muscle activity, and 50 Hz background noise, each occupying different frequency bands and requiring tailored removal strategies [3] [2].
Table 1: Performance Comparison of Real-Time Artifact Removal and Classification Methods
| Methodology | Reported Accuracy | Key Strengths | Primary Trade-offs |
|---|---|---|---|
| ICA + Wavelet Analysis [3] | High (Specific values not provided) | Effective for EOG artifacts; works within real-time constraints. | Potential loss of neural information; requires careful parameter tuning. |
| Mutual Information (Epanechnikov Kernel) [7] | 80.13% | High accuracy; lower computational cost than Gaussian kernel. | Simplicity may come at the cost of robustness for highly complex artifacts. |
| Mutual Information (Gaussian Kernel) [7] | Slightly inferior to Epanechnikov | Established method with known properties. | Higher computational cost than Epanechnikov kernel. |
| Second Order Blind Identification (SOBI) [7] | ~12% lower than best MI method | Classical, commonly used algorithm. | Lower performance in emotion classification contexts. |
A study optimizing emotion estimation for HRI demonstrated that a carefully chosen methodology could operate under real-time constraints while maintaining high accuracy in both subject-dependent and subject-independent paradigms [3] [2]. Furthermore, research on mutual information-based blind source separation revealed that the choice of algorithm directly impacts this balance; a method using an Epanechnikov kernel achieved 80.13% accuracy in emotion classification while offering a lower computational cost than a comparable Gaussian kernel approach [7]. This underscores that even within a class of algorithms, specific implementations can yield better fidelity-speed outcomes.
This protocol is adapted from methodologies proven effective for real-time affective HRI [3] [2].
1. Objective: To clean EEG signals of artifacts and extract features for emotion classification within a time frame suitable for seamless human-robot interaction.
2. Materials & Equipment:
3. Procedure:
4. Critical Timing Parameters: The entire pipeline, from Step 1 to Step 6, must complete processing within a single processing window (e.g., 1-2 seconds) to be considered "real-time" and to keep pace with the dynamic nature of HRI.
This protocol details a specific, high-performance artifact removal method suitable for emotion recognition [7].
1. Objective: To remove artifacts from EEG signals using Mutual Information with Epanechnikov kernel density estimation to improve emotion classification accuracy.
2. Materials & Equipment:
3. Procedure:
Table 2: Key Research Reagent Solutions for Real-Time EEG Processing
| Item Name | Function/Application | Specific Example/Note |
|---|---|---|
| Wearable EEG Headset | Acquires brain signals in a non-invasive, user-friendly manner. | Systems with pre-configured electrodes at AF3, AF4, T7, T8, TP7, TP8, P7, P8 are ideal for emotion estimation [3]. |
| Independent Component Analysis (ICA) | A blind source separation technique for decomposing signals into statistically independent components, crucial for isolating artifacts. | Used as a core step in real-time EOG artifact removal when combined with wavelet analysis [3]. |
| Mutual Information Algorithm | A method for blind source separation that minimizes the mutual information between estimated components. | The Epanechnikov kernel variant offers high accuracy with lower computational cost [7]. |
| MARA (Multiple Artifact Rejection Algorithm) | An automated tool for classifying independent components from ICA as artifact or brain signal. | Eliminates the need for manual component inspection, facilitating real-time processing [7]. |
| Support Vector Machine (SVM) | A supervised machine learning model used for classification tasks, such as mapping EEG features to emotional states. | Effective for emotion classification post-artifact removal, offering a good balance of accuracy and speed [7]. |
The following diagram illustrates the complete real-time EEG processing pipeline for human-robot interaction, integrating both the signal processing and the closed-loop HRI components.
Real-Time EEG Processing Pipeline for HRI
Achieving a functional balance between signal fidelity and processing speed is not merely a technical optimization problem but a fundamental requirement for the success of real-time affective Human-Robot Interaction. The protocols and data presented herein provide a foundation for researchers to build upon. The choice between artifact removal methods like ICA-wavelet hybrids and mutual information-based approaches should be guided by the specific accuracy and latency demands of the target HRI application. Future work must continue to refine these models, pushing the boundaries of computational efficiency while preserving the rich information contained within neural signals to enable ever more seamless and natural human-robot collaboration.
In human-robot interaction (HRI) research, neurophysiological monitoring through electroencephalography (EEG) provides critical insight into human cognitive and affective states. These signals enable robots to adapt their behavior in real-time, fostering seamless collaboration. However, electrophysiological data are persistently contaminated by artifacts from ocular, muscular, and environmental sources, which can corrupt interpretation and derobot decision-making. Validating the performance of artifact removal algorithms is therefore paramount. This application note establishes rigorous protocols for evaluating such methods using three core metrics: Accuracy, Selectivity, and Dipolarity. These metrics collectively ensure that artifact removal preserves neural signals of interest while effectively eliminating contaminants, which is crucial for developing reliable real-time HRI systems.
Accuracy quantifies the fidelity of the cleaned neural signal by measuring the deviation between the processed output and an uncontaminated ground-truth signal. In real-time HRI, high accuracy is essential to prevent robots from misinterpreting a user's cognitive state [3].
The most common accuracy metric is the Root Mean Square Error (RMSE), calculated as:
RMSE = √( Σ(ŷᵢ - yᵢ)² / N )
where ŷᵢ is the cleaned signal, yᵢ is the ground-truth signal, and N is the number of data points. A lower RMSE indicates higher accuracy.
Selectivity evaluates an algorithm's ability to isolate and remove artifacts without distorting the underlying neural activity. This is particularly important in HRI, where signals like event-related desynchronization (ERD) must be preserved to accurately decode human motor intention or cognitive load [57].
Selectivity can be assessed by comparing the power in a frequency band of interest (e.g., the alpha band for ERD) before and after artifact removal in a known paradigm.
Dipolarity is a physiological plausibility check for components identified by blind source separation (BSS) methods like Independent Component Analysis (ICA). It measures how well the scalp topography of a component can be explained by a single equivalent dipole in the brain. Neural sources typically originate in the brain and thus have high dipolarity, while artifacts (from eyes, muscles, or heart) do not [57].
A common measure is the dipolarity index or the residual variance from a single-dipole fit. Components with a residual variance below a threshold (e.g., <15%) are considered "near-dipolar" and likely neural in origin.
This protocol is designed for a controlled, quantitative evaluation of artifact removal accuracy when a ground-truth signal is available.
S_clean).S_clean with a recorded artifact signal (A) at a known Signal-to-Noise Ratio (SNR) to create a contaminated signal (S_contaminated).S_contaminated to obtain the cleaned signal (S_cleaned).S_cleaned and S_clean.This protocol validates selectivity using the well-established phenomenon of event-related desynchronization (ERD) in a real data scenario where absolute ground truth is unavailable.
This protocol is used to validate the components classified as "neural" by a BSS algorithm, ensuring they correspond to physiologically plausible brain sources.
The following tables summarize quantitative data from benchmark studies, providing a reference for expected performance in HRI research.
Table 1: Comparative Performance of Blind Source Separation (BSS) Methods on Real EEG Data During a Foot Movement Task (n=18 subjects). Performance is measured by the ability to reduce muscle artifacts while preserving the Event-Related Desynchronization (ERD). Adapted from [57].
| BSS Method | Artifact Reduction Efficacy | ERD Preservation Quality | Key Advantage |
|---|---|---|---|
| Extended Infomax | High | High | Best overall performance in the comparison |
| FastICA | High | Medium-High | Robust for super-Gaussian sources |
| TDSEP/SOBI | High | Medium-High | Exploits temporal structure |
| Fourier-ICA | High | Medium | Optimized for oscillatory sources |
| Spatio-Spectral Decomposition (SSD) | High | Medium | Maximizes SNR in a band of interest |
Table 2: Example Residual Variance (Dipolarity) and Classification Outcomes for Independent Components from a Single Subject. Components with RV < 15% are classified as neural.
| Component Index | Residual Variance (%) | Proposed Classification | Notes |
|---|---|---|---|
| IC 1 | 4.2% | Neural | Frontal theta, retained |
| IC 2 | 78.5% | Artifact | Lateral eye movement, rejected |
| IC 3 | 9.1% | Neural | Mu rhythm, retained |
| IC 4 | 95.2% | Artifact | Muscle artifact, rejected |
| IC 5 | 6.8% | Neural | Occipital alpha, retained |
Table 3: Essential Software and Analytical Tools for Validating Artifact Removal in HRI Research.
| Tool / Resource | Function | Application in Protocol |
|---|---|---|
| EEGLAB | An open-source MATLAB toolbox for processing electrophysiological data. | Core platform for running ICA, dipole fitting, and calculating time-frequency features [57]. |
| BCILAB | A toolbox for building brain-computer interface models. | Useful for prototyping and testing real-time classification pipelines within HRI systems. |
| ICMARC | An automated independent component classifier. | Automates the classification of ICA components as neural or artifactual based on multiple features, including dipolarity [57]. |
| DIPFIT | A plugin within EEGLAB for equivalent dipole modeling. | Directly calculates the residual variance for dipolarity assessment of components (Protocol 3) [57]. |
| Boundary Element Model (BEM) | A volume conduction head model. | Serves as the forward model for the DIPFIT toolbox to compute the single-dipole fit. |
Electroencephalography (EEG) is a crucial tool for studying brain dynamics in real-world settings, including human-robot interaction (HRI). However, motion artifacts significantly compromise EEG signal quality, making robust artifact removal essential for generating reliable neural data. This application note provides a structured comparison of three prominent artifact removal methods—Independent Component Analysis (ICA), Artifact Subspace Reconstruction (ASR), and iCanClean—focusing on their efficacy in mobile scenarios relevant to HRI research.
ICA is a blind source separation technique that decomposes multi-channel EEG data into statistically independent components, which are then classified and manually inspected to remove artifactual sources [58] [59]. The Adaptive Mixture ICA (AMICA) algorithm is recognized as one of the most powerful variants, though it is computationally intensive [59]. ICA operates as a stationary method, assuming a single spatial filter applies throughout the recording, making it highly effective for separating constant, fixed-source artifacts like eye blinks and muscle activity [60]. However, its performance depends on data stationarity, and it struggles with non-stationary, high-amplitude motion artifacts [58] [60].
ASR is an automated, real-time capable method designed to remove large-amplitude, non-stationary artifacts using a sliding-window approach [46] [58] [60]. Based on principal component analysis (PCA), ASR calibrates on a clean baseline data segment. It then identifies and reconstructs artifact-dominated subspaces in new data that exceed a user-defined standard deviation threshold (parameter "k") [46] [60]. ASR is particularly effective for removing transient, high-amplitude artifacts such as motion-induced spikes and cable sway, thereby improving data stationarity for subsequent ICA decomposition [58] [60].
iCanClean is a novel framework that uses canonical correlation analysis (CCA) with reference noise signals to identify and subtract noisy subspaces from EEG data [6] [46] [36]. It can operate with dedicated noise sensors (e.g., in a dual-layer electrode setup) or generate pseudo-reference noise signals from the EEG itself by applying a temporary notch filter [46]. A key parameter is the R² cleaning aggressiveness, which determines the correlation threshold for noise removal [46] [36]. iCanClean is designed for real-time implementation and effectively handles multiple simultaneous artifact types, including motion, muscle, and eye artifacts [6].
The following tables summarize key performance metrics and characteristics from empirical studies.
Table 1: Performance Comparison on Phantom and Human Data
| Metric | iCanClean | ASR | ICA | Notes & Context |
|---|---|---|---|---|
| Data Quality Score (Phantom) | 55.9% | 27.6% | N/A | "Brain + All Artifacts" condition; baseline was 15.7% [6] |
| Good ICs After Cleaning | 13.2 | Varies | 8.4 (pre-cleaning) | iCanClean increased viable brain components by 57% [36] |
| ERP Congruency Effect | Identified | Identified | N/A | During running; P300 effect found with iCanClean & ASR [46] |
| Single-Trial Classification | N/A | Effective | Effective | ASR+ICA pipeline outperformed minimal cleaning (69% vs 55%) [58] |
| Optimal Parameters | R²=0.65, 4s window [36] | k=20-30 [60] | AMICA with sample rejection [59] | Parameter sweeps identified optimal settings |
Table 2: Method Characteristics and Applicability
| Characteristic | ICA | ASR | iCanClean |
|---|---|---|---|
| Primary Strength | Separates fixed-source artifacts | Removes transient, high-amplitude bursts | Removes multiple, co-occurring artifact types |
| Real-Time Capability | Low (offline) | High | High |
| Computational Speed | Slow (hours) [6] | Fast | Fast |
| Hardware Dependency | Standard EEG | Standard EEG | Optimal with dual-layer EEG [46] [36] |
| Key Limitation | Requires stationarity; slow | Requires clean calibration data | Performance depends on noise reference quality |
This methodology, used to validate iCanClean, provides ground-truth data for quantitative comparisons [6] [34].
This protocol is designed for evaluating methods on human data during tasks like walking or running, relevant for dynamic HRI [46].
This protocol tests robust pipeline combinations for highly dynamic scenarios like sports [58].
The following diagram illustrates the typical data processing flow and the role of each artifact removal method.
Artifact Removal Processing Workflow: This diagram illustrates how ICA, ASR, and iCanClean can be integrated into an EEG processing pipeline. ASR and iCanClean, being real-time capable, are often used as a preprocessing step before ICA, especially for data with significant motion artifacts. They remove large, non-stationary bursts, thereby creating a more stationary data set that improves the subsequent ICA decomposition [58] [60]. iCanClean can be used in place of ASR when reliable noise references are available [46] [36].
Table 3: Key Materials and Tools for Mobile EEG Artifact Removal Research
| Tool / Material | Function & Application | Example / Note |
|---|---|---|
| High-Density EEG System | Acquires neural data; essential for effective source separation with ICA. | 100+ channels recommended for mobile studies [6]. |
| Dual-Layer EEG Cap | Provides dedicated noise references for iCanClean. | 120 scalp electrodes + 120 noise electrodes [36]. |
| Phantom Head Apparatus | Validates cleaning algorithms with known ground-truth signals. | Contains embedded artificial brain and artifact sources [6] [34]. |
| Inertial Measurement Units (IMUs) | Captures motion dynamics; can be used as reference for artifact removal. | Head-mounted IMUs to correlate motion with EEG artifacts [35]. |
| AMICA Algorithm | Performs high-quality ICA decomposition, robust to some data imperfections. | Can be run with integrated sample rejection [59]. |
| EEGLAB + clean_rawdata Plugin | Implements ASR and other cleaning functions within a standard EEG analysis environment. | Default ASR parameter (k) of 20-30 is recommended [60]. |
For human-robot interaction research requiring real-time artifact removal:
The combination of ASR followed by ICA (ASRICA) presents a robust and widely accessible pipeline for preprocessing EEG data in movement-intensive HRI studies [58].
This application note investigates a critical challenge in human-robot interaction (HRI): the degradation of emotion recognition accuracy due to signal artifacts. For HRI to be truly effective and natural in real-world settings, robots must reliably infer human emotional states. However, this process is often compromised by artifacts—unwanted noise originating from motion, physiological sources, or instrumentation—that corrupt the biological signals used for affective computing. This study provides a quantitative comparison of emotion recognition performance with and without specialized artifact removal procedures, detailing the experimental protocols and reagent solutions necessary to implement these methods in real-time HRI research.
The implementation of artifact removal protocols leads to substantial improvements in the accuracy of emotion recognition systems. The following table summarizes quantitative findings from key studies, comparing performance with and without artifact removal.
Table 1: Impact of Artifact Removal on Emotion Recognition Accuracy
| Modality | Artifact Removal Method | Classifier / Model | Accuracy Without Removal | Accuracy With Removal | Performance Gain | Citation |
|---|---|---|---|---|---|---|
| EEG | Mutual Information (Epanechnikov Kernel) | SVM with HOC & Hjorth Features | ~68% (estimated) | 80.13% | ~12% | [7] |
| EEG | Mutual Information (Gaussian Kernel) | SVM with HOC & Hjorth Features | ~68% (estimated) | 78.50% | ~10.5% | [7] |
| EEG | SOBI (Baseline Method) | SVM with HOC & Hjorth Features | ~68% (estimated) | 75.70% | ~7.7% | [7] |
| EEG | Real-time Ocular Artifact Removal | Subject-Dependent & Independent Models | Not Reported | Maintained High Accuracy | Enabled Real-Time Operation | [3] |
| Multimodal (Audio, Text, Motion) | N/A (Multimodal Fusion) | Multimodal Fusion Network | N/A | 71.04% (on IEMOCAP) | Outperforms single modalities | [61] |
As the data demonstrates, advanced artifact removal can directly increase classification accuracy by over 10% in EEG-based systems [7]. Furthermore, the choice of algorithm is significant, with the Mutual Information method using an Epanechnikov kernel outperforming both the Gaussian kernel variant and the classical SOBI algorithm [7]. Beyond direct accuracy gains, these methods are essential for transitioning from offline analysis to real-time emotion estimation, a prerequisite for dynamic HRI [3].
For context, multimodal fusion approaches, which integrate several clean signal streams (e.g., audio, text, and motion), currently represent the state-of-the-art, achieving the highest benchmark accuracy on standard datasets [61]. This underscores that artifact removal on individual modalities is a foundational step toward robust multimodal systems.
To ensure reproducibility and facilitate adoption in HRI research, this section outlines detailed protocols for the featured experiments.
This protocol is adapted from the work of Grilo et al. [7], which provides a clear pipeline from raw data to classified emotion.
3.1.1 Objective: To accurately classify human emotions from EEG signals by implementing a novel Blind Source Separation (BSS) algorithm for artifact removal.
3.1.2 Materials: The "Research Reagent Solutions" and key materials required for this protocol are listed below.
Table 2: Research Reagent Solutions for EEG Emotion Recognition
| Item Name | Function / Description | Example / Specification |
|---|---|---|
| EEG Acquisition System | Records brain electrical activity; requires non-invasive, multi-electrode setup. | A system from a university hospital, e.g., 64-channel setup [7]. |
| Emotion Elicitation Stimuli | Standardized audio-visual materials to induce target emotions for ground-truth labeling. | Videos from a standardized database designed to elicit happiness and disgust [7]. |
| Signal Processing Library | Software toolkit for implementing BSS, feature extraction, and classification algorithms. | Custom MATLAB or Python scripts for Mutual Information BSS, HOC, and Hjorth parameters. |
| Self-Assessment Manikin (SAM) | A questionnaire for participants to self-report valence and arousal levels, providing emotion labels. | 9-point pictorial scale based on Russell's circumplex model [7]. |
| Support Vector Machine (SVM) | A supervised machine learning model for classifying extracted features into emotion categories. | SVM with linear or RBF kernel, as implemented in scikit-learn or a similar library. |
3.1.3 Procedure:
The following workflow diagram illustrates the key steps of this protocol:
Diagram 1: Workflow for EEG-based Emotion Recognition with Artifact Removal.
This protocol is optimized for real-time human-robot interaction scenarios, where low latency is critical [3].
3.2.1 Objective: To estimate a user's emotional state from EEG with minimal delay, enabling dynamic robot response.
3.2.2 Key Adaptations for Real-Time Operation:
The streamlined dataflow for this real-time system is shown below:
Diagram 2: Real-time EEG processing workflow for HRI applications.
Successful implementation of the aforementioned protocols requires a suite of specialized tools and datasets. The following table catalogs essential resources for researchers in this field.
Table 3: Essential Research Tools and Datasets for HRI Emotion Recognition
| Category | Item | Specific Use-Case |
|---|---|---|
| Datasets | IEMOCAP (Interactive Emotional Dyadic Motion Capture) | Benchmarking multimodal (audio, visual, text) emotion recognition systems [61]. |
| SEED | Evaluating EEG-based emotion recognition methodologies [3]. | |
| Hardware | Wearable EEG Headsets (e.g., from OpenBCI) | Mobile, real-time acquisition of brain activity for naturalistic HRI studies [42]. |
| RGB-D Cameras (e.g., Microsoft Kinect) | Capturing facial expressions and body kinematics for visual emotion recognition [62]. | |
| Software/Algorithms | MARA (Multiple Artifact Rejection Algorithm) | Automated identification and rejection of artifact components from decomposed EEG signals [7]. |
| CNNs & Transformers | High-accuracy model architectures for facial expression and speech emotion recognition [61] [63]. | |
| Mutual Information BSS | Advanced artifact removal for EEG signals, with Epanechnikov kernel for optimal performance [7]. |
This case study unequivocally demonstrates that dedicated artifact removal is not merely a pre-processing step but a critical determinant of performance in emotion-aware HRI systems. The data shows that advanced techniques like Mutual Information BSS can improve EEG-based emotion recognition accuracy by over 10%, making the difference between a unreliable and a functional system [7]. The provided protocols and toolkit offer a clear pathway for researchers to integrate these methods, thereby enhancing the robustness, realism, and effectiveness of human-robot interactions. Future work will focus on standardizing these protocols across diverse populations and robotic platforms, and on further optimizing algorithms for low-power, real-time operation.
The accurate interpretation of physiological signals is fundamental to advancing human-robot interaction (HRI). A critical methodological consideration in this domain is the choice between subject-dependent and subject-independent paradigms for building computational models. This distinction governs how models are trained, validated, and ultimately deployed in real-world HRI scenarios, where real-time processing and adaptability are paramount. The choice of paradigm directly influences the system's ability to generalize across users and its requirement for individualized calibration data, presenting a key trade-off between performance and practicality. This application note provides a structured comparison of these paradigms, detailing their respective performances, outlining standardized protocols for their implementation, and discussing their specific implications for HRI research involving real-time artifact removal.
The table below synthesizes key quantitative findings from empirical studies comparing subject-dependent and subject-independent model performance across EEG-based emotion recognition, human activity recognition (HAR), and motor imagery (MI) tasks.
Table 1: Comparative Performance of Subject-Dependent vs. Subject-Independent Models
| Study Context | Metric | Subject-Dependent Model Performance | Subject-Independent Model Performance | Citation |
|---|---|---|---|---|
| Human Activity Recognition (HAR) | Relative Performance | Person-Specific Models (PSMs) outperform Person-Independent Models (PIMs) by 43.5% | PIMs outperform PSMs by 55.9% | [64] [65] |
| EEG-based Emotion Recognition | Viability | Achieves high accuracy on the SEED database under real-time constraints | Maintains high accuracy on the SEED database, proving viable for cross-subject applications | [2] [3] |
| EEG Artifact Removal | Approach | Subject-dependent artifact removal (SD-AR) enhances classifier performance, especially in subjects with poor motor imagery skills | Standardized, generalized artifact removal methods are commonly used (e.g., ICA, Surface Laplacian) | [66] |
| Error-Related Potential (ErrP) Detection | Focus | Often yields higher within-subject accuracy | Critical for real-world HRI; focus on developing robust cross-subject classifiers | [26] |
To ensure reproducible and comparable results in HRI research, adhering to standardized experimental protocols for both paradigms is essential. The following sections detail the methodologies for dataset partitioning and model evaluation.
This protocol is designed to assess model performance for a specific individual, maximizing the use of that subject's data.
1. Research Reagent Solutions
2. Step-by-Step Methodology
This protocol evaluates a model's ability to generalize to new, unseen users, which is critical for plug-and-play HRI systems.
1. Research Reagent Solutions
2. Step-by-Step Methodology
The following diagram illustrates the logical workflow and key decision points for selecting and implementing the appropriate modeling paradigm within an HRI research pipeline.
The choice between subject-dependent and subject-independent paradigms profoundly impacts the design of real-time artifact removal systems in HRI.
Subject-Dependent Artifact Removal: In this paradigm, artifact removal can be highly personalized. The SD-AR approach demonstrates that tailoring the preprocessing pipeline (e.g., selectively applying Spatial Laplacian and ICA based on individual user performance) significantly improves Motor Imagery classification, particularly for users with poor skills [66]. This is highly relevant for therapeutic HRI applications where a robot interacts repeatedly with a single patient, and maximum performance for that individual is the priority.
Subject-Independent Artifact Removal: For dynamic HRI where a robot must interact with multiple unknown users, a generalized, subject-independent artifact removal method is mandatory. Techniques like the Mutual Information-based BSS with Epanechnikov kernel [7] or deep learning models like AnEEG [37] are trained on large, diverse datasets to remove common artifacts (EOG, EMG) without requiring user-specific calibration. This enables a "plug-and-play" interaction, which is essential for robots in public spaces or industrial settings with rotating shifts.
Furthermore, for implicit communication channels like Error-Related Potentials (ErrPs)—where a robot must detect a user's perceived error without explicit command—developing robust subject-independent classifiers is a primary focus. This ensures the robot can adapt to new users immediately, a cornerstone of fluid and intuitive HRI [26].
The evaluation of subject-dependent versus subject-independent paradigms reveals a fundamental trade-off between personalized accuracy and broad generalization. Subject-dependent models, exemplified by PSMs, offer superior performance for individual users, making them ideal for dedicated assistive or rehabilitative HRI. Conversely, subject-independent models, or PIMs, provide the generalization necessary for scalable, multi-user applications. The decision framework and standardized protocols provided here offer researchers a clear pathway for selecting, implementing, and evaluating the appropriate paradigm. As HRI evolves, hybrid approaches—such as initial subject-independent models that gradually personalize over time—present a promising avenue for future research, aiming to bridge the performance gap while maintaining the practicality essential for real-world deployment.
Affective human-robot interaction (HRI) requires robust and real-time analysis of user states, often leveraging electroencephalography (EEG) due to its high temporal resolution and non-invasive nature [3] [67]. However, the practical deployment of these systems is severely hampered by a critical, unsolved problem: the effective and standardized removal of artifacts from EEG signals in real-time scenarios. Artifacts—unwanted noise originating from ocular movements, muscle activity, or environmental sources—corrupt the neural signal, leading to misinterpretation of brain patterns and unreliable robot control or emotion estimation [3] [37] [19]. While numerous artifact removal techniques exist, the field is characterized by a lack of standardization, inconsistent performance reporting, and significant methodological gaps, particularly for online, closed-loop HRI systems. This document outlines the primary gaps in the literature and the associated standardization challenges, providing application notes and experimental protocols to guide future research in this critical area.
A systematic analysis of current research reveals several interconnected gaps that impede the development of robust real-time artifact removal pipelines for HRI.
Table 1: Key Gaps in Real-Time EEG Artifact Removal for HRI
| Gap Category | Specific Challenge | Implication for HRI |
|---|---|---|
| Algorithm Selection & Validation | Lack of consensus on optimal algorithms for real-time, low-channel count systems [31]. | Pipelines optimized for high-density lab EEG may fail with wearable HRI setups. |
| Real-Time Performance & Computational Cost | Inadequate reporting of processing latency and computational demands [3]. | Methods are not evaluable for true real-time (low-latency) application on edge devices. |
| Generalizability Across Contexts | Poor cross-session and cross-subject performance of artifact removal methods [3]. | Models require frequent recalibration, disrupting seamless interaction. |
| Artefact-Specific Pipeline Optimization | Rare use of tailored pipelines for specific artifact categories (ocular, motion, etc.) [31]. | A one-size-fits-all approach reduces efficiency and risks removing neural information. |
| Integration with Downstream HRI Tasks | Disconnect between artifact removal performance and its impact on final task accuracy (e.g., emotion classification, robot control) [7]. | High artifact removal metrics do not guarantee improved HRI performance. |
The shift towards wearable EEG with dry electrodes and low channel counts creates specific problems. Techniques like Independent Component Analysis (ICA), a staple in conventional EEG processing, are less effective with reduced spatial information [31]. A recent review highlights that only a few studies explicitly address the peculiarities of artifacts in wearable systems, which differ from those in high-density lab settings [31]. Furthermore, auxiliary sensors (e.g., IMUs) that could enhance artifact detection under ecological conditions are still significantly underutilized [31].
For HRI, "real-time" implies a strict upper bound on processing latency. Many studies propose advanced methods, including deep learning models like Generative Adversarial Networks (GANs) with LSTM layers [37] or mutual information-based Blind Source Separation (BSS) [7], but fail to quantify their processing time and computational footprint adequately. Without this data, it is impossible to judge their suitability for a closed-loop interaction running on portable hardware.
The absence of common standards makes it difficult to compare methods and reproduce results, slowing down collective progress.
Table 2: Standardization Challenges in Artifact Removal Research
| Standardization Area | Current Status | Recommended Direction |
|---|---|---|
| Performance Metrics | Inconsistent use of metrics (e.g., NMSE, RMSE, CC, SNR, SAR) [37]. | Mandatory reporting of a core set of metrics, including task-relevant classification accuracy. |
| Public Datasets | Scarcity of public, high-quality datasets containing motion artifacts from HRI-relevant scenarios [31]. | Community effort to create and share benchmark datasets with various artifacts and HRI tasks. |
| Reporting of Computational Cost | Rarely reported [3]. | Mandatory reporting of latency (per sample/epoch) and computational load. |
| Validation Paradigms | Over-reliance on offline, within-subject validation [19]. | Promotion of online, subject-independent, and cross-session validation protocols. |
| Definition of "Clean" Signal | No gold standard for ground truth in real data [37]. | Clear documentation of the method used to establish the reference signal (e.g., expert annotation, semi-simulation). |
To address these gaps, researchers should adopt rigorous and standardized experimental protocols.
This protocol is designed for the comparative evaluation of different artifact removal algorithms in an HRI context.
Objective: To quantitatively compare the performance and real-time capability of multiple artifact removal algorithms (e.g., ICA, ASR, wavelet-based, deep learning) using a shared dataset and evaluation framework.
Materials and Reagents:
Procedure:
This protocol tests the integrated artifact removal and HRI control system in an online, closed-loop setting.
Objective: To validate the performance of a complete real-time BCI pipeline, including artifact removal, for a specific HRI task like robotic hand control [15].
Procedure:
The following diagrams illustrate the core workflows and relationships discussed in this document.
This diagram outlines the complete data flow from signal acquisition to robot interaction, highlighting the critical role of artifact removal.
This flowchart provides a logical framework for selecting and validating an artifact removal method based on HRI requirements.
Table 3: Key Research Reagent Solutions for Real-Time EEG Artifact Removal
| Item / Technique | Function in Research | Application Notes |
|---|---|---|
| Independent Component Analysis (ICA) [3] [67] | Blind source separation to identify and remove artifact-related components. | Powerful but computationally heavy; less effective with low-channel counts; often used as a benchmark. |
| Automatic Subspace Reconstruction (ASR) [31] | Statistical method for removing high-variance, non-stationary artifacts in real-time. | Suited for online use; works well with wearable EEG; requires parameter tuning. |
| Wavelet Transform [31] | Time-frequency decomposition to isolate and remove artifacts in specific frequency bands. | Effective for transient artifacts like eye blinks; offers a good balance of performance and speed. |
| Generative Adversarial Networks (GANs) [37] | Deep learning model to generate artifact-free EEG signals from noisy inputs. | State-of-the-art performance; requires large datasets for training and significant computational resources. |
| Mutual Information-based BSS [7] | Blind source separation using mutual information with kernels (e.g., Epanechnikov) to maximize independence. | Emerging method; shown to be effective with lower computational cost than some alternatives. |
| Recursive Least Squares (RLS) [37] | Adaptive filtering technique for noise cancellation, often using a reference signal. | Very low latency and computational cost; ideal for edge computing; requires a reference signal. |
Real-time artifact removal is a foundational pillar for enabling robust and intuitive Human-Robot Interaction. This synthesis demonstrates that while established techniques like ICA and emerging methods like ASR and iCanClean provide powerful solutions, no single algorithm is universally superior. Success hinges on a carefully optimized pipeline tailored to specific HRI tasks, whether for affective computing in healthcare or industrial collaboration. Future progress will be driven by deep learning models adaptable to individual users, the strategic integration of auxiliary sensors, and a stronger focus on standardizing validation protocols for wearable systems. For biomedical research, these advancements promise more reliable brain-computer interfaces for therapeutic applications, paving the way for HRI systems that can genuinely understand and adapt to human cognitive and emotional states in real-time.