P300 BCI Robotic Control

A novel brain-computer interface for robotic arm control using P300 potentials from visual stimulation

97% Recognition Accuracy
2-DoF Robotic Control
Real-time Neural Processing
BCI Robotic Control System Diagram
Figure 1: The procedure of controlling a robot by a BCI system
Completed

Brain-Computer Interface for Robotic Control

Novel control algorithm for a 2-DoF robotic arm using a P300-based brain-computer interface

RECOGNITION ACCURACY 97%
PUBLISHED 2019
TECHNOLOGY EEG + P300 + Robotics

Project Overview

This research developed a novel control algorithm based on a P300-based brain-computer interface (BCI) to control a 2-DoF robotic arm. The system allows users to control robotic movement through brain signals alone, providing a potential pathway for assistive technology for patients with motor impairments such as Amyotrophic Lateral Sclerosis (ALS) and Spinal Cord Injury (SCI).

Eight subjects (five men and three women) performed a 2-dimensional target tracking task in a simulated environment. Their EEG signals from the visual cortex were recorded, and P300 components were extracted and evaluated to deliver a real-time BCI-based controller. The volunteer's intention was recognized and decoded as an appropriate command to control the cursor, which was then used to control a simulated robotic arm in a 2-dimensional space.

Key Achievement

The system successfully demonstrated point-to-point control of the robotic arm in a 2-dimensional space with high reliability. The best classification results were obtained with a multi-classifier solution with a recognition rate of 97 percent, without requiring channel selection before classification.

๐Ÿง  P300 Neural Detection

Robust detection of P300 event-related potentials from EEG signals using optimized processing and classification algorithms.

๐Ÿค– Robotic Arm Control

2-DoF robotic arm control with precise movement capabilities in a 2D plane using inverse dynamics control algorithm.

โšก Real-Time Processing

End-to-end real-time processing from EEG signal acquisition through feature extraction, classification, and robotic control.

Methodology

System Architecture

The BCI system was implemented in four key steps:

  1. Visual Stimulus Design: A custom GUI with four bulbs representing directions (up, down, left, right) that flash randomly to elicit P300 responses
  2. EEG Signal Processing: Recording brain signals from the visual cortex followed by filtering and feature extraction
  3. Classification: Using K-nearest neighbors (KNN) and support vector machine (SVM) classifiers to interpret brain signals
  4. Robotic Control: Converting classified brain signals into movement commands for the robotic arm
BCI Graphical User Interface
Visual Stimulus GUI: Four bulbs representing directional controls flash randomly to elicit P300 responses from the subject.
2-DoF Robotic Arm
2-DoF Robotic Arm: The simulated robotic arm with two revolute joints for movement in a 2D plane.

Feature Extraction and Classification

The signal processing pipeline included:

Feature Extraction Process
Feature Extraction Process: The pipeline for extracting meaningful features from raw EEG signals, including filtering, wavelet transformation, and dimensionality reduction.

Robotic Control System

The robotic arm was controlled using an inverse dynamics control method:

Results & Applications

Classification Results
Classification Results: Comparison of different classifiers showing KNN with 3 neighbors achieving the highest recognition rate of 97%.

Drawing Task Demonstration

To validate the system's practical application, subjects were asked to draw shapes using the BCI-controlled robotic arm:

  1. The subject imagines the desired shape and the required movements
  2. For each movement, the subject focuses on the corresponding direction's bulb in the BCI panel
  3. The BCI system detects the P300 response and translates it into a direction command
  4. The robotic arm moves the end-effector (with attached marker) accordingly
  5. By sequentially focusing on different directions, complete shapes can be drawn
Drawing Task Results
Drawing Task Results: Sample of a letter 'G' drawn by a subject using the BCI-controlled robotic arm. The drawing demonstrates the system's accuracy in following directional commands.

Original Publication

Garakani, G., Ghane, H., & Menhaj, M. B. (2019). Control of a 2-DoF robotic arm using a P300-based brain-computer interface. AUT Journal of Modeling and Simulation, 51(2), 153-162. DOI: 10.22060/miscj.2019.15569.5136

View Full Paper

Impact and Applications

This research demonstrates the potential for BCI systems to provide intuitive control mechanisms for robotic devices, particularly for:

๐Ÿ‘จโ€โš•๏ธ Assistive Technology

Helping patients with motor impairments like ALS and SCI control robotic devices for improved independence and quality of life.

๐Ÿฅ Rehabilitation

Supporting rehabilitation processes for those recovering from stroke or other neurological conditions affecting motor function.

๐Ÿ”ฌ Industrial Applications

Hands-free control for robots in specialized environments where manual control is difficult or unsafe.

๐Ÿฆพ Prosthetics

Foundation for more advanced neural control systems for next-generation prosthetic limbs.

Interested in Collaboration?

This research demonstrates the potential of BCI technology for robotic control applications. If you're interested in brain-computer interfaces or need a neural control system for assistive technology, let's discuss how we can work together.

๐Ÿ“„ View CV ๐Ÿ’ผ Connect on LinkedIn