Ajay Vikram P

Ajay Vikram P
github | linkedin | medium

Decoding Hand Kinematics in Brain-Computer Interfaces with State-Space Models

Overview

This project focused on decoding continuous hand movements from brain activity. While most BCI models classify discrete movement intentions, we tackled the more complex task of predicting continuous hand trajectories — essential for fine motor control in real-world applications.


Dataset & Task

We used the Nonhuman Primate Reaching with Multichannel Sensorimotor Cortex Electrophysiology dataset, which contains spike recordings from two macaques (Indy and Loco) reaching for targets arranged in an 8×8 grid.

Note: The above link is for reference, the actual loading and processing are done using the neurobench code harness


Models Trained

We trained 2 State-Space Models on this dataset:

  1. Legendre Memory Units (LMUs)
  2. Structured State Space Sequence Models (S4)

Refer to my blog post on State-Space Models if you are new to the topic.


Model Architecture

Note: The models were trained with a sub window binning method on the spikes.


Results

Model Parameters Test R2 Score
ANN 2D 5K 0.62
ANN 3D 24K 0.65
LSTM 44K 0.58
EEGNet 11K 0.56
LMU 7K 0.70
S4 70K 0.75

*The results shown above are for the 3rd session of Monkey 1

Back