Structured generative models for real-time user intent embedding and control-space customization
Andrew Thompson, Fiona Neylon, Brenna Argall | Northwestern University + Shirley Ryan AbilityLab
This project explores the use of variational autoencoders (VAEs) to learn structured latent control spaces from body-machine interface (BoMI) data. The goal is to enable intuitive, low-dimensional user intent embeddings that generalize across users and time, and to support robust control of high-dimensional robotic systems from sparse or noisy human input.
Unlike PCA-based mappings used during deployment, the SSVAE-based approach is generative and probabilistic. It allows for intent inference under uncertainty, meaningful and continuous interpolation across embeddings, and latent-level customization—capabilities that support safer and more generalizable assistive control.
Experiment | Goal | Outcome |
---|---|---|
Latent stability across sessions | Assess how latent axes shift with session index | Axes remain consistent across 3+ sessions with single-user training |
Cross-user generalization | Train on one user, test on another | Latents retain structure, but decoder needs fine-tuning |
Control-space mapping | Latent → 6-DOF robot control mapping | Mapping feasible with linear decoder or MLP; interpretable axes emerge |
Few-shot retraining | Fine-tune encoder with small new-user dataset | Rapid convergence observed with 20–50 examples |
Additionally, ablation studies were on all of the additional cost terms to see their effect.
This work builds directly on the PCA-based BoMI system used in a 190+ session longitudinal study, where we gathered data from and evaluated teleoperation performance from a cohort of 10 individuals with cervical spinal cord injuries (cSCI). The SSVAE model was trained entirely on data collected during those sessions. Whereas PCA was used for control deployment, the SSVAE supports:
Preprint (In Review):
Structured Semi-Supervised Generative Methods for Learning Robust Control Embeddings from Human Motion Data
Andrew Thompson, Fiona Neylon, Brenna Argall
All models and training scripts are currently housed in private repositories. Code excerpts, latent visualizations, and sanitized motion samples are available upon request.