Robotics and Mixed Reality
Our research efforts focus on developing methods to better localize robots and humans together, and approaches for leveraging spatial intelligence and the sensing capabilities of Mixed Reality devices to enable easier, safer, and more effective…
Event-VAE-RL
Representation Learning and Reinforcement Learning from Event Cameras This repository will provide a codebase to train and evaluate models as seen in this paper: Representation Learning for Event-based Visuomotor Policies Sai Vemprala, Sami Mian, Ashish…
Eye Control
By taking a broad view of the component parts of eye tracking – hardware, driver interface, API surface, AAC and computer control – we can find new ways to empower people who need gaze-tracking technology…
Project DeepEyes
Project DeepEyes aims to develop no cost and high-quality eye tracking for every person with motor-neuron disabilities.