Both these projects were motivated by earlier work from 2011 which demnostrated eye-gaze controlled navigation working with UG students in BME and EE at CCNY.
In both instances the goal was to move the the point of gaze to the center of the screen, simply put "where you look is where it goes".
This is the popular AR.Drone quadricopter controlled with this policy using an EyeLink 2000 eye-tracker. The drone pivots left and right when the operator looks into those directions. It tilts up and down (making it fly backwards and forwards) when the operator looks up or down. Simple, isn't it?
EyeDrone was implemented by Lucas C Parra in C++ with much help from Michael Quintian.
The chin rest here is not really needed as the basic control algorithm is inherently stable and miscalibrations are of little concern. When we first tried this we had not realized how stable it was.
John M. Ettikkalayil, MS student in BME, reimplemented this using an open source sourftware. This solution is much less expensive as one can use a simple grayscale camera. A full description of the system is on the
neural engineering lab site:
This is an image map scrolling application implemented in matlab using psychtoolbox using a plugin for the EyeLink eye-tracker. The object being looked at is moved gradually to the center of the screen. A bit of logic was needed to scroll smoothly.
EyeMap was implemented in matlab by Mohammod Arafat with help of Krishan Mistry and Lucas C Parra.
If you want to get involved or learn more about this project contact Lucas C Parra.