MWC 2014 - Throw fruit and turn Kindle pages using only your eyes (eye-tracking for 5 dollars)
**Integrating an eye tracker into the hardware**
The Eye Tribe -- First look (Wired UK)
The Eye Tribe released its first eye-tracking product to developers in December -- a long, thin $99 module that attaches to a Windows laptop, computer or tablet. It sold out immediately and the company is now working on a second batch. But it also has a more exciting proposition in the pipeline -- a software development kit module for Android phones that it eventually wants to see integrated into the a wide range of mobile devices.
Everything from simple scrolling, to flicking over the pages of a Kindle to playing games like Fruit Ninja -- as demoed to Wired.co.uk -- could be controlled using only your eyes.
Most of the requisite hardware is already built into phones. The Eye Tribe just needs to persuade companies to integrate the technology.
All that's required is a camera sensor with infrared capabilities. "What we know is that in Q4 this year, sensors are coming out that can switch between regular camera and infrared camera."
“OEM vendors could likely add this sensor to their handsets for just five dollars”
If modifying the device to add eye-tracking only adds 5 dollars to the manufacturing cost, then I’m sure that at least one of the smartphone, tablet, notebook, and laptop manufacturers will make the supposedly easy camera modification.
**See before touch**
I think that most of the time, a person will see a widget that they want to touch before they actually reach out, and physically touch it
(The only times where I’m not looking is when I press the Android Navigation Bar buttons that are near the bottom edge of the screen. Although, on a larger Nexus 10, I usually have to look at them first).
**Eyes + consecutively touching the same few buttons**
On certain tasks, it might be convenient and fast to have the option of touching “single tap where I’m looking”, and “swipe up where I’m looking” buttons. You would only need one or two buttons that are close to you (kind of like the Navigation Bar buttons at the bottom).
Look, touch an easy-to-reach spot, look, and then touch the same button again. You don’t have to keep changing your hand and finger positions between each tap.
“Looking at icons on a desktop instantly highlights them, and you can then tap anywhere on the screen to open up the selected app.”
I guess that in one of their demos, they temporarily made the entire screen a “single tap where I’m looking” button.
Besides the three default buttons in the Navigation Bar, you could add “single tap where I’m looking”, and “swipe up where I’m looking” (to perhaps simulate a Page Down for reading) buttons, and those alone should allow you to do a lot of things).
If you have a vertically propped up tablet with an external keyboard, you could remap a keyboard button to be the “tap where I’m looking” button.
Even without the above options, I still think the ability to have a page automatically scroll down when your eyes reach the bottom of the page, or have an e-book automatically turn the page when the gaze reaches the corner of the text would be pretty good features to have. They would be especially handy for computer interaction while cooking and eating, and interacting with a vertically set up touch device that is more than an arms-length away while doing other stuff on the desktop.