Eye Control By Microsoft
How can Windows empower every person and every
organization on the planet to achieve more? Our ongoing mission is
behind a feature that Windows Insiders voted as one of their top 10
favorites: Eye Control.
While Eye Control was first released in-box
during the Fall Creators Update, it’s seen some great improvements with
the April 2018 Update. We sat down with Microsoft engineer Jake Cohen to
get the story behind the accessibility feature that enables users to
control Windows with their eyes and the help of a compatible
eye-tracking device (such as Tobii and EyeTech).
“Accessibility has been super important for us
for the past 20-plus years,” Jake explained. “For the past few years,
we’ve working hard to really aspire towards our mission statement by
empowering every person of every level of ability.”
The Windows team started on the road to
developing Eye Control during the 2014 Microsoft company-wide hackathon,
when Steve Gleason, an NFL football player who had played with the New
Orleans Saints, emailed Microsoft with a challenge.
Living with ALS, Steve wanted to spark technology that could help
people with ALS and that would help him communicate more easily, play
with his son, and move his wheelchair independently.
Since the hackathon, Microsoft has been working
closely with Team Gleason, the nonprofit foundation founded by Steve,
to develop technologies that will empower people living with ALS.
Continuing this work, the Windows team has been steadily implementing
built-in eye tracking to Windows 10.
With a compatible device, Eye Control
leverages infrared lighting and cameras to detect where a user’s eyes
are looking relative to the screen. Windows takes that information and
enables a user to control a mouse or keyboard.
“Eye Control starts with a launch pad, which is
UI that’s always present on the screen,” Jake said. “When you dwell
your eyes on an icon, which is the act of fixating your eyes somewhere
on the screen and waiting, it’ll activate a click. So it’s basically a
press and hold with your eyes.”
“You have access on the launch pad to the
mouse, the keyboard, text-to-speech, and now in the April 2018 release,
many more options like Quick Access, Start, Task View, Device
Calibration, Settings, and more. For browsing the Web or scrolling an
app, you can also fixate your eyes somewhere on the screen and then use
the arrows that are provided to scroll up and down using your eyes.”
What’s in store for the future of Eye Control?
Jake’s team is continuing to work with Microsoft Research and Team
Gleason to collect feedback from the ALS community and make Eye Control
even more useful.
“It’s really inspiring to get this feedback
because we hear people say, ‘This is amazing technology. This is really
helping me,’” Jake said. “And also, ‘This is the next thing I need.’
It’s about empowering them to do everything they can think of, not just a
subset of interactions or abilities.”
Jake also mentioned that sparking more third-party tools is a priority for the team.
“The next step we’re taking is releasing public
developer APIs and open-source libraries to allow third-party
developers to build apps and experiences that can leverage eye
tracking,” Jake said.
“Imagine all of the gaps that third-party
developers can fill for customers who are living with mobility
impairments. It comes down to Microsoft’s core roots. We can’t fulfill
this mission statement alone to empower everyone–we have to build a
platform that empowers everyone to empower other people. I’m excited see
what developers can come up with in order to make an impact.”
(Source: windows.com )
(Source: windows.com )
Comments
Post a Comment