Search

Your Eyes Control Your Smartphone Via New Gaze-Tracking Tool - Carnegie Mellon University

https://ift.tt/CoziOFH

Software that tracks the eyes with precision can solve this problem. Andy Kong, a senior majoring in computer science, had been interested in eye-tracking technologies since he first came to CMU. He found commercial versions pricey, so he wrote a program that used a laptop's built-in camera to track the user's eyes, which in turn moved the cursor around the screen — an important early step toward EyeMU.

"Current phones only respond when we ask them for things, whether by speech, taps or button clicks," Kong said. "If the phone is widely used now, imagine how much more useful it would be if we could predict what the user wanted by analyzing gaze or other biometrics."

It wasn't easy to streamline the package so it could work at speed on a smartphone.

"That's a resource constraint. You must make sure your algorithms are fast enough," Ahuja said. "If it takes too long, your eye will move along."

Kong, the paper's lead author, presented the team's findings with Ahuja, Harrison and Assistant Professor of HCII Mayank Goel at last year's International Conference on Multimodal Interaction. Having a peer-reviewed paper accepted to a major conference was a huge achievement for Kong, an undergraduate researcher.

Kong and Ahuja advanced that early prototype by using Google's Face Mesh tool to study the gaze patterns of users looking at different areas of the screen and render the mapping data. Next, the team developed a gaze predictor that uses the smartphone's front-facing camera to lock in what the viewer is looking at and register it as the target.

The team made the tool more productive by combining the gaze predictor with the smartphone's built-in motion sensors to enable commands. For example, a user could look at a notification long enough to secure it as a target and flick the phone to the left to dismiss it or to the right to respond to the notification. Similarly, a user might pull the phone closer to enlarge an image or move the phone away to disengage the gaze control, all while holding a large latte in the other hand.

"The big tech companies like Google and Apple have gotten pretty close with gaze prediction, but just staring at something alone doesn't get you there," Harrison said. "The real innovation in this project is the addition of a second modality, such as flicking the phone left or right, combined with gaze prediction. That's what makes it powerful. It seems so obvious in retrospect, but it's a clever idea that makes EyeMU much more intuitive."

Adblock test (Why?)



"Smartphone" - Google News
April 23, 2022 at 02:48AM
https://ift.tt/JI2kYyc

Your Eyes Control Your Smartphone Via New Gaze-Tracking Tool - Carnegie Mellon University
"Smartphone" - Google News
https://ift.tt/K87hUtk
https://ift.tt/CoziOFH

Bagikan Berita Ini

0 Response to "Your Eyes Control Your Smartphone Via New Gaze-Tracking Tool - Carnegie Mellon University"

Post a Comment


Powered by Blogger.