Title: Automatically Identifying Targets Users Interact with During Real World Tasks
Author: Amy Hurst, Scott E. Hudson, Jennifer Mankoff
Publisher: IUI '10, February 7-10, 2010 Hong Kong
Summary
Information about location and size of targets on a screen has been a popular area of research. The ability to analyze user actions in that environment is essential for answering questions about usability, performance and daily use. Knowing the target size and location is necessary in assessing the pointing performance of impaired individuals.
The researchers in this article attempted to develop a slightly new Accessibility API. They relied on Microsoft's existing API (MSAA API) combined with a hybrid solution of their own that relied on machine learning and computer vision. They found that their hybrid approach resulted in a 75% success rate. They looked at 8 popular applications: MS Outlook, web browsers, MS Word, Windows Explorer, Media Player and MS PowerPoint.
The applications of their work include improved computer accessibility, support for automatic extraction of a task sequence, and automatically scripting common actions. CRUMBS was used to capture information about the interaction. This is currently only limited to Microsoft machines.
Discussion
I'm glad they included pictures in this because I don't think I would've been able to understand it otherwise. That said, the pictures really help to illustrate their computer vision and machine learning algorithms. I didn't understand a ton of what they talked about but I did understand their general idea.
Summary
Information about location and size of targets on a screen has been a popular area of research. The ability to analyze user actions in that environment is essential for answering questions about usability, performance and daily use. Knowing the target size and location is necessary in assessing the pointing performance of impaired individuals.
The researchers in this article attempted to develop a slightly new Accessibility API. They relied on Microsoft's existing API (MSAA API) combined with a hybrid solution of their own that relied on machine learning and computer vision. They found that their hybrid approach resulted in a 75% success rate. They looked at 8 popular applications: MS Outlook, web browsers, MS Word, Windows Explorer, Media Player and MS PowerPoint.
The applications of their work include improved computer accessibility, support for automatic extraction of a task sequence, and automatically scripting common actions. CRUMBS was used to capture information about the interaction. This is currently only limited to Microsoft machines.
Discussion
I'm glad they included pictures in this because I don't think I would've been able to understand it otherwise. That said, the pictures really help to illustrate their computer vision and machine learning algorithms. I didn't understand a ton of what they talked about but I did understand their general idea.
I am also glad they put in pictures, made the article more interesting
ReplyDeleteThis research could be extremely important when trying to conserve screen space. With the advent of smaller screens on netbooks and laptop, it is important to determine the correct size for targets to be.
ReplyDelete