Gaze tracking can be a powerful tool to control a computer or robotic limb. There are many gaze tracking technologies, but vision based gaze trackers, which work off of a standard web cam, have the potential to be quickly accessible. This report assumes that noisy estimates of the head position and gaze direction are known, and then presents a method to find a more stable estimate of the user's gaze on a table plane. The method works by combining estimates of each eye's gaze intersection with the viewing plane. The basic combination achieves fair precision but poor accuracy, so the report describes a simple way to improve accuracy while maintaining good aspects of the system. The system achieves an accuracy of 6.9 cm in the X direction and 5.7 cm in the Y direction for a gaze point experiment.
Title
Gaze Tracking in the Plane
Published
2019-05-17
Full Collection Name
Electrical Engineering & Computer Sciences Technical Reports
Other Identifiers
EECS-2019-73
Type
Text
Extent
13 p
Archive
The Engineering Library
Usage Statement
Researchers may make free and open use of the UC Berkeley Library’s digitized public domain materials. However, some materials in our online collections may be protected by U.S. copyright law (Title 17, U.S.C.). Use or reproduction of materials protected by copyright beyond that allowed by fair use (Title 17, U.S.C. § 107) requires permission from the copyright owners. The use or reproduction of some materials may also be restricted by terms of University of California gift or purchase agreements, privacy and publicity rights, or trademark law. Responsibility for determining rights status and permissibility of any use or reproduction rests exclusively with the researcher. To learn more or make inquiries, please see our permissions policies (https://www.lib.berkeley.edu/about/permissions-policies).