Publication:
Gaze-based prediction of pen-based virtual interaction tasks

Placeholder

Departments

School / College / Institute

Program

KU Authors

Co-Authors

Publication Date

Language

Embargo Status

Journal Title

Journal ISSN

Volume Title

Alternative Title

Abstract

In typical human-computer interaction, users convey their intentions through traditional input devices (e.g. keyboards, mice, joysticks) coupled with standard graphical user interface elements. Recently, pen-based interaction has emerged as a more intuitive alternative to these traditional means. However, existing pen-based systems are limited by the fact that they rely heavily on auxiliary mode switching mechanisms during interaction (e.g. hard or soft modifier keys, buttons, menus). In this paper, we describe how eye gaze movements that naturally occur during pen-based interaction can be used to reduce dependency on explicit mode selection mechanisms in pen-based systems. In particular, we show that a range of virtual manipulation commands, that would otherwise require auxiliary mode switching elements, can be issued with an 88% success rate with the aid of users' natural eye gaze behavior during pen-only interaction. (C) 2014 Elsevier Ltd. All rights reserved.

Source

Publisher

Academic Press Ltd- Elsevier Science Ltd

Subject

Computer science, Cybernetics, Ergonomics, Psychology

Citation

Has Part

Source

International Journal of Human-Computer Studies

Book Series Title

Edition

DOI

10.1016/j.ijhcs.2014.09.005

item.page.datauri

Link

Rights

Copyrights Note

Endorsement

Review

Supplemented By

Referenced By

0

Views

0

Downloads

View PlumX Details