Using qualitative eye-tracking data to inform audio presentation of dynamic Web content

    Research output: Contribution to journalArticlepeer-review


    Presenting Web content through screen readers can be a challenging task, but this is the only means of access for many blind and visually impaired users. The difficulties are more acute when the information forms part of an interactive process, such as the increasingly common Web 2.0 applications. If the process is to be completed correctly and efficiently it is vital that appropriate information is given to the user at an appropriate time. Designing a non-visual interface that achieves these aims is a non-trivial task, for which several approaches are possible. The one taken here is to use eye-tracking to understand how sighted users interact with the content, and to gain insight into how they benefit from the information, then apply this understanding to design a non-visual user interface. This paper describes how this technique was applied to develop audio interfaces for two common types of interactionauto-suggest lists and pop-up calendars. Although the resulting interfaces were quite different, one largely mirroring the visual representation and the other not, evaluations showed that the approach was effective, with both audio implementations effective and popular with participants. © 2010 Taylor & Francis.
    Original languageEnglish
    Pages (from-to)281-301
    Number of pages20
    JournalNew Review of Hypermedia and Multimedia
    Issue number3
    Publication statusPublished - Dec 2010


    • Accessibility
    • AJAX
    • Eye-Tracking
    • Human Computer Interaction
    • Methodology
    • Visual Disability
    • Web 2.1


    Dive into the research topics of 'Using qualitative eye-tracking data to inform audio presentation of dynamic Web content'. Together they form a unique fingerprint.

    Cite this