• Login
    View Item 
    •   YODA Home
    • CSEM Archive
    • Research Publications
    • View Item
    •   YODA Home
    • CSEM Archive
    • Research Publications
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Real Time Eye Gaze Tracking for Human Machine Interaction in the Cockpit

    Thumbnail
    Author
    Türetken, Engin; Saeedi, Sareh; Bigdeli, Siavash; Stadelmann, Patrick; Cantale, Nicolas; Lutnyk, Luis; Raubal, Martin; Dunbar, L. Andrea
    Metadata
    Show full item record
    Abstract
    The Aeronautics industry has pioneered safety from digital checklists to moving maps that improve pilot situational awareness and support safe ground movements. Today, pilots deal with increasingly complex cockpit environments and air traffic densification. Here we present an intelligent vision system, which allows real-time human-machine interaction in the cockpits to reduce pilot’s workload. The challenges for such a vision system include extreme change in background light intensity, large field-of-view and variable working distances. Adapted hardware, use of state-of-the-art computer vision techniques and machine learning algorithms in eye gaze detection allow a smooth, and accurate real-time feedback system. The current system has been over-specified to explore the optimized solutions for different use-cases. The algorithmic pipeline for eye gaze tracking was developed and iteratively optimized to obtain the speed and accuracy required for the aviation use cases. The pipeline, which is a combination of data-driven and analytics approaches, runs in real time at 60 fps with a latency of about 32ms. The eye gaze estimation error was evaluated in terms of the point of regard distance error with respect to the 3D point location. An average error of less than 1.1cm was achieved over 28 gaze points representing the cockpit instruments placed at about 80-110cm from the participants’ eyes. The angular gaze deviation goes down to less than 1° for the panels towards which an accurate eye gaze was required according to the use cases.
    Publication Reference
    Engin Türetkin, Sareh Saeedi, Siavash Bigdeli, Patrick Stadelmann, Nicolas Cantale, Luis Lutnyk, Martin Raubal, Andrea L. Dunbar, "Real time eye gaze tracking for human machine interaction in the cockpit," Proc. SPIE 12019, AI and Optical Data Sciences III, 1201904 (2 March 2022)
    Year
    2022-03-02
    Sponsors
    PEGGASUS project has received funding from the Clean Sky 2 Joint Undertaking under the European Union’s Horizon 2020 research and innovation program under grant agreement No. 821461.
    URI
    https://yoda.csem.ch/handle/20.500.12839/1025
    Collections
    • Research Publications

    Browse

    All of YODACommunities & CollectionsBy Issue DateAuthorsTitlesResearch AreasBusiness UnitsThis CollectionBy Issue DateAuthorsTitlesResearch AreasBusiness Units

    My Account

    Login

    DSpace software copyright © 2002-2022  DuraSpace
    Contact Us | Send Feedback
    DSpace Express is a service operated by 
    Atmire NV