Implementation of Eye Movement-based Digital Healthcare Communication Board: Accurate Gaze Estimation Using Triplet Loss Function
DOI:
https://doi.org/10.47611/jsrhs.v13i4.8372Keywords:
Gaze Estimation, Machine LearningAbstract
Effective communication is important in healthcare, especially for quadriplegia patients who often face barriers in expressing their needs. Traditionally, these patients rely on healthcare communication boards to interact with doctors and nurses. However, this method is often slow, cumbersome, and requires the assistance of another person which makes it an inconvenient solution. Recent advancements in gaze estimation technology, which predicts the direction of an individual's eye movements, can provide a promising alternative. This research examines the potential of gaze estimation to develop a digital healthcare communication board driven by eye movements. Such a system would empower quadriplegia patients to communicate independently which enhances both the speed and efficiency of interactions. The proposed system processes eye images to predict a gaze vector which represents the direction in which an individual is currently looking. To improve the system's accuracy, we introduce a conjugate ability-based loss function. Additionally, the proposed approach was applied to a digital healthcare system to demonstrate its feasibility and effectiveness in real-world scenarios. The system achieved an angular error of 8.7 on a public gaze estimation dataset which surpasses previous state-of-the-art methods.
Downloads
References or Bibliography
Abdelrahman, A. A., Hempel, T., Khalifa, A., Al-Hamadi, A., & Dinges, L. (2023, October). L2cs-net: Fine-grained gaze estimation in unconstrained environments. In 2023 8th International Conference on Frontiers of Signal Processing (ICFSP) (pp. 98-102). IEEE.
Autodesk. (2024, Aug 27). “Communication Board for Individuals With Disabilities”: Autodesk.
https://www.instructables.com/Communication-Board-for-Individuals-with-Disabilit/
Bonazzi, P., Rüegg, T., Bian, S., Li, Y., & Magno, M. (2023, October). Tinytracker: Ultra-fast and ultra-low-power edge vision in-sensor for gaze estimation. In 2023 IEEE SENSORS (pp. 1-4). IEEE.
Cleveland Clinic. (2024, Aug 27). “Quadriplegia”: Cleveland Clinic.
https://my.clevelandclinic.org/health/symptoms/23974-quadriplegia-tetraplegia
Fischer, T., Chang, H. J., & Demiris, Y. (2018). Rt-gene: Real-time eye gaze estimation in natural environments. In Proceedings of the European conference on computer vision (ECCV) (pp. 334-352).
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778). https://doi.org/10.48550/arXiv.1512.03385
Krafka, K., Khosla, A., Kellnhofer, P., Kannan, H., Bhandarkar, S., Matusik, W., & Torralba, A. (2016). Eye tracking for everyone. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2176-2184).
NVIDIA. (2024, Aug 27). “Gaze Estimation”: NVIDIA.
https://docs.nvidia.com/tao/tao-toolkit/text/model_zoo/cv_models/gazenet.html
Park, S., Mello, S. D., Molchanov, P., Iqbal, U., Hilliges, O., & Kautz, J. (2019). Few-shot adaptive gaze estimation. In Proceedings of the IEEE/CVF international conference on computer vision (pp. 9368-9377).
Published
How to Cite
Issue
Section
Copyright (c) 2024 Christopher Shin, Jihan Lee; Youngjee Kim

This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
Copyright holder(s) granted JSR a perpetual, non-exclusive license to distriute & display this article.


