Implementation of Eye Movement-based Digital Healthcare Communication Board: Accurate Gaze Estimation Using Triplet Loss Function

Authors

  • Christopher Shin Korea International School
  • Jihan Lee Korea International School Pangyo Campus
  • Youngjee Kim Korea International School Pangyo Campus

DOI:

https://doi.org/10.47611/jsrhs.v13i4.8372

Keywords:

Gaze Estimation, Machine Learning

Abstract

Effective communication is important in healthcare, especially for quadriplegia patients who often face barriers in expressing their needs. Traditionally, these patients rely on healthcare communication boards to interact with doctors and nurses. However, this method is often slow, cumbersome, and requires the assistance of another person which makes it an inconvenient solution. Recent advancements in gaze estimation technology, which predicts the direction of an individual's eye movements, can provide a promising alternative. This research examines the potential of gaze estimation to develop a digital healthcare communication board driven by eye movements. Such a system would empower quadriplegia patients to communicate independently which enhances both the speed and efficiency of interactions. The proposed system processes eye images to predict a gaze vector which represents the direction in which an individual is currently looking. To improve the system's accuracy, we introduce a conjugate ability-based loss function. Additionally, the proposed approach was applied to a digital healthcare system to demonstrate its feasibility and effectiveness in real-world scenarios. The system achieved an angular error of 8.7 on a public gaze estimation dataset which surpasses previous state-of-the-art methods.

Downloads

Download data is not yet available.

References or Bibliography

Abdelrahman, A. A., Hempel, T., Khalifa, A., Al-Hamadi, A., & Dinges, L. (2023, October). L2cs-net: Fine-grained gaze estimation in unconstrained environments. In 2023 8th International Conference on Frontiers of Signal Processing (ICFSP) (pp. 98-102). IEEE.

Autodesk. (2024, Aug 27). “Communication Board for Individuals With Disabilities”: Autodesk.

https://www.instructables.com/Communication-Board-for-Individuals-with-Disabilit/

Bonazzi, P., Rüegg, T., Bian, S., Li, Y., & Magno, M. (2023, October). Tinytracker: Ultra-fast and ultra-low-power edge vision in-sensor for gaze estimation. In 2023 IEEE SENSORS (pp. 1-4). IEEE.

Cleveland Clinic. (2024, Aug 27). “Quadriplegia”: Cleveland Clinic.

https://my.clevelandclinic.org/health/symptoms/23974-quadriplegia-tetraplegia

Fischer, T., Chang, H. J., & Demiris, Y. (2018). Rt-gene: Real-time eye gaze estimation in natural environments. In Proceedings of the European conference on computer vision (ECCV) (pp. 334-352).

He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778). https://doi.org/10.48550/arXiv.1512.03385

Krafka, K., Khosla, A., Kellnhofer, P., Kannan, H., Bhandarkar, S., Matusik, W., & Torralba, A. (2016). Eye tracking for everyone. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2176-2184).

NVIDIA. (2024, Aug 27). “Gaze Estimation”: NVIDIA.

https://docs.nvidia.com/tao/tao-toolkit/text/model_zoo/cv_models/gazenet.html

Park, S., Mello, S. D., Molchanov, P., Iqbal, U., Hilliges, O., & Kautz, J. (2019). Few-shot adaptive gaze estimation. In Proceedings of the IEEE/CVF international conference on computer vision (pp. 9368-9377).

Published

11-30-2024

How to Cite

Shin, C., Lee, J., & Kim, Y. (2024). Implementation of Eye Movement-based Digital Healthcare Communication Board: Accurate Gaze Estimation Using Triplet Loss Function. Journal of Student Research, 13(4). https://doi.org/10.47611/jsrhs.v13i4.8372

Issue

Section

HS Research Articles