Machine Learning-Driven Image Noise Robustness Enhancement for Reliable Gaze Estimation

Authors

  • Hyungchan Yoo North London Collegiate School Jeju
  • Tajvir Singh North London Collegiate School Jeju

DOI:

https://doi.org/10.47611/jsrhs.v13i4.8051

Keywords:

Gaze Estimation, Human Computer Interaction, Denoise

Abstract

This research paper presents the development of a novel Human-Computer Interaction (HCI) system that enables users to control a computer mouse through eye movements by utilizing machine learning techniques. Traditional input devices, such as a mouse or keyboard, often present accessibility challenges for individuals with physical disabilities or limited mobility. To address this problem, I proposed machine learning-driven gaze estimation system for human computer interaction for quadriplegia patients. The proposed system takes both eye images as input to predict gaze vector. The system interprets real-time eye movement data to execute cursor movements which offers an alternative and inclusive method of computer interaction. To improve the accuracy of the gaze estimation system, I introduced a random gaussian noise-based denoising autoencoder. Experimental results demonstrate that this approach significantly enhanced accuracy reducing the angular error by 4.7 degrees on a public gaze estimation dataset. Additionally, I conducted real-world experiments to evaluate the model's performance in tracking the gaze of actual users. The results indicate that the proposed model achieved an accuracy of 91%, representing an enhancement of up to 15% compared to state-of-the-art gaze estimation methods.

Downloads

Download data is not yet available.

References or Bibliography

Gideon, J., Su, S., & Stent, S. (2022). Unsupervised multi-view gaze representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 5001-5009).

https://doi.org/10.1109/CVPRW56347.2022.00548

He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778). https://doi.org/10.48550/arXiv.1512.03385

Park, S., Mello, S. D., Molchanov, P., Iqbal, U., Hilliges, O., & Kautz, J. (2019). Few-shot adaptive gaze estimation. In Proceedings of the IEEE/CVF international conference on computer vision (pp. 9368-9377). https://doi.org/10.48550/arXiv.1905.01941

Rosebrock, A. (2020, Feb 24). “Denoising autoencoders with Keras, TensorFlow, and Deep Learning”: Py Image Search

https://pyimagesearch.com/2020/02/24/denoising-autoencoders-with-keras-tensorflow-and-deep-learning/

Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. https://doi.org/10.48550/arXiv.1409.1556

Sun, Y., Zeng, J., Shan, S., & Chen, X. (2021). Cross-encoder for unsupervised gaze representation learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 3702-3711). https://doi.org/10.1109/ICCV48922.2021.00368

Yu, Y., & Odobez, J. M. (2020). Unsupervised representation learning for gaze estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 7314-7324). https://doi.org/10.48550/arXiv.1911.06939

Zhang, X., Park, S., Beeler, T., Bradley, D., Tang, S., & Hilliges, O. (2020). Eth-xgaze: A large scale dataset for gaze estimation under extreme head pose and gaze variation. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part V 16 (pp. 365-381). Springer International Publishing.

Published

11-30-2024

How to Cite

Yoo, H., & Singh, T. (2024). Machine Learning-Driven Image Noise Robustness Enhancement for Reliable Gaze Estimation. Journal of Student Research, 13(4). https://doi.org/10.47611/jsrhs.v13i4.8051

Issue

Section

HS Research Articles