Using a Convolutional Neural Network to Classify Subvocalized Letters Collected from the Brain by EEG

Authors

  • Shaumprovo Debnath Keystone School
  • Dr. Prasanna Kolar Southwest Research Institute, San Antonio, TX

DOI:

https://doi.org/10.47611/jsrhs.v12i3.4906

Keywords:

Convolutional Neural Network (CNN), electroencephalography (EEG), Subvocalization, Communication, Brain-computer interfaces (BCI), Artificial Intelligence

Abstract

Communication is essential to human behavior. Individuals with amyotrophic lateral sclerosis and severe traumatic brain injury encounter enormous challenges in communicating via speaking, writing or typing with others. Such lack of communication severely limits their independence and quality of life. Recent research employing invasive and expensive surgical implants of neuroprosthesis shows some degree of success in reconstructing language or images. However, these procedures are not user-friendly and not available for all affected individuals. To address these limitations, the researcher created a convolutional neural network model to classify individually thought letters – i.e., subvocalization. Using a commercially available simple electroencephalography (EEG) device data were collected from a study participant. All collected data was formatted for computer interpretation and uploaded to a Python notebook. The data was then augmented and preprocessed to make it easier for the neural network model to make predictions. The convolutional neural network model with the best performance had 63.33% classifying 38/60 samples correctly with a statistically significant p value (p<0.00001). Limitations include the reliability of the EEG headset used and time restrictions for collecting data. For future research, better hyperparameters will be investigated using additional data. Finally, the researcher plans to conduct experiments to make real-time predictions with the model once higher accuracies are achieved.

Downloads

Download data is not yet available.

References or Bibliography

Birbaumer, N. (2006). Breaking the silence: brain–computer interfaces (BCI) for communication and motor control. Psychophysiology, 43(6), 517-532. https://doi.org/10.1111/j.1469-8986.2006.00456.x

Chaudhary, U., Vlachos, I., Zimmermann, J. B., Espinosa, A., Tonin, A., Jaramillo-Gonzalez, A., ... & Birbaumer, N. (2022). Spelling interface using intracortical signals in a completely locked-in patient enabled via auditory neurofeedback training. Nature communications, 13(1), 1-9. https://doi.org/10.1038/s41467-022-28859-8

Kapur, A., Sarawgi, U., Wadkins, E., Wu, M., Hollenstein, N., & Maes, P. (2020). Non-invasive silent speech recognition in multiple sclerosis with dysphonia. In Machine Learning for Health Workshop (pp. 25-38). PMLR. https://proceedings.mlr.press/v116/kapur20a.html

Moses, D. A., Metzger, S. L., Liu, J. R., Anumanchipalli, G. K., Makin, J. G., Sun, P. F., ... & Chang, E. F. (2021). Neuroprosthesis for decoding speech in a paralyzed person with anarthria. New England Journal of Medicine, 385(3), 217-227. https://doi.org/10.1056/nejmoa2027540

Shergill, S. S., Brammer, M. J., Fukuda, R., Bullmore, E., Amaro, E., Jr, Murray, R. M., & McGuire, P. K. (2002). Modulation of activity in temporal cortex during generation of inner speech. Human brain mapping, 16(4), 219–227. https://doi.org/10.1002/hbm.10046

Tang, J., LeBel, A., Jain, S., & Huth, A. G. (2023). Semantic reconstruction of continuous language from non invasive brain recordings. Nature Neuroscience, 1-9. https://doi.org/10.1038/s41593-023-01304-9

Tyler, L. K., & Marslen-Wilson, W. (2008). Fronto-temporal brain systems supporting spoken language comprehension. Philosophical transactions of the Royal Society of London. Series B, Biological sciences, 363(1493), 1037–1054. https://doi.org/10.1098/rstb.2007.2158

Wang, F., Zhong, S. H., Peng, J., Jiang, J., & Liu, Y. (2018). Data augmentation for EEG-based emotion recognition with deep convolutional neural networks. In MultiMedia Modeling: 24th International Conference, MMM 2018, Bangkok, Thailand, February 5-7, 2018, Proceedings, Part II 24 (pp. 82-93). Springer International Publishing. https://doi.org/10.1007/978-3-319-73600-6_8

Willett, F. R., Avansino, D. T., Hochberg, L. R., Henderson, J. M., & Shenoy, K. V. (2021). High-performance brain-to-text communication via handwriting. Nature, 593(7858), 249-254. https://doi.org/10.1038/s41586-021-03506-2

Wolpaw, J. R., Birbaumer, N., Heetderks, W. J., McFarland, D. J., Peckham, P. H., Schalk, G., ... & Vaughan, T. M. (2000). Brain-computer interface technology: a review of the first international meeting. IEEE transactions on rehabilitation engineering, 8(2), 164-173. https://doi.org/10.1109/tre.2000.847807

Wolpaw, J. R., Birbaumer, N., McFarland, D. J., Pfurtscheller, G., & Vaughan, T. M. (2002). Brain-computer interfaces for communication and control. Clinical neurophysiology : official journal of the International Federation of Clinical Neurophysiology, 113(6), 767–791. https://doi.org/10.1016/s1388-2457(02)00057-3

Published

08-31-2023

How to Cite

Debnath, S., & Kolar, D. P. (2023). Using a Convolutional Neural Network to Classify Subvocalized Letters Collected from the Brain by EEG. Journal of Student Research, 12(3). https://doi.org/10.47611/jsrhs.v12i3.4906

Issue

Section

HS Research Projects