This work focuses on inner speech recognition starting from electroencephalographic (EEG) signals. Inner speech recognition is defined as the internalised process in which the person thinks in pure meanings, generally associated with an auditory imagery of own inner “voice”. The decoding of the EEG into text should be understood as the classification of a limited number of words (commands) or the presence of phonemes (units of sound that make up words). Speech-related brain computer interfaces provide effective vocal communication strategies for controlling devices through speech commands interpreted from brain signals, improving the quality of life of people who have lost the capability to speak, by restoring communication with their environment. Two public inner speech datasets are analysed. Using this data, some classification models are studied and implemented starting from basic methods such as Support Vector Machines, to ensemble methods such as the eXtreme Gradient Boosting classifier up to the use of neural networks such as Long Short Term Memory (LSTM) and Bidirectional Long Short Term Memory (BiLSTM). With the LSTM and BiLSTM models, generally not used in the literature of inner speech recognition, results in line with or superior to those present in the state-of-the-art are obtained.

Gasparini, F., Cazzaniga, E., Saibene, A. (2022). Inner speech recognition through electroencephalographic signals. In 1st Workshop on Artificial Intelligence for Human Machine Interaction, AIxHMI 2022 (pp.48-61). CEUR-WS.

Inner speech recognition through electroencephalographic signals

Gasparini, F
;
Cazzaniga, E;Saibene, A
2022

Abstract

This work focuses on inner speech recognition starting from electroencephalographic (EEG) signals. Inner speech recognition is defined as the internalised process in which the person thinks in pure meanings, generally associated with an auditory imagery of own inner “voice”. The decoding of the EEG into text should be understood as the classification of a limited number of words (commands) or the presence of phonemes (units of sound that make up words). Speech-related brain computer interfaces provide effective vocal communication strategies for controlling devices through speech commands interpreted from brain signals, improving the quality of life of people who have lost the capability to speak, by restoring communication with their environment. Two public inner speech datasets are analysed. Using this data, some classification models are studied and implemented starting from basic methods such as Support Vector Machines, to ensemble methods such as the eXtreme Gradient Boosting classifier up to the use of neural networks such as Long Short Term Memory (LSTM) and Bidirectional Long Short Term Memory (BiLSTM). With the LSTM and BiLSTM models, generally not used in the literature of inner speech recognition, results in line with or superior to those present in the state-of-the-art are obtained.
paper
BCI; EEG; inner speech recognition
English
1st Workshop on Artificial Intelligence for Human Machine Interaction, AIxHMI 2022
2022
Saibene, A; Corchs, S; Sole-Casals, J
1st Workshop on Artificial Intelligence for Human Machine Interaction, AIxHMI 2022
2022
3368
48
61
none
Gasparini, F., Cazzaniga, E., Saibene, A. (2022). Inner speech recognition through electroencephalographic signals. In 1st Workshop on Artificial Intelligence for Human Machine Interaction, AIxHMI 2022 (pp.48-61). CEUR-WS.
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/394066
Citazioni
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
Social impact