Residential College | false |
Status | 已發表Published |
A preliminary study of classifying spoken vowels with EEG signals | |
Li, Mingtao1; Pun, Sio Hang2; Chen, Fei1 | |
2021-05-04 | |
Conference Name | 10th International IEEE-EMBS Conference on Neural Engineering (NER) |
Source Publication | International IEEE/EMBS Conference on Neural Engineering, NER |
Volume | 2021-May |
Pages | 13-16 |
Conference Date | MAY 04-06, 2021 |
Conference Place | Prague, ELECTR NETWORK |
Abstract | The task of classifying vowels via brain activities has been studied in many ideal direct-speech brain-computer interfaces (DS-BCIs). The vowels in those studies usually had clear acoustic differences, mainly on the first and second formants (i.e., F1 and F2). Whereas recent studies found that those speech features were difficult to be presented in DS-BCIs based on imagined speech, the spoken speech with audible output has the potential to provide insight regarding the relationship between spoken vowels' classification accuracies and their acoustic differences. This work aimed to classify four spoken Mandarin vowels (i.e., /a/, /u/, /i/ and /ü/, and pronounced with different consonants and tones to form monosyllabic stimuli in Mandarin Chinese) by using electroencephalogram (EEG) signals. The F1 and F2 of each spoken vowel were extracted; the corresponding spoken EEG signals were analyzed with the Riemannian manifold method and further used to classify spoken vowels with a linear discriminant classifier. The acoustic analysis showed that in the F1-F2 plane, the ll ellipse was closed to the /u/ and /i/ ellipses. The classification results showed that vowels /a/., /u/ and /i/ were well classified (82.0%, 69.5% and 68.2%, respectively), but vowel /ü/ was more easily classified into /u/, /i/ and /ü/. Results in this work suggested that the spoken vowels with similar formant structures were difficult to be classified by using their spoken EEG signals. |
DOI | 10.1109/NER49283.2021.9441414 |
URL | View the original |
Indexed By | CPCI-S |
Language | 英語English |
WOS Research Area | Computer Science ; Engineering ; Neurosciences & Neurology |
WOS Subject | Computer Science, Theory & Methods ; Engineering, Biomedical ; Neurosciences |
WOS ID | WOS:000681358200004 |
Scopus ID | 2-s2.0-85107466065 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | THE STATE KEY LABORATORY OF ANALOG AND MIXED-SIGNAL VLSI (UNIVERSITY OF MACAU) |
Corresponding Author | Chen, Fei |
Affiliation | 1.Southern University of Science and Technology, Department of Electrical and Electronic Engineering, Shenzhen, 518055, China 2.State Key Laboratory of Analog and Mixed-Signal VLSI, University of Macau, Taipa, 999078, Macao |
Recommended Citation GB/T 7714 | Li, Mingtao,Pun, Sio Hang,Chen, Fei. A preliminary study of classifying spoken vowels with EEG signals[C], 2021, 13-16. |
APA | Li, Mingtao., Pun, Sio Hang., & Chen, Fei (2021). A preliminary study of classifying spoken vowels with EEG signals. International IEEE/EMBS Conference on Neural Engineering, NER, 2021-May, 13-16. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment