Decoding Neural Activity for Part-of-Speech Tagging (POS)

Salman Ahmed, Muskaan Singh, Saugat Bhattacharyya, Damien Coyle

Research output: Chapter or section in a book/report/conference proceedingChapter in a published conference proceeding

Abstract

Decoding Part of Speech(POS) tagging directly from electroencephalography (EEG) signals whilst user overtly spoke (voiced speech) sentences could improve direct speech brain-computer interfaces (BCIs) using imagined or inner speech. To the best of our knowledge, earlier work uses machine learning approach using 74,953 sentences/tokens recorded in 75 EEG sessions. The tokens can be found in 4,479 phrases consisting of terms from the English Online treebank which contains the record of weblogs, newsgroups, reviews, and Yahoo Answers. The results demonstrated the feasibility of POS decoding from EEG based on word class, word frequency, and word length with accuracy of 71%, 86%, 89%, respectively. We believe that there is significant room for improvement with more advanced artificial intelligence. In this paper, we further extend the existing work with end-to-end transformers. Our results presents transformer model outperforms benchmark traditional ML results with +20% in length, +13% for the open vs closed class and +12% in frequency. In our empirical analysis, we find the decoding performance was better when using multi-electrode recordings as compared to single-electrode recordings.

Original languageEnglish
Title of host publication2023 IEEE International Conference on Systems, Man, and Cybernetics
Subtitle of host publicationImproving the Quality of Life, SMC 2023 - Proceedings
Place of PublicationU. S. A.
PublisherIEEE
Pages3079-3084
Number of pages6
ISBN (Electronic)9798350337020
ISBN (Print)9798350337037
DOIs
Publication statusPublished - 4 Oct 2023
Event2023 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2023 - Hybrid, Honolulu, USA United States
Duration: 1 Oct 20234 Oct 2023

Publication series

NameConference Proceedings - IEEE International Conference on Systems, Man and Cybernetics
ISSN (Print)1062-922X

Conference

Conference2023 IEEE International Conference on Systems, Man, and Cybernetics, SMC 2023
Country/TerritoryUSA United States
CityHybrid, Honolulu
Period1/10/234/10/23

Funding

ACKNOWLEDGMENT I would like to express my sincere gratitude to Alex Murphy for generously sharing the dataset and granting permission to build upon his work [1]. This work was supported by a research grant from the Department for the Economy Northern Ireland under the US-Ireland R&D Partnership Programme (USI-207) VI. CONCLUSION The study is focused to reproduced and improve the initial results reported in base paper[1]. We found that the classification accuracy for decoding part of speech from EEG signals was significantly higher than chance levels, indicating that the EEG signals contain useful information about the syntactic structure of language. The classification accuracy was highest for nouns and verbs, typically associated with more distinctive neural processing than other parts of speech. Additionally, the study found that the decoding accuracy was affected by various factors, such as word class, word frequency and sentence length. Our results indicate that the deep learning model transformer model outperformed traditional SVM for word frequency, length and class. Specifically, shorter words and more frequent words were associated with higher decoding accuracy, while longer sentences were associated with lower decoding accuracy. These findings suggest that EEG signals can be used to decode part-of-speech information in real-time, potentially enabling the development of novel brain-computer interfaces for language processing and communication.

FundersFunder number
US-Ireland R&D Partnership ProgrammeUSI-207
University of Bath

    ASJC Scopus subject areas

    • Electrical and Electronic Engineering
    • Control and Systems Engineering
    • Human-Computer Interaction

    Fingerprint

    Dive into the research topics of 'Decoding Neural Activity for Part-of-Speech Tagging (POS)'. Together they form a unique fingerprint.

    Cite this