Assessing the Efficacy of the ELECTRA Pre-Trained Language Model for Multi-Class Sarcasm Subcategory Classification

Imogen Jones

Research output: Book/ReportOther report

139 Downloads (Pure)

Abstract

Sarcasm detection remains a challenging task in the discipline of natural language processing, primarily due to the large levels of nuance, subjectivity, and context-sensitivity in expression of the sentiment. Pre-trained large language models have been employed in a variety of sarcasm detection tasks, including binary sarcasm detection and the classification of sarcastic speech subcategories. However, such models remain compute-hungry solutions and thus there has been a recent trend towards attempting to mitigate this through the creation of more lightweight models- including ELECTRA. This dissertation seeks to assess the efficacy of the ELECTRA pre-trained large language model, known for its computational efficiency and performant results in various natural language processing tasks, for multi-class sarcasm subcategory classification. This research proposes a partial fine-tuning approach to generalise on sarcastic data before the model is applied in several manners to the task while employing feature engineering techniques to remove overlap between hierarchical data categories. Preliminary results yield a macro F1 Score of 0.0787 for 6-class classification and 0.2363 for3-class classification, indicating potential for further improvement and application within the field.
Original languageEnglish
Place of PublicationBath, U.K.
PublisherDepartment of Computer Science, University of Bath
Number of pages61
Publication statusPublished - May 2023

Publication series

NameDepartment of Computer Science Technical Report Series
PublisherDepartment of Computer Science
ISSN (Print)1740-9497

Bibliographical note

MSc Dissertation

Fingerprint

Dive into the research topics of 'Assessing the Efficacy of the ELECTRA Pre-Trained Language Model for Multi-Class Sarcasm Subcategory Classification'. Together they form a unique fingerprint.

Cite this