Fine-tuning Pretrained Transformers for Sentiment Analysis on Twitter Data

Authors

  • Anupriya

DOI:

https://doi.org/10.17762/msea.v70i2.2326

Abstract

Due to the noise and informal language present in Twitter data, it is difficult to perform sentiment analysis on the platform. In recent years, a number of transformer models have been developed that can perform well in this type of task. This study aims to analyze the performance of these models on Twitter data. The study utilizes a publicly-available dataset of tweets with neutral, positive, or negative sentiment. It preprocesses the data and tokenizes it using WordPiece. Three transformer models are then tuned using the labeled tweets' pre-defined weights and the models' training weights from large language modeling projects. The models are trained on a 5-phased scale. The three models' performance was evaluated using various metrics, such as accuracy, recall, and F1 score. The results show that the models performed well overall, with ELECTRA leading the way with an accuracy of 85.8%, followed by XLNet and BERT with 84.3% and 84.5% accuracy, respectively. The study also looked into the hyperparameters' impact on the performance. It revealed that batch sizes and learning rates have a significant effect on the models' performance. The results indicate that the models performed better with larger batch sizes and lower learning rates. The study concluded that the three pre-trained transformer models, namely XLNet, ELECTRA, and BERT, were able to perform well in terms of their performance when it came to analyzing Twitter data. Their findings can be beneficial for those working in the field of sentiment analysis on social media platforms.

Downloads

Published

2021-02-26

How to Cite

Anupriya, A. (2021). Fine-tuning Pretrained Transformers for Sentiment Analysis on Twitter Data. Mathematical Statistician and Engineering Applications, 70(2), 1344–1352. https://doi.org/10.17762/msea.v70i2.2326

Issue

Section

Articles