Universiti Teknologi Malaysia Institutional Repository

Sentiment analysis using pre-trained language model with no fine-tuning and less resource

Kit, Yuheng and Mohd. Mokji, Musa (2022) Sentiment analysis using pre-trained language model with no fine-tuning and less resource. IEEE Access, 10 (NA). pp. 107056-107065. ISSN 2169-3536

[img] PDF
1MB

Official URL: http://dx.doi.org/10.1109/ACCESS.2022.3212367

Abstract

Sentiment analysis has become popular when Natural Language Processing algorithms were proven to be able to process complex sentences with good accuracy. Recently, pre-trained language models such as BERT and mBERT, have been shown to be effective for improving language tasks. Most of the work in implementing the models focuses on fine-tuning BERT to achieve desirable results. However, this approach is resource-intensive and requires a long training time, up to a few hours on a GPU, depending on the dataset. Hence, this paper proposes a less complex system with less training time using the BERT model without the fine-tuning process and adopting a feature reduction algorithm to reduce sentence embeddings. The experimental results show that with 50% fewer sentence embeddings, the proposed system improves the accuracy by 1-2% with 71% less training time and 89% less memory usage. The proposed approach has also been proven to work for multilingual tasks by using a single mBERT model.

Item Type:Article
Uncontrolled Keywords:natural language processing, Sentiment analysis
Subjects:T Technology > TK Electrical engineering. Electronics Nuclear engineering
Divisions:Electrical Engineering
ID Code:104421
Deposited By: Widya Wahid
Deposited On:04 Feb 2024 09:58
Last Modified:04 Feb 2024 09:58

Repository Staff Only: item control page