Kit, Yuheng and Mohd. Mokji, Musa (2022) Sentiment analysis using pre-trained language model with no fine-tuning and less resource. IEEE Access, 10 (NA). pp. 107056-107065. ISSN 2169-3536
PDF
1MB |
Official URL: http://dx.doi.org/10.1109/ACCESS.2022.3212367
Abstract
Sentiment analysis has become popular when Natural Language Processing algorithms were proven to be able to process complex sentences with good accuracy. Recently, pre-trained language models such as BERT and mBERT, have been shown to be effective for improving language tasks. Most of the work in implementing the models focuses on fine-tuning BERT to achieve desirable results. However, this approach is resource-intensive and requires a long training time, up to a few hours on a GPU, depending on the dataset. Hence, this paper proposes a less complex system with less training time using the BERT model without the fine-tuning process and adopting a feature reduction algorithm to reduce sentence embeddings. The experimental results show that with 50% fewer sentence embeddings, the proposed system improves the accuracy by 1-2% with 71% less training time and 89% less memory usage. The proposed approach has also been proven to work for multilingual tasks by using a single mBERT model.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | natural language processing, Sentiment analysis |
Subjects: | T Technology > TK Electrical engineering. Electronics Nuclear engineering |
Divisions: | Electrical Engineering |
ID Code: | 104421 |
Deposited By: | Widya Wahid |
Deposited On: | 04 Feb 2024 09:58 |
Last Modified: | 04 Feb 2024 09:58 |
Repository Staff Only: item control page