Universiti Teknologi Malaysia Institutional Repository

Pre-trained language model with feature reduction and no fine-tuning

Kit, Y. H. and Mohd. Mokji, Musa (2022) Pre-trained language model with feature reduction and no fine-tuning. In: Control, Instrumentation and Mechatronics: Theory and Practice. Lecture Notes in Electrical Engineering, 921 (NA). Springer Science and Business Media Deutschland GmbH, Singapore, pp. 685-696. ISBN 978-981193922-8

Full text not available from this repository.

Official URL: http://dx.doi.org/10.1007/978-981-19-3923-5_59

Abstract

Pre-trained language models were proven to achieve excellent results in Natural Language Processing tasks such as Sentiment Analysis. However, the number of sentence embeddings from the base model of Bidirectional Encoder from Transformer (BERT) is 768 for a sentence, and there will be more than millions of unique numbers when the dataset is huge, leading to the increasing complexity of the system. Thus, this paper presents the feature reduction of the sentence embeddings classification with BERT to decrease the number of features and complexity by using feature reduction algorithm. With 50% fewer features, the experimental results show that the proposed system improves the accuracy by 1%−2% with 89% lesser GPU memory usage.

Item Type:Book Section
Uncontrolled Keywords:feature reduction, pre-trained language model, sentiment analysis
Subjects:T Technology > TK Electrical engineering. Electronics Nuclear engineering
Divisions:Faculty of Engineering - School of Electrical
ID Code:100760
Deposited By: Yanti Mohd Shah
Deposited On:18 May 2023 03:38
Last Modified:18 May 2023 03:38

Repository Staff Only: item control page