LLM GPT-3.5 study for sentiment analysis across Utkarsh server, Ohio supercomputer, Google Colab and PC

B. N. Lavanya*, K. V. Anitha Rathnam, Abhishek Appaji*, K. Kiran, P. Deepa Shenoy, K. R. Venugopal

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

The major objective of the present research is to inspect sentiment analysis models that have been trained on Twitter corpus by utilising the Large Language Model (LLM) gpt-3-5-turbo-16k version of the Generative Pretrained Transformer (GPT 3.5) model. Such trained models include the Bidirectional long short-term memory neural network (BiLSTM), Convolutional Neural Networks, Gated Recurrent Unit and Recurrent Neural Network which were used to perform computational tasks on the Ohio Supercomputer and Utkarsh Server, in comparison to work conducted on Google Colab and a Personal Computer (PC).This research also looks at the performance as well as the computational aspects of these models in terms of accuracy, recall, F1-score, time/memory complexity and resource requirements (CPU/GPU) throughout the training and testing phases of each model. The training accuracies in this case were concentrated between 49.91 % and 99.98 % while those of the testing accuracies of the models accounts for about 50.00-75.00 %. For instance, models including BiLSTM and RNN usually exhibit more time complexity because of the nature of the models (sequential computation), on the contrary, CNNs are less time-consuming and are more effective in terms of storage modifying layered architecture. The use of supercomputers and specialized servers reduces training time, but resource constraints on platforms such as personal computers or Colab cause considerable divergence.
Original languageEnglish
Article number103218
Number of pages10
JournalResults in Engineering
Volume24
DOIs
Publication statusPublished - 1 Dec 2024

Keywords

  • Bidirectional long short-term memory
  • (BiLSTM)
  • Convolutional neural networks (CNN)
  • Generative pretrained transformer (GPT)
  • Gated recurrent unit (GRU)
  • Natural language processing (NLP)
  • Large language model (LLM)
  • Recurrent neural network (RNN)

Fingerprint

Dive into the research topics of 'LLM GPT-3.5 study for sentiment analysis across Utkarsh server, Ohio supercomputer, Google Colab and PC'. Together they form a unique fingerprint.

Cite this