IAD Index of Academic Documents
  • Home Page
  • About
    • About Izmir Academy Association
    • About IAD Index
    • IAD Team
    • IAD Logos and Links
    • Policies
    • Contact
  • Submit A Journal
  • Submit A Conference
  • Submit Paper/Book
    • Submit a Preprint
    • Submit a Book
  • Contact
  • Iğdır Üniversitesi Fen Bilimleri Enstitüsü Dergisi
  • Cilt: 15 Sayı: 4
  • Efficient Adaptation of Large Language Models for Sentiment Analysis: A Fine-Tuning Approach

Efficient Adaptation of Large Language Models for Sentiment Analysis: A Fine-Tuning Approach

Authors : Seda Bayat Toksöz, Gültekin Işık
Pages : 1149-1164
Doi:10.21597/jist.1648466
View : 229 | Download : 317
Publication Date : 2025-12-01
Article Type : Research Paper
Abstract :This study presents a systematic comparative analysis of sentiment classification on financial news headlines using two transformer architectures, Mistral-7B and GPT-2, fine-tuned with advanced adaptation techniques—Quantized Low-Rank Adaptation (QLoRA) and Low-Rank Adaptation (LoRA).Utilising a large-scale Finance News dataset, the models are rigorously evaluated for their ability to accurately classify headlines into positive, neutral, and negative sentiments while also considering computational efficiency. Beyond overall accuracy, we report macro‑averaged precision, recall, and F1‑score, thereby providing a fuller picture of the models’ class‑wise behaviour.Empirical findings demonstrate that the Mistral-7B-based configurations substantially outperform those based on GPT-2, with Mistral-7B-QLoRA achieving the highest accuracy (0.881) and Mistral-7B-Lo RA, with a score of 0.878, while GPT-2 models demonstrate significantly lower performance (0.519 for GPT-2-LoRA and 0.517 for GPT-2-QLoRA). Detailed analyses, incorporating confusion matrices and standard evaluation metrics, underscore the superior balance of classification performance and resource efficiency offered by Mistral-7B. The study goes on to discuss limitations, including the focus on a single financial dataset, and outlines prospects for future research, including the evaluation of additional architectures and adaptation techniques across diverse domains.This work contributes to the advancement of fine-tuning strategies for large language models, offering valuable insights for optimising sentiment analysis pipelines in resource-constrained environments.
Keywords : Duygu Analizi, İnce Ayar, QLoRa (Nicelenmiş Düşük Sıralı Adaptation), LoRa (Low-Rank Adaptation), Gpt-2, Mistral-7B

ORIGINAL ARTICLE URL

* There may have been changes in the journal, article,conference, book, preprint etc. informations. Therefore, it would be appropriate to follow the information on the official page of the source. The information here is shared for informational purposes. IAD is not responsible for incorrect or missing information.


Index of Academic Documents
İzmir Academy Association
CopyRight © 2023-2026