20th AIAI 2024, 27 - 30 June 2024, Corfu, Greece

LLM Prompting versus Fine-Tuning PLMs: A Comparative Study on Keyword Generation from Customer Feedback

Aleyna Er, Banu Diri, Meltem Turhan Yöndem

Abstract:

  This study focuses on keyword generation of customer feedback analysis within the tourism sector regarding tour and hotel services provided by Setur. A dataset comprising 1000 customer surveys from 2020-2022 was crafted by annotating keywords gleaned from open-ended questions. The research employs the efficacy of fine-tuning Pre-trained Language Models (PLMs) and employing Large Language Models (LLMs) through prompting for keyword generation. The study navigates through traditional statistical methods, such as TF-IDF and KP-MINER, and contemporary approaches, such as YAKE, PageRank, and TextRank. Additionally, it explores supervised methodologies, including KEA and sequence-to-sequence models, juxtaposed against the surge in popularity of pre-trained language models like T5, Bart, GPT3.5, GPT4, and Gemini. In our experimental approach, multilingual versions of T5 and Bart models, namely MT5 and MBART, were fine-tuned to prior studies. Comparison extended to include GPT3.5, GPT4, and Gemini models, utilizing diverse prompt styles, and introducing few-shot examples to discern changes in performance. The semantic similarity evaluations between generated keywords and the source text, text length metrics, and inter-keyword semantic similarity are presented. Along with MBART and MT5, Turkish language models TRBART and TRT5 were employed. The study not only aims to contribute insights into the domain of customer feedback analysis but also to serve as a benchmark for comparing the efficiency of PLM fine-tuning and LLM prompting in keyword generation from textual data.  

*** Title, author list and abstract as seen in the Camera-Ready version of the paper that was provided to Conference Committee. Small changes that may have occurred during processing by Springer may not appear in this window.