This paper investigates the performance disparities between Small Language Models (SLMs) and Large Language Models (LLMs) in predicting stock price movements using data from two different datasets containing news articles and tweets. The study emphasizes the potential of SLMs as a more accessible and resource-efficient alternative to LLMs, enabling local and in-house deployment. Critical gaps are addressed, including the lack of direct price movement predictions, the utilization and comparison of State-of-the-Art (SotA) models, and the integration of diverse data sources. The research employed a fundamental trading strategy based on predicted stock price movement as the sole trading signal. The Phi-2 model, fine-tuned with QLoRA on consumer-grade hardware, was compared with GPT-4, serving as a SotA benchmark. Performance was evaluated using accuracy, precision, recall, and F1-score. The results indicate that the fine-tuned SML (Phi-2) outperformed the LLM (GPT-4), albeit by a small margin, demonstrating the potential of a trained SML over a general LLM. |
*** Title, author list and abstract as submitted during Camera-Ready version delivery. Small changes that may have occurred during processing by Springer may not appear in this window.