Status: Phase 2 Complete - Model Optimisation

Financial-Sentiment-LLM

A multi-task Large Language Model (LLM) pipeline that fine-tunes FinBERT to detect financial sentiment nuances across news, social media, and retail investor forums.

Metric / Dataset FinBERT Full LoRA
Overall Accuracy 85.4% 83.2%
Macro F1-Score 0.83 0.80
PhraseBank 95.9% 97.1%
Twitter 83.3% 80.5%
FiQA 81.5% 72.6%

Live Demo

Model Architecture & Training

This project implements a Multi-Task FinBERT architecture with dual prediction heads:

Component Purpose
Classification Head Predicts Negative/Neutral/Positive for news & social media
Regression Head Predicts continuous sentiment scores for forum discussions

Why Multi-Task FinBERT: Comparing Multi-Task vs. Single-Task FinBERT on the same data revealed a +4.4% accuracy boost. The improvement was especially significant on Twitter (+6.1%), where capturing sentiment nuance is critical.

We benchmarked multiple models (BERT-Base, DistilBERT, FinBERT) and training approaches (Full Fine-Tuning, LoRA):