Financial sentiment analysis via fine-tuned BERT
Top 24.8% on sourcepulse
FinBERT provides a pre-trained BERT model specifically fine-tuned for financial sentiment analysis. It aims to improve sentiment classification accuracy on financial texts by leveraging a large financial corpus and a specialized training approach. This is beneficial for researchers and developers working with financial news, reports, or social media data.
How It Works
FinBERT builds upon the BERT architecture by further training it on a large financial corpus (Reuters TRC2 subset) for language model adaptation, followed by fine-tuning on the Financial PhraseBank dataset for sentiment classification. This domain-specific pre-training and fine-tuning approach is designed to capture the nuances of financial language, leading to more accurate sentiment predictions compared to general-purpose NLP models.
Quick Start & Requirements
conda env create -f environment.yml
and conda activate finbert
.Highlighted Details
predict.py
script for easy sentiment prediction on text files.finbert_training.ipynb
) is provided for custom training.Maintenance & Community
Licensing & Compatibility
pytorch_pretrained_bert
, an older version of Hugging Face's transformers
library. Compatibility with commercial or closed-source projects is not specified.Limitations & Caveats
The project relies on an older library (pytorch_pretrained_bert
) which is noted as a priority for migration to the newer transformers
library. The TRC2 dataset used for language model training is not publicly available, requiring a separate application for access.
2 years ago
Inactive