Finance LLM for complex reasoning, driven by reinforcement learning
Top 52.2% on sourcepulse
Fin-R1 is a 7B parameter large language model specifically designed for complex financial reasoning tasks. Developed by SUFE-AIFLM-Lab and Caiyue Xingchen, it aims to provide advanced financial analysis, code generation, risk management, and compliance capabilities for financial professionals. The model achieves state-of-the-art performance on several financial benchmarks, outperforming similarly sized models and even larger distilled models in specific areas.
How It Works
Fin-R1 is built upon the Qwen2.5-7B-Instruct base model and enhanced through a two-stage fine-tuning process. This involves supervised fine-tuning (SFT) on a custom 60k-entry financial reasoning dataset (Fin-R1-Data), which was meticulously curated using a novel dual-stage data screening method combining rule-based matching, Qwen2.5-72B-Instruct for answer accuracy, and deep validation of reasoning chains for logical consistency and terminology compliance. Subsequently, the model undergoes reinforcement learning (RL) using the GRPO algorithm with format and accuracy rewards, incorporating a model-based verifier (Qwen2.5-Max) to refine output quality and generalization.
Quick Start & Requirements
pip install vllm
git clone https://huggingface.co/SUFE-AIFLM-Lab/Fin-R1
tensor-parallel-size 2
and gpu-memory-utilization 0.9
.vllm serve "/path/Fin-R1" --host 0.0.0.0 --port 8000 --gpu-memory-utilization 0.9 --max-model-len 16384 --tensor-parallel-size 2 --served-model-name "Fin-R1"
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model's outputs are for reference only and should not replace professional financial advice. Users are encouraged to apply critical thinking and their own expertise when using the model's suggestions.
4 months ago
1 week