A Symbolic Benchmark for Verifiable Chain-of-Thought Financial Reasoning
FinChain is the first benchmark designed for verifiable chain-of-thought (CoT) financial reasoning. It evaluates large language models on symbolic, multi-step problem-solving tasks grounded in financial equations. Built from scratch using a fine-grained financial taxonomy, FinChain enables step-level supervision and robust diagnostic evaluation.
📄 Paper: FinChain: A Symbolic Benchmark for Verifiable Chain-of-Thought Financial Reasoning (EMNLP 2025 submission)
- 54 topics across 12 financial domains
- 5 symbolic templates per topic (2 easy, 2 intermediate, 1 advanced)
- Executable Python traces for step-level answer verification
- ChainEval, a custom metric for evaluating both final answers and intermediate steps
This example shows a symbolic template for Compound Interest:
- Parameterized with named variables (e.g.,
principal
,rate
,time
) - Includes both natural language and step-by-step symbolic solution
- Fully executable and verifiable
finchain/
├── data/
│ └── templates/ # Symbolic prompt templates for 54 financial topics
├── eval/ # ChainEval evaluation scripts (coming soon)
└── README.md
Each instance includes:
- A financial problem generated from symbolic templates
- Gold reasoning trace with intermediate variables and calculations
- Executable code for ground-truth generation and verification
FinChain covers 54 financial topics across 12 domains:
Domains include:
- Corporate Finance
- Investment Analysis
- Personal Finance
- Financial Ratios
- Risk Management
- Sustainable Finance
- Mergers & Acquisitions
- Financial Markets
- Fintech
- Crypto Finance
- Financial Reporting
- Finance Regulation
FinChain introduces ChainEval, a joint evaluation framework for:
- ✅ Final Answer Correctness (FAC)
- 🔗 Step Alignment via:
- Semantic similarity of reasoning steps
- Numerical agreement at each step
This allows precise tracking of where models hallucinate, skip, or miscalculate.
We evaluate 30 models, including:
- GPT-4.1, GPT-4o-mini, LLaMA 3.3 70B
- Qwen3, DeepSeek-R1, Mixtral, Mathstral
- Fin-tuned models: Fino1, FinR1, WiroAI Finance Qwen
Findings:
- Larger models outperform smaller financial-tuned models
- Even top models struggle on advanced templates and multi-hop symbolic chains
- FinChain reveals reasoning gaps not captured by standard accuracy metrics
git clone https://github.com/mbzuai-nlp/finchain.git
cd finchain
Explore templates:
ls data/templates/
Evaluate predictions (scripts coming soon):
python eval/eval_chain.py --pred path/to/your_outputs.jsonl
FinChain is developed by:
Zhuohan Xie, Dhruv Sahnan, Debopriyo Banerjee, Georgi Georgiev,
Rushil Thareja, Hachem Madmoun, Jinyan Su, Aaryamonvikram Singh,
Yuxia Wang, Rui Xing, Fajri Koto, Haonan Li, Ivan Koychev,
Tanmoy Chakraborty, Salem Lahlou, Veselin Stoyanov, Preslav Nakov
Affiliations: MBZUAI, Sofia University, Quantsquare, Cornell University, IIT Delhi
For questions or collaborations, contact: zhuohan.xie@mbzuai.ac.ae
Disclaimer: FinChain uses synthetic data based on symbolic financial equations. It does not reflect real-world financial advice or regulation.