Overview of
FinanceMath.
Large Language Models (LLMs) have demonstrated impressive problem-solving capabilities across various domains, but their ability to handle knowledge-intensive mathematical reasoning in specialized fields remains understudied. To address this gap, we introduce FinanceMath, a benchmark designed to evaluate LLMs' performance in knowledge-intensive mathematical reasoning within the finance domain. FinanceMath comprises 1200 examples covering a wide range of finance subareas, with 40.2% of problems requiring interpretation of tabular data. Each problem is accompanied by expert-annotated, Python-formatted solutions.
🚨 To submit your results to the leaderboard, please send to this email with your result json files.
🚨 For more submission details, please refer to this link.
@misc{zhao2024financemath,
title={FinanceMath: Knowledge-Intensive Math Reasoning in Finance Domains},
author={Yilun Zhao and Hongjun Liu and Yitao Long and Rui Zhang and Chen Zhao and Arman Cohan},
year={2024},
eprint={2311.09797},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2311.09797},
}