|
--- |
|
dataset_info: |
|
features: |
|
- name: question |
|
dtype: string |
|
- name: answer |
|
dtype: float64 |
|
- name: context |
|
dtype: string |
|
- name: task |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 638720 |
|
num_examples: 223 |
|
download_size: 198425 |
|
dataset_size: 638720 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
--- |
|
|
|
|
|
# Finance Fundamentals: Quantity Extraction |
|
|
|
This dataset contains evaluations for extracting numbers from financial text. The source data comes from: |
|
- [TatQA](https://arxiv.org/abs/2105.07624) |
|
- [ConvFinQA](https://arxiv.org/abs/2210.03849) |
|
|
|
Each question went through additional manual review to ensure both correctness and clarity. For more information, see the [BizBench paper.](https://aclanthology.org/2024.acl-long.452.pdf) |
|
|
|
## Example |
|
Each question will contain a document context: |
|
``` |
|
The Company’s top ten clients accounted for 42.2%, 44.2% and 46.9% of its consolidated revenues during the years ended December 31, 2019, 2018 and 2017, respectively. |
|
The following table represents a disaggregation of revenue from contracts with customers by delivery location (in thousands): |
|
| | | Years Ended December 31, | | |
|
| :--- | :--- | :--- | :--- | |
|
| | 2019 | 2018 | 2017 | |
|
| Americas: | | | | |
|
| United States | $614,493 | $668,580 | $644,870 | |
|
| The Philippines | 250,888 | 231,966 | 241,211 | |
|
| Costa Rica | 127,078 | 127,963 | 132,542 | |
|
| Canada | 99,037 | 102,353 | 112,367 | |
|
| El Salvador | 81,195 | 81,156 | 75,800 | |
|
| Other | 123,969 | 118,620 | 118,853 | |
|
| Total Americas | 1,296,660 | 1,330,638 | 1,325,643 | |
|
| EMEA: | | | | |
|
| Germany | 94,166 | 91,703 | 81,634 | |
|
| Other | 223,847 | 203,251 | 178,649 | |
|
| Total EMEA | 318,013 | 294,954 | 260,283 | |
|
| Total Other | 89 | 95 | 82 | |
|
| | $1,614,762 | $1,625,687 | $1,586,008 | |
|
``` |
|
|
|
An associated question that references the context: |
|
``` |
|
What was the Total Americas amount in 2019? (thousand) |
|
``` |
|
|
|
And an answer represented as a single float value: |
|
``` |
|
1296660.0 |
|
``` |
|
|
|
## Citation |
|
|
|
If you find this data useful, please cite: |
|
``` |
|
@inproceedings{krumdick-etal-2024-bizbench, |
|
title = "{B}iz{B}ench: A Quantitative Reasoning Benchmark for Business and Finance", |
|
author = "Krumdick, Michael and |
|
Koncel-Kedziorski, Rik and |
|
Lai, Viet Dac and |
|
Reddy, Varshini and |
|
Lovering, Charles and |
|
Tanner, Chris", |
|
editor = "Ku, Lun-Wei and |
|
Martins, Andre and |
|
Srikumar, Vivek", |
|
booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)", |
|
month = aug, |
|
year = "2024", |
|
address = "Bangkok, Thailand", |
|
publisher = "Association for Computational Linguistics", |
|
url = "https://aclanthology.org/2024.acl-long.452/", |
|
doi = "10.18653/v1/2024.acl-long.452", |
|
pages = "8309--8332", |
|
abstract = "Answering questions within business and finance requires reasoning, precision, and a wide-breadth of technical knowledge. Together, these requirements make this domain difficult for large language models (LLMs). We introduce BizBench, a benchmark for evaluating models' ability to reason about realistic financial problems. BizBench comprises eight quantitative reasoning tasks, focusing on question-answering (QA) over financial data via program synthesis. We include three financially-themed code-generation tasks from newly collected and augmented QA data. Additionally, we isolate the reasoning capabilities required for financial QA: reading comprehension of financial text and tables for extracting intermediate values, and understanding financial concepts and formulas needed to calculate complex solutions. Collectively, these tasks evaluate a model{'}s financial background knowledge, ability to parse financial documents, and capacity to solve problems with code. We conduct an in-depth evaluation of open-source and commercial LLMs, comparing and contrasting the behavior of code-focused and language-focused models. We demonstrate that the current bottleneck in performance is due to LLMs' limited business and financial understanding, highlighting the value of a challenging benchmark for quantitative reasoning within this domain." |
|
} |
|
``` |