Datasets:

Modalities:
Text
Formats:
parquet
Size:
< 1K
ArXiv:
Libraries:
Datasets
pandas
File size: 4,227 Bytes
b48934d
 
 
 
 
 
 
 
 
1920fa9
 
b48934d
 
1920fa9
b48934d
1920fa9
 
b48934d
 
 
 
 
 
f092440
 
 
 
08c0a7e
f092440
 
 
6153555
f092440
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
---
dataset_info:
  features:
  - name: question
    dtype: string
  - name: answer
    dtype: float64
  - name: context
    dtype: string
  - name: task
    dtype: string
  splits:
  - name: train
    num_bytes: 638720
    num_examples: 223
  download_size: 198425
  dataset_size: 638720
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---


# Finance Fundamentals: Quantity Extraction

This dataset contains evaluations for extracting numbers from financial text. The source data comes from:
- [TatQA](https://arxiv.org/abs/2105.07624)
- [ConvFinQA](https://arxiv.org/abs/2210.03849)

Each question went through additional manual review to ensure both correctness and clarity. For more information, see the [BizBench paper.](https://aclanthology.org/2024.acl-long.452.pdf)

## Example
Each question will contain a document context:
```
The Company’s top ten clients accounted for 42.2%, 44.2% and 46.9% of its consolidated revenues during the years ended December 31, 2019, 2018 and 2017, respectively.
The following table represents a disaggregation of revenue from contracts with customers by delivery location (in thousands):
|  |  | Years Ended December 31, |  |
| :--- | :--- | :--- | :--- |
|  | 2019 | 2018 | 2017 |
| Americas: |  |  |  |
| United States | $614,493 | $668,580 | $644,870 |
| The Philippines | 250,888 | 231,966 | 241,211 |
| Costa Rica | 127,078 | 127,963 | 132,542 |
| Canada | 99,037 | 102,353 | 112,367 |
| El Salvador | 81,195 | 81,156 | 75,800 |
| Other | 123,969 | 118,620 | 118,853 |
| Total Americas | 1,296,660 | 1,330,638 | 1,325,643 |
| EMEA: |  |  |  |
| Germany | 94,166 | 91,703 | 81,634 |
| Other | 223,847 | 203,251 | 178,649 |
| Total EMEA | 318,013 | 294,954 | 260,283 |
| Total Other | 89 | 95 | 82 |
|  | $1,614,762 | $1,625,687 | $1,586,008 |
```

An associated question that references the context:
```
What was the  Total Americas  amount in 2019? (thousand)
```

And an answer represented as a single float value:
```
1296660.0
```

## Citation

If you find this data useful, please cite:
```
@inproceedings{krumdick-etal-2024-bizbench,
    title = "{B}iz{B}ench: A Quantitative Reasoning Benchmark for Business and Finance",
    author = "Krumdick, Michael  and
      Koncel-Kedziorski, Rik  and
      Lai, Viet Dac  and
      Reddy, Varshini  and
      Lovering, Charles  and
      Tanner, Chris",
    editor = "Ku, Lun-Wei  and
      Martins, Andre  and
      Srikumar, Vivek",
    booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    month = aug,
    year = "2024",
    address = "Bangkok, Thailand",
    publisher = "Association for Computational Linguistics",
    url = "https://aclanthology.org/2024.acl-long.452/",
    doi = "10.18653/v1/2024.acl-long.452",
    pages = "8309--8332",
    abstract = "Answering questions within business and finance requires reasoning, precision, and a wide-breadth of technical knowledge. Together, these requirements make this domain difficult for large language models (LLMs). We introduce BizBench, a benchmark for evaluating models' ability to reason about realistic financial problems. BizBench comprises eight quantitative reasoning tasks, focusing on question-answering (QA) over financial data via program synthesis. We include three financially-themed code-generation tasks from newly collected and augmented QA data. Additionally, we isolate the reasoning capabilities required for financial QA: reading comprehension of financial text and tables for extracting intermediate values, and understanding financial concepts and formulas needed to calculate complex solutions. Collectively, these tasks evaluate a model{'}s financial background knowledge, ability to parse financial documents, and capacity to solve problems with code. We conduct an in-depth evaluation of open-source and commercial LLMs, comparing and contrasting the behavior of code-focused and language-focused models. We demonstrate that the current bottleneck in performance is due to LLMs' limited business and financial understanding, highlighting the value of a challenging benchmark for quantitative reasoning within this domain."
}
```