Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-10318_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-10318_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-11256_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-11256_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-12194_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-12194_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-1876_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-1876_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-2814_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-2814_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-3752_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-3752_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-4690_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-4690_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-5628_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-5628_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-6566_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-6566_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-7504_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-7504_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-8442_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-8442_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-9380_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-9380_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-938_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-938_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1173_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1173_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1407-figureqa_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1407-figureqa_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1564_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1564_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1876-figureqa_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1876-figureqa_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1955_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1955_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2345-figureqa_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2345-figureqa_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2346_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2346_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2737_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2737_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2814-figureqa_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2814-figureqa_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3128_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3128_summary.json +8 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3283-figureqa_responses_sampled.json +0 -0
- eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3283-figureqa_summary.json +8 -0
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-10318_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-10318_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.91992,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-10318",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22998
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-11256_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-11256_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.92472,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-11256",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 23118
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-12194_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-12194_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.92568,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-12194",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 23142
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-1876_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-1876_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.79512,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-1876",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 19878
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-2814_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-2814_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.83552,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-2814",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 20888
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-3752_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-3752_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.86252,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-3752",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 21563
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-4690_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-4690_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.86796,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-4690",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 21699
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-5628_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-5628_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.87816,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-5628",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 21954
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-6566_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-6566_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.89024,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-6566",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22256
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-7504_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-7504_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.8962,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-7504",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22405
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-8442_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-8442_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.91004,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-8442",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22751
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-9380_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-9380_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.91696,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-9380",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22924
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-938_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_checkpoint-938_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.74036,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1/checkpoint-938",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 18509
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-chart_mixed_v1-figureqa_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.92592,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-chart_mixed_v1",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 23148
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1173_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1173_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.86412,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-1173",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 21603
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1407-figureqa_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1407-figureqa_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.87056,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-1407",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 21764
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1564_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1564_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.86492,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-1564",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 21623
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1876-figureqa_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1876-figureqa_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.88392,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-1876",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22098
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1955_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-1955_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.89048,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-1955",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22262
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2345-figureqa_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2345-figureqa_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.89192,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-2345",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22298
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2346_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2346_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.89012,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-2346",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22253
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2737_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2737_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.9052,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-2737",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22630
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2814-figureqa_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-2814-figureqa_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.9036,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-2814",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22590
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3128_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3128_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.91004,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-3128",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22751
|
8 |
+
}
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3283-figureqa_responses_sampled.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
eval_llava/eval_data/figureqa/eval_results/llava-v1.5-13b-figureqa_checkpoint-3283-figureqa_summary.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"accuracy": 0.91204,
|
3 |
+
"model_path": "data/train_llava/llava-v1.5-13b-figureqa/checkpoint-3283",
|
4 |
+
"input_file": "figureqa_sampled.json",
|
5 |
+
"metric": "em",
|
6 |
+
"total_samples": 25000,
|
7 |
+
"correct_samples": 22801
|
8 |
+
}
|