Upload folder using huggingface_hub
Browse files- base_results.json +11 -6
- config.json +1 -1
- smash_config.json +1 -0
- smashed_results.json +11 -6
base_results.json
CHANGED
@@ -1,8 +1,13 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
"
|
4 |
-
"
|
5 |
-
"
|
6 |
-
"
|
7 |
-
"
|
|
|
|
|
|
|
|
|
|
|
8 |
}
|
|
|
1 |
{
|
2 |
+
"perplexity_y_gt": 38109.7109375,
|
3 |
+
"inference_elapsed_time_ms_@1": 398.0410270690918,
|
4 |
+
"inference_latency_ms_@1": 39.80410270690918,
|
5 |
+
"inference_throughput_batches_per_ms_@1": 0.02512303838032305,
|
6 |
+
"Loading model_emissions": 7.3388832823058995e-06,
|
7 |
+
"Loading model_energy_consumed": 2.5236441197042344e-05,
|
8 |
+
"Inference_emissions": 1.7892046241417072e-05,
|
9 |
+
"Inference_energy_consumed": 6.152592369944487e-05,
|
10 |
+
"tracker_emissions": 2.896018336263697e-05,
|
11 |
+
"tracker_energy_consumed": 9.958626351897996e-05,
|
12 |
+
"disk_memory": 3158.1982421875
|
13 |
}
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/tmp/models/
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/tmp/models/tmpsyoybu74tgrh01ji",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
smash_config.json
CHANGED
@@ -16,6 +16,7 @@
|
|
16 |
"device": "cuda",
|
17 |
"save_fns": [],
|
18 |
"load_fns": [
|
|
|
19 |
"transformers",
|
20 |
"transformers"
|
21 |
],
|
|
|
16 |
"device": "cuda",
|
17 |
"save_fns": [],
|
18 |
"load_fns": [
|
19 |
+
"transformers",
|
20 |
"transformers",
|
21 |
"transformers"
|
22 |
],
|
smashed_results.json
CHANGED
@@ -1,8 +1,13 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
"
|
4 |
-
"
|
5 |
-
"
|
6 |
-
"
|
7 |
-
"
|
|
|
|
|
|
|
|
|
|
|
8 |
}
|
|
|
1 |
{
|
2 |
+
"perplexity_y_gt": 22252.1484375,
|
3 |
+
"inference_elapsed_time_ms_@1": 1357.9652862548828,
|
4 |
+
"inference_latency_ms_@1": 135.7965286254883,
|
5 |
+
"inference_throughput_batches_per_ms_@1": 0.007363958490852802,
|
6 |
+
"Loading model_emissions": 1.1571457192724436e-05,
|
7 |
+
"Loading model_energy_consumed": 3.9791121860781054e-05,
|
8 |
+
"Inference_emissions": 3.394370871224601e-05,
|
9 |
+
"Inference_energy_consumed": 0.00011672326374115302,
|
10 |
+
"tracker_emissions": 4.893639156223708e-05,
|
11 |
+
"tracker_energy_consumed": 0.0001682790583457543,
|
12 |
+
"disk_memory": 3150.1982421875
|
13 |
}
|