Upload folder using huggingface_hub
Browse files- base_results.json +6 -11
- config.json +1 -1
- smash_config.json +0 -1
- smashed_results.json +6 -11
base_results.json
CHANGED
@@ -1,13 +1,8 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
"
|
4 |
-
"
|
5 |
-
"
|
6 |
-
"
|
7 |
-
"
|
8 |
-
"Inference_emissions": 1.7792599123990086e-05,
|
9 |
-
"Inference_energy_consumed": 6.118395187149493e-05,
|
10 |
-
"tracker_emissions": 2.894878941998153e-05,
|
11 |
-
"tracker_energy_consumed": 9.954708282175866e-05,
|
12 |
-
"disk_memory": 3158.1982421875
|
13 |
}
|
|
|
1 |
{
|
2 |
+
"Loading model_emissions": 5.795339523503365e-06,
|
3 |
+
"Loading model_energy_consumed": 1.9928610317922197e-05,
|
4 |
+
"Inference_emissions": 1.594539373029533e-05,
|
5 |
+
"Inference_energy_consumed": 5.483191049086271e-05,
|
6 |
+
"tracker_emissions": 2.5390839033500574e-05,
|
7 |
+
"tracker_energy_consumed": 8.731225059232328e-05
|
|
|
|
|
|
|
|
|
|
|
8 |
}
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "/tmp/models/
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "/tmp/models/tmpdew89rqs40hssuqn",
|
3 |
"architectures": [
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
smash_config.json
CHANGED
@@ -16,7 +16,6 @@
|
|
16 |
"device": "cuda",
|
17 |
"save_fns": [],
|
18 |
"load_fns": [
|
19 |
-
"transformers",
|
20 |
"transformers",
|
21 |
"transformers"
|
22 |
],
|
|
|
16 |
"device": "cuda",
|
17 |
"save_fns": [],
|
18 |
"load_fns": [
|
|
|
19 |
"transformers",
|
20 |
"transformers"
|
21 |
],
|
smashed_results.json
CHANGED
@@ -1,13 +1,8 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
"
|
4 |
-
"
|
5 |
-
"
|
6 |
-
"
|
7 |
-
"
|
8 |
-
"Inference_emissions": 3.463975955665669e-05,
|
9 |
-
"Inference_energy_consumed": 0.00011911679495420192,
|
10 |
-
"tracker_emissions": 5.0698589660335453e-05,
|
11 |
-
"tracker_energy_consumed": 0.00017433878255303535,
|
12 |
-
"disk_memory": 3150.1982421875
|
13 |
}
|
|
|
1 |
{
|
2 |
+
"Loading model_emissions": 1.1244527912208446e-05,
|
3 |
+
"Loading model_energy_consumed": 3.866690019844378e-05,
|
4 |
+
"Inference_emissions": 3.2272753648677847e-05,
|
5 |
+
"Inference_energy_consumed": 0.00011097729973239058,
|
6 |
+
"tracker_emissions": 4.660710352313183e-05,
|
7 |
+
"tracker_energy_consumed": 0.00016026926470704363
|
|
|
|
|
|
|
|
|
|
|
8 |
}
|