sharpenb commited on
Commit
a8fe035
·
verified ·
1 Parent(s): c1b5117

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. base_results.json +6 -11
  2. config.json +1 -1
  3. smash_config.json +0 -1
  4. smashed_results.json +6 -11
base_results.json CHANGED
@@ -1,13 +1,8 @@
1
  {
2
- "perplexity_y_gt": 38109.7109375,
3
- "inference_elapsed_time_ms_@1": 400.4822006225586,
4
- "inference_latency_ms_@1": 40.04822006225586,
5
- "inference_throughput_batches_per_ms_@1": 0.02496989874819599,
6
- "Loading model_emissions": 7.406049849317766e-06,
7
- "Loading model_energy_consumed": 2.5467408914281957e-05,
8
- "Inference_emissions": 1.7792599123990086e-05,
9
- "Inference_energy_consumed": 6.118395187149493e-05,
10
- "tracker_emissions": 2.894878941998153e-05,
11
- "tracker_energy_consumed": 9.954708282175866e-05,
12
- "disk_memory": 3158.1982421875
13
  }
 
1
  {
2
+ "Loading model_emissions": 5.795339523503365e-06,
3
+ "Loading model_energy_consumed": 1.9928610317922197e-05,
4
+ "Inference_emissions": 1.594539373029533e-05,
5
+ "Inference_energy_consumed": 5.483191049086271e-05,
6
+ "tracker_emissions": 2.5390839033500574e-05,
7
+ "tracker_energy_consumed": 8.731225059232328e-05
 
 
 
 
 
8
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/tmp/models/tmphs3okmnxfnpy7s4b",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/tmp/models/tmpdew89rqs40hssuqn",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
smash_config.json CHANGED
@@ -16,7 +16,6 @@
16
  "device": "cuda",
17
  "save_fns": [],
18
  "load_fns": [
19
- "transformers",
20
  "transformers",
21
  "transformers"
22
  ],
 
16
  "device": "cuda",
17
  "save_fns": [],
18
  "load_fns": [
 
19
  "transformers",
20
  "transformers"
21
  ],
smashed_results.json CHANGED
@@ -1,13 +1,8 @@
1
  {
2
- "perplexity_y_gt": 22252.1484375,
3
- "inference_elapsed_time_ms_@1": 1410.4688110351562,
4
- "inference_latency_ms_@1": 141.04688110351563,
5
- "inference_throughput_batches_per_ms_@1": 0.0070898412795536445,
6
- "Loading model_emissions": 1.2611034407058323e-05,
7
- "Loading model_energy_consumed": 4.336594765240736e-05,
8
- "Inference_emissions": 3.463975955665669e-05,
9
- "Inference_energy_consumed": 0.00011911679495420192,
10
- "tracker_emissions": 5.0698589660335453e-05,
11
- "tracker_energy_consumed": 0.00017433878255303535,
12
- "disk_memory": 3150.1982421875
13
  }
 
1
  {
2
+ "Loading model_emissions": 1.1244527912208446e-05,
3
+ "Loading model_energy_consumed": 3.866690019844378e-05,
4
+ "Inference_emissions": 3.2272753648677847e-05,
5
+ "Inference_energy_consumed": 0.00011097729973239058,
6
+ "tracker_emissions": 4.660710352313183e-05,
7
+ "tracker_energy_consumed": 0.00016026926470704363
 
 
 
 
 
8
  }