sharpenb commited on
Commit
5fe5153
·
verified ·
1 Parent(s): a8fe035

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. base_results.json +11 -6
  2. config.json +1 -1
  3. smash_config.json +1 -0
  4. smashed_results.json +11 -6
base_results.json CHANGED
@@ -1,8 +1,13 @@
1
  {
2
- "Loading model_emissions": 5.795339523503365e-06,
3
- "Loading model_energy_consumed": 1.9928610317922197e-05,
4
- "Inference_emissions": 1.594539373029533e-05,
5
- "Inference_energy_consumed": 5.483191049086271e-05,
6
- "tracker_emissions": 2.5390839033500574e-05,
7
- "tracker_energy_consumed": 8.731225059232328e-05
 
 
 
 
 
8
  }
 
1
  {
2
+ "perplexity_y_gt": 38109.7109375,
3
+ "inference_elapsed_time_ms_@1": 398.0410270690918,
4
+ "inference_latency_ms_@1": 39.80410270690918,
5
+ "inference_throughput_batches_per_ms_@1": 0.02512303838032305,
6
+ "Loading model_emissions": 7.3388832823058995e-06,
7
+ "Loading model_energy_consumed": 2.5236441197042344e-05,
8
+ "Inference_emissions": 1.7892046241417072e-05,
9
+ "Inference_energy_consumed": 6.152592369944487e-05,
10
+ "tracker_emissions": 2.896018336263697e-05,
11
+ "tracker_energy_consumed": 9.958626351897996e-05,
12
+ "disk_memory": 3158.1982421875
13
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/tmp/models/tmpdew89rqs40hssuqn",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/tmp/models/tmpsyoybu74tgrh01ji",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
smash_config.json CHANGED
@@ -16,6 +16,7 @@
16
  "device": "cuda",
17
  "save_fns": [],
18
  "load_fns": [
 
19
  "transformers",
20
  "transformers"
21
  ],
 
16
  "device": "cuda",
17
  "save_fns": [],
18
  "load_fns": [
19
+ "transformers",
20
  "transformers",
21
  "transformers"
22
  ],
smashed_results.json CHANGED
@@ -1,8 +1,13 @@
1
  {
2
- "Loading model_emissions": 1.1244527912208446e-05,
3
- "Loading model_energy_consumed": 3.866690019844378e-05,
4
- "Inference_emissions": 3.2272753648677847e-05,
5
- "Inference_energy_consumed": 0.00011097729973239058,
6
- "tracker_emissions": 4.660710352313183e-05,
7
- "tracker_energy_consumed": 0.00016026926470704363
 
 
 
 
 
8
  }
 
1
  {
2
+ "perplexity_y_gt": 22252.1484375,
3
+ "inference_elapsed_time_ms_@1": 1357.9652862548828,
4
+ "inference_latency_ms_@1": 135.7965286254883,
5
+ "inference_throughput_batches_per_ms_@1": 0.007363958490852802,
6
+ "Loading model_emissions": 1.1571457192724436e-05,
7
+ "Loading model_energy_consumed": 3.9791121860781054e-05,
8
+ "Inference_emissions": 3.394370871224601e-05,
9
+ "Inference_energy_consumed": 0.00011672326374115302,
10
+ "tracker_emissions": 4.893639156223708e-05,
11
+ "tracker_energy_consumed": 0.0001682790583457543,
12
+ "disk_memory": 3150.1982421875
13
  }