Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json with huggingface_hub
Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
"name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
-
"version": "2.8.0
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
9 |
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
@@ -80,12 +80,12 @@
|
|
80 |
"environment": {
|
81 |
"cpu": " AMD EPYC 7R32",
|
82 |
"cpu_count": 16,
|
83 |
-
"cpu_ram_mb": 66697.
|
84 |
"system": "Linux",
|
85 |
"machine": "x86_64",
|
86 |
"platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
|
87 |
"processor": "x86_64",
|
88 |
-
"python_version": "3.10.
|
89 |
"gpu": [
|
90 |
"NVIDIA A10G"
|
91 |
],
|
@@ -93,7 +93,7 @@
|
|
93 |
"gpu_vram_mb": 24146608128,
|
94 |
"optimum_benchmark_version": "0.6.0.dev0",
|
95 |
"optimum_benchmark_commit": null,
|
96 |
-
"transformers_version": "4.
|
97 |
"transformers_commit": null,
|
98 |
"accelerate_version": "1.10.0",
|
99 |
"accelerate_commit": null,
|
@@ -103,7 +103,7 @@
|
|
103 |
"optimum_commit": null,
|
104 |
"timm_version": "1.0.19",
|
105 |
"timm_commit": null,
|
106 |
-
"peft_version":
|
107 |
"peft_commit": null
|
108 |
},
|
109 |
"print_report": true,
|
@@ -113,75 +113,75 @@
|
|
113 |
"overall": {
|
114 |
"memory": {
|
115 |
"unit": "MB",
|
116 |
-
"max_ram":
|
117 |
"max_global_vram": 720.896,
|
118 |
"max_process_vram": 0.0,
|
119 |
"max_reserved": 65.011712,
|
120 |
-
"max_allocated": 42.
|
121 |
},
|
122 |
"latency": {
|
123 |
"unit": "s",
|
124 |
"values": [
|
125 |
-
0.
|
126 |
-
0.
|
127 |
-
0.
|
128 |
-
0.
|
129 |
-
0.
|
130 |
],
|
131 |
"count": 5,
|
132 |
-
"total": 0.
|
133 |
-
"mean": 0.
|
134 |
-
"p50": 0.
|
135 |
-
"p90": 0.
|
136 |
-
"p95": 0.
|
137 |
-
"p99": 0.
|
138 |
-
"stdev": 0.
|
139 |
-
"stdev_":
|
140 |
},
|
141 |
"throughput": {
|
142 |
"unit": "samples/s",
|
143 |
-
"value":
|
144 |
},
|
145 |
"energy": {
|
146 |
"unit": "kWh",
|
147 |
-
"cpu": 6.
|
148 |
-
"ram": 3.
|
149 |
-
"gpu":
|
150 |
-
"total": 2.
|
151 |
},
|
152 |
"efficiency": {
|
153 |
"unit": "samples/kWh",
|
154 |
-
"value":
|
155 |
}
|
156 |
},
|
157 |
"warmup": {
|
158 |
"memory": {
|
159 |
"unit": "MB",
|
160 |
-
"max_ram":
|
161 |
"max_global_vram": 720.896,
|
162 |
"max_process_vram": 0.0,
|
163 |
"max_reserved": 65.011712,
|
164 |
-
"max_allocated": 42.
|
165 |
},
|
166 |
"latency": {
|
167 |
"unit": "s",
|
168 |
"values": [
|
169 |
-
0.
|
170 |
-
0.
|
171 |
],
|
172 |
"count": 2,
|
173 |
-
"total": 0.
|
174 |
-
"mean": 0.
|
175 |
-
"p50": 0.
|
176 |
-
"p90": 0.
|
177 |
-
"p95": 0.
|
178 |
-
"p99": 0.
|
179 |
-
"stdev": 0.
|
180 |
-
"stdev_":
|
181 |
},
|
182 |
"throughput": {
|
183 |
"unit": "samples/s",
|
184 |
-
"value":
|
185 |
},
|
186 |
"energy": null,
|
187 |
"efficiency": null
|
@@ -189,32 +189,32 @@
|
|
189 |
"train": {
|
190 |
"memory": {
|
191 |
"unit": "MB",
|
192 |
-
"max_ram":
|
193 |
"max_global_vram": 720.896,
|
194 |
"max_process_vram": 0.0,
|
195 |
"max_reserved": 65.011712,
|
196 |
-
"max_allocated": 42.
|
197 |
},
|
198 |
"latency": {
|
199 |
"unit": "s",
|
200 |
"values": [
|
201 |
-
0.
|
202 |
-
0.
|
203 |
-
0.
|
204 |
],
|
205 |
"count": 3,
|
206 |
-
"total": 0.
|
207 |
-
"mean": 0.
|
208 |
-
"p50": 0.
|
209 |
-
"p90": 0.
|
210 |
-
"p95": 0.
|
211 |
-
"p99": 0.
|
212 |
-
"stdev":
|
213 |
-
"stdev_":
|
214 |
},
|
215 |
"throughput": {
|
216 |
"unit": "samples/s",
|
217 |
-
"value":
|
218 |
},
|
219 |
"energy": null,
|
220 |
"efficiency": null
|
|
|
3 |
"name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
|
4 |
"backend": {
|
5 |
"name": "pytorch",
|
6 |
+
"version": "2.8.0",
|
7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
8 |
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
9 |
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
|
|
80 |
"environment": {
|
81 |
"cpu": " AMD EPYC 7R32",
|
82 |
"cpu_count": 16,
|
83 |
+
"cpu_ram_mb": 66697.248768,
|
84 |
"system": "Linux",
|
85 |
"machine": "x86_64",
|
86 |
"platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
|
87 |
"processor": "x86_64",
|
88 |
+
"python_version": "3.10.18",
|
89 |
"gpu": [
|
90 |
"NVIDIA A10G"
|
91 |
],
|
|
|
93 |
"gpu_vram_mb": 24146608128,
|
94 |
"optimum_benchmark_version": "0.6.0.dev0",
|
95 |
"optimum_benchmark_commit": null,
|
96 |
+
"transformers_version": "4.52.4",
|
97 |
"transformers_commit": null,
|
98 |
"accelerate_version": "1.10.0",
|
99 |
"accelerate_commit": null,
|
|
|
103 |
"optimum_commit": null,
|
104 |
"timm_version": "1.0.19",
|
105 |
"timm_commit": null,
|
106 |
+
"peft_version": "0.17.0",
|
107 |
"peft_commit": null
|
108 |
},
|
109 |
"print_report": true,
|
|
|
113 |
"overall": {
|
114 |
"memory": {
|
115 |
"unit": "MB",
|
116 |
+
"max_ram": 1586.307072,
|
117 |
"max_global_vram": 720.896,
|
118 |
"max_process_vram": 0.0,
|
119 |
"max_reserved": 65.011712,
|
120 |
+
"max_allocated": 42.172928
|
121 |
},
|
122 |
"latency": {
|
123 |
"unit": "s",
|
124 |
"values": [
|
125 |
+
0.2943692932128906,
|
126 |
+
0.009220095634460449,
|
127 |
+
0.008275967597961426,
|
128 |
+
0.008241151809692383,
|
129 |
+
0.008268799781799317
|
130 |
],
|
131 |
"count": 5,
|
132 |
+
"total": 0.32837530803680426,
|
133 |
+
"mean": 0.06567506160736085,
|
134 |
+
"p50": 0.008275967597961426,
|
135 |
+
"p90": 0.18030961418151858,
|
136 |
+
"p95": 0.23733945369720455,
|
137 |
+
"p99": 0.28296332530975343,
|
138 |
+
"stdev": 0.11434771850559726,
|
139 |
+
"stdev_": 174.11132278676263
|
140 |
},
|
141 |
"throughput": {
|
142 |
"unit": "samples/s",
|
143 |
+
"value": 152.264798163192
|
144 |
},
|
145 |
"energy": {
|
146 |
"unit": "kWh",
|
147 |
+
"cpu": 6.759366175000084e-06,
|
148 |
+
"ram": 3.179594461111101e-06,
|
149 |
+
"gpu": 1.0571675123999692e-05,
|
150 |
+
"total": 2.0510635760110876e-05
|
151 |
},
|
152 |
"efficiency": {
|
153 |
"unit": "samples/kWh",
|
154 |
+
"value": 487551.9275442461
|
155 |
}
|
156 |
},
|
157 |
"warmup": {
|
158 |
"memory": {
|
159 |
"unit": "MB",
|
160 |
+
"max_ram": 1586.307072,
|
161 |
"max_global_vram": 720.896,
|
162 |
"max_process_vram": 0.0,
|
163 |
"max_reserved": 65.011712,
|
164 |
+
"max_allocated": 42.172928
|
165 |
},
|
166 |
"latency": {
|
167 |
"unit": "s",
|
168 |
"values": [
|
169 |
+
0.2943692932128906,
|
170 |
+
0.009220095634460449
|
171 |
],
|
172 |
"count": 2,
|
173 |
+
"total": 0.3035893888473511,
|
174 |
+
"mean": 0.15179469442367555,
|
175 |
+
"p50": 0.15179469442367555,
|
176 |
+
"p90": 0.2658543734550476,
|
177 |
+
"p95": 0.2801118333339691,
|
178 |
+
"p99": 0.2915178012371063,
|
179 |
+
"stdev": 0.14257459878921508,
|
180 |
+
"stdev_": 93.92594341359114
|
181 |
},
|
182 |
"throughput": {
|
183 |
"unit": "samples/s",
|
184 |
+
"value": 26.35138214274844
|
185 |
},
|
186 |
"energy": null,
|
187 |
"efficiency": null
|
|
|
189 |
"train": {
|
190 |
"memory": {
|
191 |
"unit": "MB",
|
192 |
+
"max_ram": 1586.307072,
|
193 |
"max_global_vram": 720.896,
|
194 |
"max_process_vram": 0.0,
|
195 |
"max_reserved": 65.011712,
|
196 |
+
"max_allocated": 42.172928
|
197 |
},
|
198 |
"latency": {
|
199 |
"unit": "s",
|
200 |
"values": [
|
201 |
+
0.008275967597961426,
|
202 |
+
0.008241151809692383,
|
203 |
+
0.008268799781799317
|
204 |
],
|
205 |
"count": 3,
|
206 |
+
"total": 0.024785919189453127,
|
207 |
+
"mean": 0.008261973063151042,
|
208 |
+
"p50": 0.008268799781799317,
|
209 |
+
"p90": 0.008274534034729004,
|
210 |
+
"p95": 0.008275250816345215,
|
211 |
+
"p99": 0.008275824241638183,
|
212 |
+
"stdev": 1.5010837059176696e-05,
|
213 |
+
"stdev_": 0.18168586298260936
|
214 |
},
|
215 |
"throughput": {
|
216 |
"unit": "samples/s",
|
217 |
+
"value": 726.2187802040175
|
218 |
},
|
219 |
"energy": null,
|
220 |
"efficiency": null
|