IlyasMoutawwakil HF Staff commited on
Commit
f5f2e49
·
verified ·
1 Parent(s): 13e4cbe

Upload cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_openai-community/gpt2/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cuda_training_transformers_text-generation_openai-community/gpt2",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.3.0+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-generation",
9
  "library": "transformers",
@@ -76,10 +76,10 @@
76
  "environment": {
77
  "cpu": " AMD EPYC 7R32",
78
  "cpu_count": 16,
79
- "cpu_ram_mb": 66697.29792,
80
  "system": "Linux",
81
  "machine": "x86_64",
82
- "platform": "Linux-5.10.215-203.850.amzn2.x86_64-x86_64-with-glibc2.35",
83
  "processor": "x86_64",
84
  "python_version": "3.10.12",
85
  "gpu": [
@@ -89,15 +89,15 @@
89
  "gpu_vram_mb": 24146608128,
90
  "optimum_benchmark_version": "0.2.1",
91
  "optimum_benchmark_commit": null,
92
- "transformers_version": "4.41.1",
93
  "transformers_commit": null,
94
- "accelerate_version": "0.30.1",
95
  "accelerate_commit": null,
96
- "diffusers_version": "0.27.2",
97
  "diffusers_commit": null,
98
  "optimum_version": null,
99
  "optimum_commit": null,
100
- "timm_version": "1.0.3",
101
  "timm_commit": null,
102
  "peft_version": null,
103
  "peft_commit": null
@@ -107,33 +107,33 @@
107
  "overall": {
108
  "memory": {
109
  "unit": "MB",
110
- "max_ram": 1123.81952,
111
- "max_global_vram": 3406.299136,
112
  "max_process_vram": 0.0,
113
- "max_reserved": 2759.852032,
114
  "max_allocated": 2523.776
115
  },
116
  "latency": {
117
  "unit": "s",
118
  "count": 5,
119
- "total": 1.1130573043823244,
120
- "mean": 0.22261146087646488,
121
- "stdev": 0.2966186784115287,
122
- "p50": 0.07342182159423828,
123
- "p90": 0.5207038085937501,
124
- "p95": 0.6682699951171873,
125
- "p99": 0.7863229443359374,
126
  "values": [
127
- 0.815836181640625,
128
- 0.0780052490234375,
129
- 0.07342182159423828,
130
- 0.07328665924072265,
131
- 0.07250739288330078
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "samples/s",
136
- "value": 44.92131699162318
137
  },
138
  "energy": null,
139
  "efficiency": null
@@ -141,30 +141,30 @@
141
  "warmup": {
142
  "memory": {
143
  "unit": "MB",
144
- "max_ram": 1123.81952,
145
- "max_global_vram": 3406.299136,
146
  "max_process_vram": 0.0,
147
- "max_reserved": 2759.852032,
148
  "max_allocated": 2523.776
149
  },
150
  "latency": {
151
  "unit": "s",
152
  "count": 2,
153
- "total": 0.8938414306640625,
154
- "mean": 0.44692071533203126,
155
- "stdev": 0.36891546630859373,
156
- "p50": 0.44692071533203126,
157
- "p90": 0.7420530883789063,
158
- "p95": 0.7789446350097656,
159
- "p99": 0.8084578723144531,
160
  "values": [
161
- 0.815836181640625,
162
- 0.0780052490234375
163
  ]
164
  },
165
  "throughput": {
166
  "unit": "samples/s",
167
- "value": 8.950133352015863
168
  },
169
  "energy": null,
170
  "efficiency": null
@@ -172,31 +172,31 @@
172
  "train": {
173
  "memory": {
174
  "unit": "MB",
175
- "max_ram": 1123.81952,
176
- "max_global_vram": 3406.299136,
177
  "max_process_vram": 0.0,
178
- "max_reserved": 2759.852032,
179
  "max_allocated": 2523.776
180
  },
181
  "latency": {
182
  "unit": "s",
183
  "count": 3,
184
- "total": 0.21921587371826173,
185
- "mean": 0.07307195790608724,
186
- "stdev": 0.0004030032788678646,
187
- "p50": 0.07328665924072265,
188
- "p90": 0.07339478912353516,
189
- "p95": 0.07340830535888672,
190
- "p99": 0.07341911834716797,
191
  "values": [
192
- 0.07342182159423828,
193
- 0.07328665924072265,
194
- 0.07250739288330078
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
- "value": 82.11084213332911
200
  },
201
  "energy": null,
202
  "efficiency": null
 
3
  "name": "cuda_training_transformers_text-generation_openai-community/gpt2",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.3.1+cu121",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "text-generation",
9
  "library": "transformers",
 
76
  "environment": {
77
  "cpu": " AMD EPYC 7R32",
78
  "cpu_count": 16,
79
+ "cpu_ram_mb": 66697.293824,
80
  "system": "Linux",
81
  "machine": "x86_64",
82
+ "platform": "Linux-5.10.219-208.866.amzn2.x86_64-x86_64-with-glibc2.35",
83
  "processor": "x86_64",
84
  "python_version": "3.10.12",
85
  "gpu": [
 
89
  "gpu_vram_mb": 24146608128,
90
  "optimum_benchmark_version": "0.2.1",
91
  "optimum_benchmark_commit": null,
92
+ "transformers_version": "4.42.3",
93
  "transformers_commit": null,
94
+ "accelerate_version": "0.31.0",
95
  "accelerate_commit": null,
96
+ "diffusers_version": "0.29.2",
97
  "diffusers_commit": null,
98
  "optimum_version": null,
99
  "optimum_commit": null,
100
+ "timm_version": "1.0.7",
101
  "timm_commit": null,
102
  "peft_version": null,
103
  "peft_commit": null
 
107
  "overall": {
108
  "memory": {
109
  "unit": "MB",
110
+ "max_ram": 1716.031488,
111
+ "max_global_vram": 3559.391232,
112
  "max_process_vram": 0.0,
113
+ "max_reserved": 2910.846976,
114
  "max_allocated": 2523.776
115
  },
116
  "latency": {
117
  "unit": "s",
118
  "count": 5,
119
+ "total": 0.856558563232422,
120
+ "mean": 0.1713117126464844,
121
+ "stdev": 0.25528712586502056,
122
+ "p50": 0.04334796905517578,
123
+ "p90": 0.42704709930419926,
124
+ "p95": 0.5544658714294433,
125
+ "p99": 0.6564008891296387,
126
  "values": [
127
+ 0.6818846435546875,
128
+ 0.0447907829284668,
129
+ 0.04330086517333984,
130
+ 0.04334796905517578,
131
+ 0.04323430252075195
132
  ]
133
  },
134
  "throughput": {
135
  "unit": "samples/s",
136
+ "value": 58.37312490498423
137
  },
138
  "energy": null,
139
  "efficiency": null
 
141
  "warmup": {
142
  "memory": {
143
  "unit": "MB",
144
+ "max_ram": 1716.031488,
145
+ "max_global_vram": 3559.391232,
146
  "max_process_vram": 0.0,
147
+ "max_reserved": 2910.846976,
148
  "max_allocated": 2523.776
149
  },
150
  "latency": {
151
  "unit": "s",
152
  "count": 2,
153
+ "total": 0.7266754264831543,
154
+ "mean": 0.36333771324157715,
155
+ "stdev": 0.3185469303131104,
156
+ "p50": 0.36333771324157715,
157
+ "p90": 0.6181752574920654,
158
+ "p95": 0.6500299505233764,
159
+ "p99": 0.6755137049484252,
160
  "values": [
161
+ 0.6818846435546875,
162
+ 0.0447907829284668
163
  ]
164
  },
165
  "throughput": {
166
  "unit": "samples/s",
167
+ "value": 11.009041600205336
168
  },
169
  "energy": null,
170
  "efficiency": null
 
172
  "train": {
173
  "memory": {
174
  "unit": "MB",
175
+ "max_ram": 1716.031488,
176
+ "max_global_vram": 3559.391232,
177
  "max_process_vram": 0.0,
178
+ "max_reserved": 2910.846976,
179
  "max_allocated": 2523.776
180
  },
181
  "latency": {
182
  "unit": "s",
183
  "count": 3,
184
+ "total": 0.12988313674926757,
185
+ "mean": 0.04329437891642252,
186
+ "stdev": 4.663027563417732e-05,
187
+ "p50": 0.04330086517333984,
188
+ "p90": 0.04333854827880859,
189
+ "p95": 0.043343258666992186,
190
+ "p99": 0.04334702697753906,
191
  "values": [
192
+ 0.04330086517333984,
193
+ 0.04334796905517578,
194
+ 0.04323430252075195
195
  ]
196
  },
197
  "throughput": {
198
  "unit": "samples/s",
199
+ "value": 138.586120188551
200
  },
201
  "energy": null,
202
  "efficiency": null