Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json with huggingface_hub
Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
"name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
|
3 |
"backend": {
|
4 |
"name": "pytorch",
|
5 |
-
"version": "2.8.0
|
6 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
7 |
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
8 |
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
@@ -79,12 +79,12 @@
|
|
79 |
"environment": {
|
80 |
"cpu": " AMD EPYC 7R32",
|
81 |
"cpu_count": 16,
|
82 |
-
"cpu_ram_mb": 66697.
|
83 |
"system": "Linux",
|
84 |
"machine": "x86_64",
|
85 |
"platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
|
86 |
"processor": "x86_64",
|
87 |
-
"python_version": "3.10.
|
88 |
"gpu": [
|
89 |
"NVIDIA A10G"
|
90 |
],
|
@@ -92,7 +92,7 @@
|
|
92 |
"gpu_vram_mb": 24146608128,
|
93 |
"optimum_benchmark_version": "0.6.0.dev0",
|
94 |
"optimum_benchmark_commit": null,
|
95 |
-
"transformers_version": "4.
|
96 |
"transformers_commit": null,
|
97 |
"accelerate_version": "1.10.0",
|
98 |
"accelerate_commit": null,
|
@@ -102,7 +102,7 @@
|
|
102 |
"optimum_commit": null,
|
103 |
"timm_version": "1.0.19",
|
104 |
"timm_commit": null,
|
105 |
-
"peft_version":
|
106 |
"peft_commit": null
|
107 |
},
|
108 |
"print_report": true,
|
|
|
2 |
"name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
|
3 |
"backend": {
|
4 |
"name": "pytorch",
|
5 |
+
"version": "2.8.0",
|
6 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
7 |
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
8 |
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
|
|
79 |
"environment": {
|
80 |
"cpu": " AMD EPYC 7R32",
|
81 |
"cpu_count": 16,
|
82 |
+
"cpu_ram_mb": 66697.248768,
|
83 |
"system": "Linux",
|
84 |
"machine": "x86_64",
|
85 |
"platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
|
86 |
"processor": "x86_64",
|
87 |
+
"python_version": "3.10.18",
|
88 |
"gpu": [
|
89 |
"NVIDIA A10G"
|
90 |
],
|
|
|
92 |
"gpu_vram_mb": 24146608128,
|
93 |
"optimum_benchmark_version": "0.6.0.dev0",
|
94 |
"optimum_benchmark_commit": null,
|
95 |
+
"transformers_version": "4.52.4",
|
96 |
"transformers_commit": null,
|
97 |
"accelerate_version": "1.10.0",
|
98 |
"accelerate_commit": null,
|
|
|
102 |
"optimum_commit": null,
|
103 |
"timm_version": "1.0.19",
|
104 |
"timm_commit": null,
|
105 |
+
"peft_version": "0.17.0",
|
106 |
"peft_commit": null
|
107 |
},
|
108 |
"print_report": true,
|