IlyasMoutawwakil HF Staff commited on
Commit
30ab6c2
·
verified ·
1 Parent(s): 7b110e7

Upload cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json with huggingface_hub

Browse files
cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark_config.json CHANGED
@@ -2,7 +2,7 @@
2
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
3
  "backend": {
4
  "name": "pytorch",
5
- "version": "2.8.0+cu128",
6
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
  "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
8
  "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
@@ -79,12 +79,12 @@
79
  "environment": {
80
  "cpu": " AMD EPYC 7R32",
81
  "cpu_count": 16,
82
- "cpu_ram_mb": 66697.240576,
83
  "system": "Linux",
84
  "machine": "x86_64",
85
  "platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
86
  "processor": "x86_64",
87
- "python_version": "3.10.12",
88
  "gpu": [
89
  "NVIDIA A10G"
90
  ],
@@ -92,7 +92,7 @@
92
  "gpu_vram_mb": 24146608128,
93
  "optimum_benchmark_version": "0.6.0.dev0",
94
  "optimum_benchmark_commit": null,
95
- "transformers_version": "4.55.2",
96
  "transformers_commit": null,
97
  "accelerate_version": "1.10.0",
98
  "accelerate_commit": null,
@@ -102,7 +102,7 @@
102
  "optimum_commit": null,
103
  "timm_version": "1.0.19",
104
  "timm_commit": null,
105
- "peft_version": null,
106
  "peft_commit": null
107
  },
108
  "print_report": true,
 
2
  "name": "cuda_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM",
3
  "backend": {
4
  "name": "pytorch",
5
+ "version": "2.8.0",
6
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
  "model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
8
  "processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
 
79
  "environment": {
80
  "cpu": " AMD EPYC 7R32",
81
  "cpu_count": 16,
82
+ "cpu_ram_mb": 66697.248768,
83
  "system": "Linux",
84
  "machine": "x86_64",
85
  "platform": "Linux-5.10.240-238.959.amzn2.x86_64-x86_64-with-glibc2.35",
86
  "processor": "x86_64",
87
+ "python_version": "3.10.18",
88
  "gpu": [
89
  "NVIDIA A10G"
90
  ],
 
92
  "gpu_vram_mb": 24146608128,
93
  "optimum_benchmark_version": "0.6.0.dev0",
94
  "optimum_benchmark_commit": null,
95
+ "transformers_version": "4.52.4",
96
  "transformers_commit": null,
97
  "accelerate_version": "1.10.0",
98
  "accelerate_commit": null,
 
102
  "optimum_commit": null,
103
  "timm_version": "1.0.19",
104
  "timm_commit": null,
105
+ "peft_version": "0.17.0",
106
  "peft_commit": null
107
  },
108
  "print_report": true,