IlyasMoutawwakil
HF Staff
Upload cuda_inference_timm_image-classification_timm/resnet50.a1_in1k/file.json with huggingface_hub
eada27c
verified
| { | |
| "config": { | |
| "name": "cuda_inference_timm_image-classification_timm/resnet50.a1_in1k", | |
| "backend": { | |
| "name": "pytorch", | |
| "version": "2.2.2", | |
| "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", | |
| "task": "image-classification", | |
| "model": "timm/resnet50.a1_in1k", | |
| "library": "timm", | |
| "device": "cuda", | |
| "device_ids": "0", | |
| "seed": 42, | |
| "inter_op_num_threads": null, | |
| "intra_op_num_threads": null, | |
| "hub_kwargs": { | |
| "revision": "main", | |
| "force_download": false, | |
| "local_files_only": false, | |
| "trust_remote_code": false | |
| }, | |
| "no_weights": false, | |
| "device_map": null, | |
| "torch_dtype": null, | |
| "eval_mode": true, | |
| "to_bettertransformer": false, | |
| "low_cpu_mem_usage": null, | |
| "attn_implementation": null, | |
| "cache_implementation": null, | |
| "autocast_enabled": false, | |
| "autocast_dtype": null, | |
| "torch_compile": false, | |
| "torch_compile_target": "forward", | |
| "torch_compile_config": {}, | |
| "quantization_scheme": null, | |
| "quantization_config": {}, | |
| "deepspeed_inference": false, | |
| "deepspeed_inference_config": {}, | |
| "peft_type": null, | |
| "peft_config": {} | |
| }, | |
| "scenario": { | |
| "name": "inference", | |
| "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", | |
| "iterations": 1, | |
| "duration": 1, | |
| "warmup_runs": 1, | |
| "input_shapes": { | |
| "batch_size": 1, | |
| "num_choices": 2, | |
| "sequence_length": 2 | |
| }, | |
| "new_tokens": null, | |
| "latency": true, | |
| "memory": true, | |
| "energy": true, | |
| "forward_kwargs": {}, | |
| "generate_kwargs": { | |
| "max_new_tokens": 2, | |
| "min_new_tokens": 2 | |
| }, | |
| "call_kwargs": { | |
| "num_inference_steps": 2 | |
| } | |
| }, | |
| "launcher": { | |
| "name": "process", | |
| "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", | |
| "device_isolation": true, | |
| "device_isolation_action": "error", | |
| "start_method": "spawn" | |
| }, | |
| "environment": { | |
| "cpu": " AMD EPYC 7R32", | |
| "cpu_count": 16, | |
| "cpu_ram_mb": 66697.29792, | |
| "system": "Linux", | |
| "machine": "x86_64", | |
| "platform": "Linux-5.10.214-202.855.amzn2.x86_64-x86_64-with-glibc2.35", | |
| "processor": "x86_64", | |
| "python_version": "3.10.14", | |
| "gpu": [ | |
| "NVIDIA A10G" | |
| ], | |
| "gpu_count": 1, | |
| "gpu_vram_mb": 24146608128, | |
| "optimum_benchmark_version": "0.2.0", | |
| "optimum_benchmark_commit": null, | |
| "transformers_version": "4.40.2", | |
| "transformers_commit": null, | |
| "accelerate_version": "0.30.0", | |
| "accelerate_commit": null, | |
| "diffusers_version": "0.27.2", | |
| "diffusers_commit": null, | |
| "optimum_version": null, | |
| "optimum_commit": null, | |
| "timm_version": "0.9.16", | |
| "timm_commit": null, | |
| "peft_version": null, | |
| "peft_commit": null | |
| } | |
| }, | |
| "report": { | |
| "forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 945.463296, | |
| "max_global_vram": 1047.003136, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 396.361728, | |
| "max_allocated": 253.912576 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "count": 179, | |
| "total": 0.9962271671295163, | |
| "mean": 0.005565514900164897, | |
| "stdev": 0.0002488616081462451, | |
| "p50": 0.005605440139770507, | |
| "p90": 0.0057956352233886725, | |
| "p95": 0.005870796680450439, | |
| "p99": 0.006352179040908813, | |
| "values": [ | |
| 0.0062341117858886715, | |
| 0.0057825279235839844, | |
| 0.005980160236358643, | |
| 0.005783552169799804, | |
| 0.0059023361206054685, | |
| 0.00582860803604126, | |
| 0.005847040176391601, | |
| 0.005988351821899414, | |
| 0.005825535774230957, | |
| 0.005786623954772949, | |
| 0.005738495826721191, | |
| 0.00568012809753418, | |
| 0.005637119770050049, | |
| 0.005730303764343261, | |
| 0.005656576156616211, | |
| 0.0057118721008300784, | |
| 0.005566400051116943, | |
| 0.005565440177917481, | |
| 0.005565440177917481, | |
| 0.005634047985076904, | |
| 0.005629951953887939, | |
| 0.005697535991668701, | |
| 0.005745664119720459, | |
| 0.005699584007263184, | |
| 0.005723135948181152, | |
| 0.00568832015991211, | |
| 0.005635072231292724, | |
| 0.005598207950592041, | |
| 0.005661695957183838, | |
| 0.005831679821014404, | |
| 0.005714975833892823, | |
| 0.005750783920288086, | |
| 0.005708799839019775, | |
| 0.005798912048339844, | |
| 0.005772287845611572, | |
| 0.005628928184509278, | |
| 0.00572211217880249, | |
| 0.005706751823425293, | |
| 0.005712959766387939, | |
| 0.005703680038452149, | |
| 0.005616608142852783, | |
| 0.00555622386932373, | |
| 0.005549056053161621, | |
| 0.005566463947296142, | |
| 0.005605440139770507, | |
| 0.0055808000564575196, | |
| 0.005587007999420166, | |
| 0.005626880168914795, | |
| 0.005676032066345215, | |
| 0.005665791988372803, | |
| 0.005725183963775635, | |
| 0.005757952213287353, | |
| 0.005761023998260498, | |
| 0.005755904197692871, | |
| 0.005731391906738281, | |
| 0.005735424041748047, | |
| 0.005660672187805176, | |
| 0.005646336078643799, | |
| 0.005675007820129394, | |
| 0.0057784318923950195, | |
| 0.005676032066345215, | |
| 0.005641215801239013, | |
| 0.005752831935882568, | |
| 0.005615647792816162, | |
| 0.005577727794647217, | |
| 0.005608448028564453, | |
| 0.0056483840942382815, | |
| 0.005665823936462402, | |
| 0.005650432109832764, | |
| 0.005794816017150879, | |
| 0.0056442880630493165, | |
| 0.005727231979370117, | |
| 0.005607423782348633, | |
| 0.00572211217880249, | |
| 0.005623807907104492, | |
| 0.005814271926879883, | |
| 0.005708799839019775, | |
| 0.005605408191680908, | |
| 0.005640160083770752, | |
| 0.005639167785644531, | |
| 0.005724160194396972, | |
| 0.0057497601509094234, | |
| 0.0057487359046936035, | |
| 0.005761023998260498, | |
| 0.00577839994430542, | |
| 0.005816319942474365, | |
| 0.005656576156616211, | |
| 0.00570470380783081, | |
| 0.005881855964660645, | |
| 0.00561356782913208, | |
| 0.005836800098419189, | |
| 0.005612512111663818, | |
| 0.007020544052124023, | |
| 0.006162432193756104, | |
| 0.00586956787109375, | |
| 0.005570559978485107, | |
| 0.005627903938293457, | |
| 0.005463071823120117, | |
| 0.005391359806060791, | |
| 0.0053350400924682614, | |
| 0.005341184139251709, | |
| 0.005436416149139404, | |
| 0.0054568958282470706, | |
| 0.005637119770050049, | |
| 0.005470208168029785, | |
| 0.005561344146728516, | |
| 0.005668863773345947, | |
| 0.005683199882507324, | |
| 0.005448703765869141, | |
| 0.005362688064575195, | |
| 0.005315584182739258, | |
| 0.00535756778717041, | |
| 0.005338111877441406, | |
| 0.005311488151550293, | |
| 0.005362688064575195, | |
| 0.005312511920928955, | |
| 0.0052930560111999515, | |
| 0.0053637118339538575, | |
| 0.005345280170440674, | |
| 0.0052971520423889164, | |
| 0.005362688064575195, | |
| 0.005315584182739258, | |
| 0.005330880165100097, | |
| 0.005354496002197265, | |
| 0.005342207908630371, | |
| 0.005303296089172363, | |
| 0.005416959762573242, | |
| 0.0052899842262268066, | |
| 0.005311488151550293, | |
| 0.005306367874145508, | |
| 0.005310463905334473, | |
| 0.005313536167144775, | |
| 0.0053043198585510255, | |
| 0.0053441920280456546, | |
| 0.005280767917633057, | |
| 0.005325823783874512, | |
| 0.005323775768280029, | |
| 0.005319680213928223, | |
| 0.005311488151550293, | |
| 0.0053350400924682614, | |
| 0.0053043198585510255, | |
| 0.005339136123657226, | |
| 0.005305344104766845, | |
| 0.0053002238273620605, | |
| 0.005296127796173096, | |
| 0.005336063861846924, | |
| 0.005282815933227539, | |
| 0.006303743839263916, | |
| 0.006523903846740723, | |
| 0.005336063861846924, | |
| 0.0053309440612792965, | |
| 0.005406720161437988, | |
| 0.005380095958709717, | |
| 0.0053606400489807126, | |
| 0.005325823783874512, | |
| 0.00538316822052002, | |
| 0.00531660795211792, | |
| 0.0053432960510253905, | |
| 0.005314559936523438, | |
| 0.0053350400924682614, | |
| 0.00530944013595581, | |
| 0.005342207908630371, | |
| 0.0053606400489807126, | |
| 0.005343232154846191, | |
| 0.0053002238273620605, | |
| 0.005309375762939453, | |
| 0.005296127796173096, | |
| 0.00531763219833374, | |
| 0.005296127796173096, | |
| 0.005326848030090332, | |
| 0.005292064189910889, | |
| 0.0053012480735778805, | |
| 0.005320703983306885, | |
| 0.0053043198585510255, | |
| 0.005311456203460694, | |
| 0.005326848030090332, | |
| 0.005321728229522705, | |
| 0.005291007995605469, | |
| 0.005342207908630371 | |
| ] | |
| }, | |
| "throughput": { | |
| "unit": "samples/s", | |
| "value": 179.67789466709928 | |
| }, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 6.292989539931902e-08, | |
| "ram": 3.44065424296931e-08, | |
| "gpu": 1.3212679588359838e-07, | |
| "total": 2.2946323371261053e-07 | |
| }, | |
| "efficiency": { | |
| "unit": "samples/kWh", | |
| "value": 4357996.633362372 | |
| } | |
| } | |
| } | |
| } |